10 resultados para global optimization algorithms
Resumo:
Optimization is a very important field for getting the best possible value for the optimization function. Continuous optimization is optimization over real intervals. There are many global and local search techniques. Global search techniques try to get the global optima of the optimization problem. However, local search techniques are used more since they try to find a local minimal solution within an area of the search space. In Continuous Constraint Satisfaction Problems (CCSP)s, constraints are viewed as relations between variables, and the computations are supported by interval analysis. The continuous constraint programming framework provides branch-and-prune algorithms for covering sets of solutions for the constraints with sets of interval boxes which are the Cartesian product of intervals. These algorithms begin with an initial crude cover of the feasible space (the Cartesian product of the initial variable domains) which is recursively refined by interleaving pruning and branching steps until a stopping criterion is satisfied. In this work, we try to find a convenient way to use the advantages in CCSP branchand- prune with local search of global optimization applied locally over each pruned branch of the CCSP. We apply local search techniques of continuous optimization over the pruned boxes outputted by the CCSP techniques. We mainly use steepest descent technique with different characteristics such as penalty calculation and step length. We implement two main different local search algorithms. We use “Procure”, which is a constraint reasoning and global optimization framework, to implement our techniques, then we produce and introduce our results over a set of benchmarks.
Resumo:
Despite the extensive literature in finding new models to replace the Markowitz model or trying to increase the accuracy of its input estimations, there is less studies about the impact on the results of using different optimization algorithms. This paper aims to add some research to this field by comparing the performance of two optimization algorithms in drawing the Markowitz Efficient Frontier and in real world investment strategies. Second order cone programming is a faster algorithm, appears to be more efficient, but is impossible to assert which algorithm is better. Quadratic Programming often shows superior performance in real investment strategies.
Resumo:
Due to the progresses made in the branch of embedded technologies, manufacturers are becoming able to pack their shop floor level manufacturing resources with even more complex functionalities. This technological progression is radically changing the way production systems are designed and deployed, as well as, monitored and controlled. The dissemination of smart devices inside production processes confers new visibility on the production system while enabling for a more efficient and effective management of the operations. By turning the current manufacturing resources functionalities into services based on a Service Oriented Architecture (SOA), in order to expose them as a service to the user, the binomial manufacturing resource/service will push the entire manufacturing enterprise visibility to another level while enabling the global optimization of the operations and processes of a production system while, at the same time, supporting its accommodation to the operational spike easily and with reduced impact on production. The present work implements a Cloud Manufacturing infrastructure for achieving the resource/service value-added i.e. to facilitate the creation of services that are the composition of currently available atomic services. In this context, manufacturing resource virtualization (i.e. formalization of resources capabilities into services accessible inside and outside the enterprise) and semantic representation/description are the pillars for achieving resource service composition. In conclusion, the present work aims to act on the manufacturing resource layer where physical resources and shop floor capabilities are going to be provided to the user as a SaaS (Software as a Service) and/or IaaS (Infrastructure as a Service).
Resumo:
Thesis presented in partial fulfillment of the requirements for the degree of Doctor of Philosophy in the subject of Electrical and Computer Engineering
Resumo:
Dissertação apresentada na faculdade de Ciências e Tecnologia da Universidade Nova de Lisboa para a obtenção do grau de Mestre em Engenharia Electrotécnica e de Computadores
Resumo:
Submitted in partial fulfillment for the Requirements for the Degree of PhD in Mathematics, in the Speciality of Statistics in the Faculdade de Ciências e Tecnologia
Resumo:
In developed countries, civil infrastructures are one of the most significant investments of governments, corporations, and individuals. Among these, transportation infrastructures, including highways, bridges, airports, and ports, are of huge importance, both economical and social. Most developed countries have built a fairly complete network of highways to fit their needs. As a result, the required investment in building new highways has diminished during the last decade, and should be further reduced in the following years. On the other hand, significant structural deteriorations have been detected in transportation networks, and a huge investment is necessary to keep these infrastructures safe and serviceable. Due to the significant importance of bridges in the serviceability of highway networks, maintenance of these structures plays a major role. In this paper, recent progress in probabilistic maintenance and optimization strategies for deteriorating civil infrastructures with emphasis on bridges is summarized. A novel model including interaction between structural safety analysis,through the safety index, and visual inspections and non destructive tests, through the condition index, is presented. Single objective optimization techniques leading to maintenance strategies associated with minimum expected cumulative cost and acceptable levels of condition and safety are presented. Furthermore, multi-objective optimization is used to simultaneously consider several performance indicators such as safety, condition, and cumulative cost. Realistic examples of the application of some of these techniques and strategies are also presented.
Resumo:
A Work Project, presented as part of the requirements for the Award of a Masters Degree in Management from the NOVA – School of Business and Economics
Resumo:
A Work Project, presented as part of the requirements for the Award of a Masters Degree in Management from the NOVA – School of Business and Economics
Resumo:
Breast cancer is the most common cancer among women, being a major public health problem. Worldwide, X-ray mammography is the current gold-standard for medical imaging of breast cancer. However, it has associated some well-known limitations. The false-negative rates, up to 66% in symptomatic women, and the false-positive rates, up to 60%, are a continued source of concern and debate. These drawbacks prompt the development of other imaging techniques for breast cancer detection, in which Digital Breast Tomosynthesis (DBT) is included. DBT is a 3D radiographic technique that reduces the obscuring effect of tissue overlap and appears to address both issues of false-negative and false-positive rates. The 3D images in DBT are only achieved through image reconstruction methods. These methods play an important role in a clinical setting since there is a need to implement a reconstruction process that is both accurate and fast. This dissertation deals with the optimization of iterative algorithms, with parallel computing through an implementation on Graphics Processing Units (GPUs) to make the 3D reconstruction faster using Compute Unified Device Architecture (CUDA). Iterative algorithms have shown to produce the highest quality DBT images, but since they are computationally intensive, their clinical use is currently rejected. These algorithms have the potential to reduce patient dose in DBT scans. A method of integrating CUDA in Interactive Data Language (IDL) is proposed in order to accelerate the DBT image reconstructions. This method has never been attempted before for DBT. In this work the system matrix calculation, the most computationally expensive part of iterative algorithms, is accelerated. A speedup of 1.6 is achieved proving the fact that GPUs can accelerate the IDL implementation.