975 resultados para Non-lineal optimization
Resumo:
We have investigated and extensively tested three families of non-convex optimization approaches for solving the transmission network expansion planning problem: simulated annealing (SA), genetic algorithms (GA), and tabu search algorithms (TS). The paper compares the main features of the three approaches and presents an integrated view of these methodologies. A hybrid approach is then proposed which presents performances which are far better than the ones obtained with any of these approaches individually. Results obtained in tests performed with large scale real-life networks are summarized.
Resumo:
We have investigated and extensively tested three families of non-convex optimization approaches for solving the transmission network expansion planning problem: simulated annealing (SA), genetic algorithms (GA), and tabu search algorithms (TS). The paper compares the main features of the three approaches and presents an integrated view of these methodologies. A hybrid approach is then proposed which presents performances which are far better than the ones obtained with any of these approaches individually. Results obtained in tests performed with large scale real-life networks are summarized.
Resumo:
The electro-dynamical tethers emit waves in structured denominated Alfven wings. The Derivative Nonlineal Schrödinger Equation (DNLS) possesses the capacity to describe the propagation of circularly polarized Alfven waves of finite amplitude in cold plasmas. The DNLS equation is truncated to explore the coherent, weakly nonlinear, cubic coupling of three waves near resonance, one wave being linearly unstable and the other waves damped. In this article is presented a theoretical and numerical analysis when the growth rate of the unstable wave is next to zero considering two damping models: Landau and resistive. The DNLS equation presents a chaotic dynamics when is consider only three wave truncation. The evolution to chaos possesses three routes: hard transition, period-doubling and intermittence of type I.
Resumo:
We propose an iterative procedure to minimize the sum of squares function which avoids the nonlinear nature of estimating the first order moving average parameter and provides a closed form of the estimator. The asymptotic properties of the method are discussed and the consistency of the linear least squares estimator is proved for the invertible case. We perform various Monte Carlo experiments in order to compare the sample properties of the linear least squares estimator with its nonlinear counterpart for the conditional and unconditional cases. Some examples are also discussed
Resumo:
We propose an iterative procedure to minimize the sum of squares function which avoids the nonlinear nature of estimating the first order moving average parameter and provides a closed form of the estimator. The asymptotic properties of the method are discussed and the consistency of the linear least squares estimator is proved for the invertible case. We perform various Monte Carlo experiments in order to compare the sample properties of the linear least squares estimator with its nonlinear counterpart for the conditional and unconditional cases. Some examples are also discussed
Resumo:
This paper presents a methodology that aims to increase the probability of delivering power to any load point of the electrical distribution system by identifying new investments in distribution components. The methodology is based on statistical failure and repair data of the distribution power system components and it uses fuzzy-probabilistic modelling for system component outage parameters. Fuzzy membership functions of system component outage parameters are obtained by statistical records. A mixed integer non-linear optimization technique is developed to identify adequate investments in distribution networks components that allow increasing the availability level for any customer in the distribution system at minimum cost for the system operator. To illustrate the application of the proposed methodology, the paper includes a case study that considers a real distribution network.
Resumo:
Important research effort has been devoted to the topic of optimal planning of distribution systems. The non linear nature of the system, the need to consider a large number of scenarios and the increasing necessity to deal with uncertainties make optimal planning in distribution systems a difficult task. Heuristic techniques approaches have been proposed to deal with these issues, overcoming some of the inherent difficulties of classic methodologies. This paper considers several methodologies used to address planning problems of electrical power distribution networks, namely mixedinteger linear programming (MILP), ant colony algorithms (AC), genetic algorithms (GA), tabu search (TS), branch exchange (BE), simulated annealing (SA) and the Bender´s decomposition deterministic non-linear optimization technique (BD). Adequacy of theses techniques to deal with uncertainties is discussed. The behaviour of each optimization technique is compared from the point of view of the obtained solution and of the methodology performance. The paper presents results of the application of these optimization techniques to a real case of a 10-kV electrical distribution system with 201 nodes that feeds an urban area.
Resumo:
The elastic behavior of the demand consumption jointly used with other available resources such as distributed generation (DG) can play a crucial role for the success of smart grids. The intensive use of Distributed Energy Resources (DER) and the technical and contractual constraints result in large-scale non linear optimization problems that require computational intelligence methods to be solved. This paper proposes a Particle Swarm Optimization (PSO) based methodology to support the minimization of the operation costs of a virtual power player that manages the resources in a distribution network and the network itself. Resources include the DER available in the considered time period and the energy that can be bought from external energy suppliers. Network constraints are considered. The proposed approach uses Gaussian mutation of the strategic parameters and contextual self-parameterization of the maximum and minimum particle velocities. The case study considers a real 937 bus distribution network, with 20310 consumers and 548 distributed generators. The obtained solutions are compared with a deterministic approach and with PSO without mutation and Evolutionary PSO, both using self-parameterization.
Resumo:
This thesis investigates a method for human-robot interaction (HRI) in order to uphold productivity of industrial robots like minimization of the shortest operation time, while ensuring human safety like collision avoidance. For solving such problems an online motion planning approach for robotic manipulators with HRI has been proposed. The approach is based on model predictive control (MPC) with embedded mixed integer programming. The planning strategies of the robotic manipulators mainly considered in the thesis are directly performed in the workspace for easy obstacle representation. The non-convex optimization problem is approximated by a mixed-integer program (MIP). It is further effectively reformulated such that the number of binary variables and the number of feasible integer solutions are drastically decreased. Safety-relevant regions, which are potentially occupied by the human operators, can be generated online by a proposed method based on hidden Markov models. In contrast to previous approaches, which derive predictions based on probability density functions in the form of single points, such as most likely or expected human positions, the proposed method computes safety-relevant subsets of the workspace as a region which is possibly occupied by the human at future instances of time. The method is further enhanced by combining reachability analysis to increase the prediction accuracy. These safety-relevant regions can subsequently serve as safety constraints when the motion is planned by optimization. This way one arrives at motion plans that are safe, i.e. plans that avoid collision with a probability not less than a predefined threshold. The developed methods have been successfully applied to a developed demonstrator, where an industrial robot works in the same space as a human operator. The task of the industrial robot is to drive its end-effector according to a nominal sequence of grippingmotion-releasing operations while no collision with a human arm occurs.
Resumo:
Analisa-se se os funcionamentos inferenciais têm uma estrutura própria dos sistemas dinâmicos não lineais, estudados a partir de quatro gráficas humorísticas. Os primeiros resultados com o tratamento estadístico lineal de K-medias projetam a presencia de perfis de diferentes funcionamentos inferenciais em função das diferentes piadas. Os resultados com a técnica da wavelet, proveniente dos sistemas dinâmicos não lineais, mostram patrões dos funcionamentos inferenciais que dão conta de sua natureza multifractal, sem uma sequencialidade fixa e sem uma organização aparente. Isto implica que é necessário revisar a concepção de estádios sequenciais fixos como os que dominam os estudos de desenvolvimento cognitivo.
Resumo:
Este documento se centra en la presentación de información y análisis de la misma a la hora de establecer la manera en que empresas del sector de extracción de gas natural y generación de energía a base de dicho recurso, toman decisiones en cuanto a inversión, centrándose en la lógica que usan a la hora de emprender este proceso. Esto debido a la constante necesidad de establecer procesos que permitan tomar decisiones más acertadas, incluyendo todas las herramientas posibles para tal fin. La lógica es una de estas herramientas, pues permite encadenar factores con el fin de obtener resultados positivos. Por tal razón, se hace importante conocer el uso de esta herramienta, teniendo en cuentas de qué manera y en que contextos es usada. Con el fin de tener una mayor orientación, este estudio estará centrado en un sector específico, el cual es el de la extracción de petróleo y gas natural. Lo anterior entendiendo la necesidad existente de fundamentación teórica que permita establecer de manera clara la forma apropiada de tomar decisiones en un sector tan diverso y complejo como lo es el mencionado. El contexto empresarial actual exige una visión global, no basada en la lógica lineal causal que hoy se tiene como referencia. El sector de extracción de petróleo y gas natural es un ejemplo particular en cuanto a la manera en cuanto se toman decisiones en inversión, puesto que en su mayoría son empresas de capital intensivo, las cuales mantienen un flujo elevado de recursos monetarios.
Resumo:
This paper shows the Particle Swarm Optimization algorithm with a Differential Evolution. Each candidate solution is sampled in the interval [?5, 5] D where D indicates the dimension of the search space, and the evolution is performed with a classical PSO algorithm and a classical DE/x/1 algorithm according to a random threshold. Moreover, this paper provides concepts to deal with non-linear optimization through the use of PSO.
Resumo:
A lo largo del presente trabajo se investiga la viabilidad de la descomposición automática de espectros de radiación gamma por medio de algoritmos de resolución de sistemas de ecuaciones algebraicas lineales basados en técnicas de pseudoinversión. La determinación de dichos algoritmos ha sido realizada teniendo en cuenta su posible implementación sobre procesadores de propósito específico de baja complejidad. En el primer capítulo se resumen las técnicas para la detección y medida de la radiación gamma que han servido de base para la confección de los espectros tratados en el trabajo. Se reexaminan los conceptos asociados con la naturaleza de la radiación electromagnética, así como los procesos físicos y el tratamiento electrónico que se hallan involucrados en su detección, poniendo de relieve la naturaleza intrínsecamente estadística del proceso de formación del espectro asociado como una clasificación del número de detecciones realizadas en función de la energía supuestamente continua asociada a las mismas. Para ello se aporta una breve descripción de los principales fenómenos de interacción de la radiación con la materia, que condicionan el proceso de detección y formación del espectro. El detector de radiación es considerado el elemento crítico del sistema de medida, puesto que condiciona fuertemente el proceso de detección. Por ello se examinan los principales tipos de detectores, con especial hincapié en los detectores de tipo semiconductor, ya que son los más utilizados en la actualidad. Finalmente, se describen los subsistemas electrónicos fundamentales para el acondicionamiento y pretratamiento de la señal procedente del detector, a la que se le denomina con el término tradicionalmente utilizado de Electrónica Nuclear. En lo que concierne a la espectroscopia, el principal subsistema de interés para el presente trabajo es el analizador multicanal, el cual lleva a cabo el tratamiento cualitativo de la señal, y construye un histograma de intensidad de radiación en el margen de energías al que el detector es sensible. Este vector N-dimensional es lo que generalmente se conoce con el nombre de espectro de radiación. Los distintos radionúclidos que participan en una fuente de radiación no pura dejan su impronta en dicho espectro. En el capítulo segundo se realiza una revisión exhaustiva de los métodos matemáticos en uso hasta el momento ideados para la identificación de los radionúclidos presentes en un espectro compuesto, así como para determinar sus actividades relativas. Uno de ellos es el denominado de regresión lineal múltiple, que se propone como la aproximación más apropiada a los condicionamientos y restricciones del problema: capacidad para tratar con espectros de baja resolución, ausencia del concurso de un operador humano (no supervisión), y posibilidad de ser soportado por algoritmos de baja complejidad capaces de ser instrumentados sobre procesadores dedicados de alta escala de integración. El problema del análisis se plantea formalmente en el tercer capítulo siguiendo las pautas arriba mencionadas y se demuestra que el citado problema admite una solución en la teoría de memorias asociativas lineales. Un operador basado en este tipo de estructuras puede proporcionar la solución al problema de la descomposición espectral deseada. En el mismo contexto, se proponen un par de algoritmos adaptativos complementarios para la construcción del operador, que gozan de unas características aritméticas especialmente apropiadas para su instrumentación sobre procesadores de alta escala de integración. La característica de adaptatividad dota a la memoria asociativa de una gran flexibilidad en lo que se refiere a la incorporación de nueva información en forma progresiva.En el capítulo cuarto se trata con un nuevo problema añadido, de índole altamente compleja. Es el del tratamiento de las deformaciones que introducen en el espectro las derivas instrumentales presentes en el dispositivo detector y en la electrónica de preacondicionamiento. Estas deformaciones invalidan el modelo de regresión lineal utilizado para describir el espectro problema. Se deriva entonces un modelo que incluya las citadas deformaciones como una ampliación de contribuciones en el espectro compuesto, el cual conlleva una ampliación sencilla de la memoria asociativa capaz de tolerar las derivas en la mezcla problema y de llevar a cabo un análisis robusto de contribuciones. El método de ampliación utilizado se basa en la suposición de pequeñas perturbaciones. La práctica en el laboratorio demuestra que, en ocasiones, las derivas instrumentales pueden provocar distorsiones severas en el espectro que no pueden ser tratadas por el modelo anterior. Por ello, en el capítulo quinto se plantea el problema de medidas afectadas por fuertes derivas desde el punto de vista de la teoría de optimización no lineal. Esta reformulación lleva a la introducción de un algoritmo de tipo recursivo inspirado en el de Gauss-Newton que permite introducir el concepto de memoria lineal realimentada. Este operador ofrece una capacidad sensiblemente mejorada para la descomposición de mezclas con fuerte deriva sin la excesiva carga computacional que presentan los algoritmos clásicos de optimización no lineal. El trabajo finaliza con una discusión de los resultados obtenidos en los tres principales niveles de estudio abordados, que se ofrecen en los capítulos tercero, cuarto y quinto, así como con la elevación a definitivas de las principales conclusiones derivadas del estudio y con el desglose de las posibles líneas de continuación del presente trabajo.---ABSTRACT---Through the present research, the feasibility of Automatic Gamma-Radiation Spectral Decomposition by Linear Algebraic Equation-Solving Algorithms using Pseudo-Inverse Techniques is explored. The design of the before mentioned algorithms has been done having into account their possible implementation on Specific-Purpose Processors of Low Complexity. In the first chapter, the techniques for the detection and measurement of gamma radiation employed to construct the spectra being used throughout the research are reviewed. Similarly, the basic concepts related with the nature and properties of the hard electromagnetic radiation are also re-examined, together with the physic and electronic processes involved in the detection of such kind of radiation, with special emphasis in the intrinsic statistical nature of the spectrum build-up process, which is considered as a classification of the number of individual photon-detections as a function of the energy associated to each individual photon. Fbr such, a brief description of the most important matter-energy interaction phenomena conditioning the detection and spectrum formation processes is given. The radiation detector is considered as the most critical element in the measurement system, as this device strongly conditions the detection process. Fbr this reason, the characteristics of the most frequent detectors are re-examined, with special emphasis on those of semiconductor nature, as these are the most frequently employed ones nowadays. Finally, the fundamental electronic subsystems for preaconditioning and treating of the signal delivered by the detector, classically addresed as Nuclear Electronics, is described. As far as Spectroscopy is concerned, the subsystem most interesting for the scope covered by the present research is the so-called Multichannel Analyzer, which is devoted to the cualitative treatment of the signal, building-up a hystogram of radiation intensity in the range of energies in which the detector is sensitive. The resulting N-dimensional vector is generally known with the ñame of Radiation Spectrum. The different radio-nuclides contributing to the spectrum of a composite source will leave their fingerprint in the resulting spectrum. Through the second chapter, an exhaustive review of the mathematical methods devised to the present moment to identify the radio-nuclides present in the composite spectrum and to quantify their relative contributions, is reviewed. One of the more popular ones is the so-known Múltiple Linear Regression, which is proposed as the best suited approach according to the constraints and restrictions present in the formulation of the problem, i.e., the need to treat low-resolution spectra, the absence of control by a human operator (un-supervision), and the possibility of being implemented as low-complexity algorithms amenable of being supported by VLSI Specific Processors. The analysis problem is formally stated through the third chapter, following the hints established in this context, and it is shown that the addressed problem may be satisfactorily solved under the point of view of Linear Associative Memories. An operator based on this kind of structures may provide the solution to the spectral decomposition problem posed. In the same context, a pair of complementary adaptive algorithms useful for the construction of the solving operator are proposed, which share certain special arithmetic characteristics that render them specially suitable for their implementation on VLSI Processors. The adaptive nature of the associative memory provides a high flexibility to this operator, in what refers to the progressive inclusión of new information to the knowledge base. Through the fourth chapter, this fact is treated together with a new problem to be considered, of a high interest but quite complex nature, as is the treatment of the deformations appearing in the spectrum when instrumental drifts in both the detecting device and the pre-acconditioning electronics are to be taken into account. These deformations render the Linear Regression Model proposed almost unuseful to describe the resulting spectrum. A new model including the drifts is derived as an extensión of the individual contributions to the composite spectrum, which implies a simple extensión of the Associative Memory, which renders this suitable to accept the drifts in the composite spectrum, thus producing a robust analysis of contributions. The extensión method is based on the Low-Amplitude Perturbation Hypothesis. Experimental practice shows that in certain cases the instrumental drifts may provoke severe distortions in the resulting spectrum, which can not be treated with the before-mentioned hypothesis. To cover also these less-frequent cases, through the fifth chapter, the problem involving strong drifts is treated under the point of view of Non-Linear Optimization Techniques. This reformulation carries the study to the consideration of recursive algorithms based on the Gauss-Newton methods, which allow the introduction of Feed-Back Memories, computing elements with a sensibly improved capability to decompose spectra affected by strong drifts. The research concludes with a discussion of the results obtained in the three main levéis of study considerad, which are presented in chapters third, fourth and fifth, toghether with the review of the main conclusions derived from the study and the outline of the main research lines opened by the present work.
Resumo:
La presente tesis analiza la mejora de la resistencia estructural ante vuelco de autocares enfocando dos vías de actuación: análisis y propuestas de requisitos reglamentarios a nivel europeo y la generación de herramientas que ayuden al diseño y a la verificación de estos requisitos. Los requisitos reglamentarios de resistencia estructural a vuelco contemplan la superestructura de los vehículos pero no para los asientos y sistemas de retención. La influencia de los pasajeros retenidos es superior a la incluida en reglamentación (Reg. 66.01) debiendo considerarse unida al vehículo un porcentaje de la masa de los pasajeros del 91% para cinturón de tres puntos y del 52% para cinturón subabdominal frente al 50% reglamentario para todos los casos. Se ha determinado la cinemática y dinámica del vuelco normativo en sus diferentes fases, formulando las energías en las fases iniciales (hasta el impacto contra el suelo) y determinando la fase final de deformación a través del análisis secuencial de ensayos de módulos reales. Se han determinado los esfuerzos para los asientos que se dividen en dos fases diferenciadas temporalmente: una primera debida a la deformación estructural y una segunda debida al esfuerzo del pasajero retenido que se produce en sentido opuesto (con una deceleración del pasajero en torno a 3.3 g). Se ha caracterizado a través de ensayos cuasi.estáticos el comportamiento de perfiles a flexión y de las uniones estructurales de las principales zonas del vehículo (piso, ventana y techo) verificándose la validez del comportamiento plástico teórico Kecman.García para perfiles de hasta 4 mm de espesor y caracterizando la resistencia y rigidez en la zona elástica de las uniones en función del tipo de refuerzo, materiales y perfiles (análisis de más de 180 probetas). Se ha definido un método de ensayo cuasi.estático para asientos ante esfuerzos de vuelco, ensayándose 19 butacas y determinándose que son resistentes (salvo las uniones a vehículo con pinzas), que son capaces de absorber hasta más de un 17% de la energía absorbida, aunque algunos necesitan optimización para llegar a contribuir en el mecanismo de deformación estructural. Se han generado modelos simplificados para introducir en los modelos barra.rótula plástica: un modelo combinado unión+rótula plástica (que incluye la zona de rigidez determinada en función del tipo de unión) para la superestructura y un modelo simplificado de muelles no.lineales para los asientos. Igualmente se ha generado la metodología de diseño a través de ensayos virtuales con modelos de detalle de elementos finitos tanto de las uniones como de los asientos. Se ha propuesto una metodología de diseño basada en obtener el “mecanismo óptimo de deformación estructural” (elevando la zona de deformación lateral a nivel de ventana y en pilar o en costilla en techo). Para ello se abren dos vías: diseño de la superestructura (selección de perfiles y generación de uniones resistentes) o combinación con asientos (que en lugar de solo resistir las cargas pueden llegar a modificar el mecanismo de deformación). Se ha propuesto una metodología de verificación alternativa al vuelco de vehículo completo que contempla el cálculo cuasi.estático con modelos simplificados barra.rótula plástica más el ensayo de una sección representativa con asientos y utillajes antropomórficos retenidos que permite validar el diseño de las uniones, determinar el porcentaje de energía que debe absorberse por deformación estructural (factor C) y verificar el propio asiento como sistema de retención. ABSTRACT This research analyzes the improvement of the structural strength of buses and coaches under rollover from two perspectives: regulatory requirements at European level and generation of tools that will help to the design and to the verification of requirements. European Regulations about rollover structural strength includes requirements for the superstructure of the vehicles but not about seats, anchorages and restraint systems. The influence of the retained passengers is higher than the one included currently in the Regulations (Reg. 66.01), being needed to consider a 91% of the passenger mass as rigidly joint to the vehicle (for 3 points’ belt, a 52% for 2 points’ belt) instead of the 50% included in the Regulation. Kinematic and dynamic of the normative rollover has been determined from testing of different sections, formulating the energies of the first phases (up to the first impact with the ground) and determining the last deformation phase through sequential analysis of movements and deformations. The efforts due to rollover over the seats have been established, being divided in two different temporal phases: a first one due to the structural deformation of the vehicle and a second one due to the effort of the restrained passenger being this second one in opposite sense (with a passenger deceleration around 3.3 g). From quasi.static testing, the behavior of the structural tubes under flexural loads, including the principal joints in the vehicle (floor, window and roof), the validity of the theoretical plastic behavior according Kecman.García theories have been verified up to 4 mm of thickness. Strength of the joints as well as the stiffness of the elastic zone has been determined in function of main parameters: type of reinforcement, materials and section of the tubes (more than 180 test specimens). It has been defined a quasi.static testing methodology to characterize the seats and restrain system behavior under rollover, testing 19 double seats and concluding that they are resistant (excepting clamping joints), that they can absorb more than a 17 of the absorbed energy, and that some of them need optimization to contribute in the structural deformation mechanism. It has been generated simplified MEF models, to analyze in a beam.plastic hinge model: a combined model joint+plastic hinge (including the stiffness depending on the type of joint) for the superstructure and a simplified model with non.lineal springs to represent the seats. It has been detailed methodologies for detailed design of joints and seats from virtual testing (MEF models). A design methodology based in the “optimized structural deformation mechanism” (increasing the height of deformation of the lateral up to window level) is proposed. Two possibilities are analyzed: design of the superstructure based on the selection of profiles and design of strength joints (were seats only resist the efforts and contribute in the energy absorption) or combination structure.seats, were seats contributes in the deformation mechanism. An alternative methodology to the rollover of a vehicle that includes the quasi.static calculation with simplified models “beam.joint+plastic hinge” plus the testing of a representative section of the vehicle including seats and anthropomorphic ballast restrained by the safety belts is presented. The test of the section allows validate the design of the joints, determine the percentage of energy to be absorbed by structural deformation (factor C) and verify the seat as a retention system.
Resumo:
It has been reported that for certain colour samples, the chromatic adaptation transform CAT02 imbedded in the CIECAM02 colour appearance model predicts corresponding colours with negative tristimulus values (TSVs), which can cause problems in certain applications. To overcome this problem, a mathematical approach is proposed for modifying CAT02. This approach combines a non-negativity constraint for the TSVs of corresponding colours with the minimization of the colour differences between those values for the corresponding colours obtained by visual observations and the TSVs of the corresponding colours predicted by the model, which is a constrained non-linear optimization problem. By solving the non-linear optimization problem, a new matrix is found. The performance of the CAT02 transform with various matrices including the original CAT02 matrix, and the new matrix are tested using visual datasets and the optimum colours. Test results show that the CAT02 with the new matrix predicted corresponding colours without negative TSVs for all optimum colours and the colour matching functions of the two CIE standard observers under the test illuminants considered. However, the accuracy with the new matrix for predicting the visual data is approximately 1 CIELAB colour difference unit worse compared with the original CAT02. This indicates that accuracy has to be sacrificed to achieve the non-negativity constraint for the TSVs of the corresponding colours.