36 resultados para penalty function method

em Universidad Politécnica de Madrid


Relevância:

90.00% 90.00%

Publicador:

Resumo:

La evaluación de la seguridad de estructuras antiguas de fábrica es un problema abierto.El material es heterogéneo y anisótropo, el estado previo de tensiones difícil de conocer y las condiciones de contorno inciertas. A comienzos de los años 50 se demostró que el análisis límite era aplicable a este tipo de estructuras, considerándose desde entonces como una herramienta adecuada. En los casos en los que no se produce deslizamiento la aplicación de los teoremas del análisis límite estándar constituye una herramienta formidable por su simplicidad y robustez. No es necesario conocer el estado real de tensiones. Basta con encontrar cualquier solución de equilibrio, y que satisfaga las condiciones de límite del material, en la seguridad de que su carga será igual o inferior a la carga real de inicio de colapso. Además esta carga de inicio de colapso es única (teorema de la unicidad) y se puede obtener como el óptimo de uno cualquiera entre un par de programas matemáticos convexos duales. Sin embargo, cuando puedan existir mecanismos de inicio de colapso que impliquen deslizamientos, cualquier solución debe satisfacer tanto las restricciones estáticas como las cinemáticas, así como un tipo especial de restricciones disyuntivas que ligan las anteriores y que pueden plantearse como de complementariedad. En este último caso no está asegurada la existencia de una solución única, por lo que es necesaria la búsqueda de otros métodos para tratar la incertidumbre asociada a su multiplicidad. En los últimos años, la investigación se ha centrado en la búsqueda de un mínimo absoluto por debajo del cual el colapso sea imposible. Este método es fácil de plantear desde el punto de vista matemático, pero intratable computacionalmente, debido a las restricciones de complementariedad 0 y z 0 que no son ni convexas ni suaves. El problema de decisión resultante es de complejidad computacional No determinista Polinomial (NP)- completo y el problema de optimización global NP-difícil. A pesar de ello, obtener una solución (sin garantía de exito) es un problema asequible. La presente tesis propone resolver el problema mediante Programación Lineal Secuencial, aprovechando las especiales características de las restricciones de complementariedad, que escritas en forma bilineal son del tipo y z = 0; y 0; z 0 , y aprovechando que el error de complementariedad (en forma bilineal) es una función de penalización exacta. Pero cuando se trata de encontrar la peor solución, el problema de optimización global equivalente es intratable (NP-difícil). Además, en tanto no se demuestre la existencia de un principio de máximo o mínimo, existe la duda de que el esfuerzo empleado en aproximar este mínimo esté justificado. En el capítulo 5, se propone hallar la distribución de frecuencias del factor de carga, para todas las soluciones de inicio de colapso posibles, sobre un sencillo ejemplo. Para ello, se realiza un muestreo de soluciones mediante el método de Monte Carlo, utilizando como contraste un método exacto de computación de politopos. El objetivo final es plantear hasta que punto está justificada la busqueda del mínimo absoluto y proponer un método alternativo de evaluación de la seguridad basado en probabilidades. Las distribuciones de frecuencias, de los factores de carga correspondientes a las soluciones de inicio de colapso obtenidas para el caso estudiado, muestran que tanto el valor máximo como el mínimo de los factores de carga son muy infrecuentes, y tanto más, cuanto más perfecto y contínuo es el contacto. Los resultados obtenidos confirman el interés de desarrollar nuevos métodos probabilistas. En el capítulo 6, se propone un método de este tipo basado en la obtención de múltiples soluciones, desde puntos de partida aleatorios y calificando los resultados mediante la Estadística de Orden. El propósito es determinar la probabilidad de inicio de colapso para cada solución.El método se aplica (de acuerdo a la reducción de expectativas propuesta por la Optimización Ordinal) para obtener una solución que se encuentre en un porcentaje determinado de las peores. Finalmente, en el capítulo 7, se proponen métodos híbridos, incorporando metaheurísticas, para los casos en que la búsqueda del mínimo global esté justificada. Abstract Safety assessment of the historic masonry structures is an open problem. The material is heterogeneous and anisotropic, the previous state of stress is hard to know and the boundary conditions are uncertain. In the early 50's it was proven that limit analysis was applicable to this kind of structures, being considered a suitable tool since then. In cases where no slip occurs, the application of the standard limit analysis theorems constitutes an excellent tool due to its simplicity and robustness. It is enough find any equilibrium solution which satisfy the limit constraints of the material. As we are certain that this load will be equal to or less than the actual load of the onset of collapse, it is not necessary to know the actual stresses state. Furthermore this load for the onset of collapse is unique (uniqueness theorem), and it can be obtained as the optimal from any of two mathematical convex duals programs However, if the mechanisms of the onset of collapse involve sliding, any solution must satisfy both static and kinematic constraints, and also a special kind of disjunctive constraints linking the previous ones, which can be formulated as complementarity constraints. In the latter case, it is not guaranted the existence of a single solution, so it is necessary to look for other ways to treat the uncertainty associated with its multiplicity. In recent years, research has been focused on finding an absolute minimum below which collapse is impossible. This method is easy to set from a mathematical point of view, but computationally intractable. This is due to the complementarity constraints 0 y z 0 , which are neither convex nor smooth. The computational complexity of the resulting decision problem is "Not-deterministic Polynomialcomplete" (NP-complete), and the corresponding global optimization problem is NP-hard. However, obtaining a solution (success is not guaranteed) is an affordable problem. This thesis proposes solve that problem through Successive Linear Programming: taking advantage of the special characteristics of complementarity constraints, which written in bilinear form are y z = 0; y 0; z 0 ; and taking advantage of the fact that the complementarity error (bilinear form) is an exact penalty function. But when it comes to finding the worst solution, the (equivalent) global optimization problem is intractable (NP-hard). Furthermore, until a minimum or maximum principle is not demonstrated, it is questionable that the effort expended in approximating this minimum is justified. XIV In chapter 5, it is proposed find the frequency distribution of the load factor, for all possible solutions of the onset of collapse, on a simple example. For this purpose, a Monte Carlo sampling of solutions is performed using a contrast method "exact computation of polytopes". The ultimate goal is to determine to which extent the search of the global minimum is justified, and to propose an alternative approach to safety assessment based on probabilities. The frequency distributions for the case study show that both the maximum and the minimum load factors are very infrequent, especially when the contact gets more perfect and more continuous. The results indicates the interest of developing new probabilistic methods. In Chapter 6, is proposed a method based on multiple solutions obtained from random starting points, and qualifying the results through Order Statistics. The purpose is to determine the probability for each solution of the onset of collapse. The method is applied (according to expectations reduction given by the Ordinal Optimization) to obtain a solution that is in a certain percentage of the worst. Finally, in Chapter 7, hybrid methods incorporating metaheuristics are proposed for cases in which the search for the global minimum is justified.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

This paper describes a fully automatic simultaneous lung vessel and airway enhancement filter. The approach consists of a Frangi-based multiscale vessel enhancement filtering specifically designed for lung vessel and airway detection, where arteries and veins have high contrast with respect to the lung parenchyma, and airway walls are hollow tubular structures with a non negative response using the classical Frangi's filter. The features extracted from the Hessian matrix are used to detect centerlines and approximate walls of airways, decreasing the filter response in those areas by applying a penalty function to the vesselness measure. We validate the segmentation method in 20 CT scans with different pathological states within the VESSEL12 challenge framework. Results indicate that our approach obtains good results, decreasing the number of false positives in airway walls.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

The algorithms and graphic user interface software package ?OPT-PROx? are developed to meet food engineering needs related to canned food thermal processing simulation and optimization. The adaptive random search algorithm and its modification coupled with penalty function?s approach, and the finite difference methods with cubic spline approximation are utilized by ?OPT-PROx? package (http://tomakechoice. com/optprox/index.html). The diversity of thermal food processing optimization problems with different objectives and required constraints are solvable by developed software. The geometries supported by the ?OPT-PROx? are the following: (1) cylinder, (2) rectangle, (3) sphere. The mean square error minimization principle is utilized in order to estimate the heat transfer coefficient of food to be heated under optimal condition. The developed user friendly dialogue and used numerical procedures makes the ?OPT-PROx? software useful to food scientists in research and education, as well as to engineers involved in optimization of thermal food processing.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

El objetivo de este trabajo es la elaboración de un procedimiento para la medida del coeficiente de absorción sonora normal en un tubo de impedancia. Para ello se han estudiado los fundamentos básicos de la ecuación de ondas y sus soluciones. Se han considerado las soluciones pertinentes que describen el comportamiento de una onda sonora dentro de un tubo rígido. Se ha considerado también la teoría básica de funciones de transferencia. Estas teorías son claves a la hora de poder desarrollar el procedimiento de medida, ya que el coeficiente de absorción acústica se obtendrá con la ayuda de un tubo de impedancias que mide las funciones de transferencia entre dos posiciones de micrófonos incorporados en una de las caras del tubo. La utilización de esta técnica tiene como principal ventaja, la necesidad de poco espacio en un laboratorio y el empleo de muestras pequeñas de material. La implementación de los visto teóricamente a su aplicación práctica se ha hecho a través de un procedimiento de medida que sigue la Norma UNE-EN ISO 10534-2 (2002) “Determinación del coeficiente de absorción sonoro y la impedancia en tubos de impedancia Parte 2: método función de transferencia”. El valor del coeficiente de absorción se puede obtener a través de una instrumentación específica y un programa computador. Para poder validar los cálculos que realiza el programa utilizado, se ha realizado una batería de medidas del coeficiente de absorción a diferentes tipos de materiales acústicos, y los cálculos se han hecho por la vía del programa y por la vía de una hoja de cálculo. Como parte del procedimiento de medida se ha calculado la incertidumbre en las medidas. En definitiva se pretende contribuir con este trabajo a establecer un procedimiento de medida del comportamiento acústico de diversos materiales. SUMMARY. The aim of this work is the development of a procedure for measuring the sound absorption coefficient normal of an impedance tube. To this end we have studied the basics of the wave equation and its solutions. We have considered the relevant solutions that describe the behavior of a sound wave in a rigid tube. It has also considered the basic theory of transfer functions. These theories are key when we want to develop the measurement method, since the absorption coefficient is obtained with the aid of an impedance tube measuring transfer functions between two positions of microphones incorporated into one side of the tube. The use of this technique has the main advantage, the need of little space on a laboratory and use of small samples of material. The implementation of theoretically seen to his practical application has been made through a measurement procedure following the UNE-EN ISO 10534-2 (2002) "Determination of sound absorption coefficient and impedance in impedance tubes Part 2 : transfer function method ". The value of the absorption coefficient can be obtained through a specific instrumentation and computer software. In order to validate the calculations performed by the program used, there has been realized a series of measures of the absorption coefficient at different types of acoustical materials, and calculations were made by means of the program and by means of a spreadsheet. As part of the measurement procedure has been estimated uncertainty in the measurements. Ultimately it’s tried to contribute with this work to establish a procedure measuring the acoustic behavior of various materials.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

El objetivo del presente Trabajo Fin de Máster es estudiar la precisión de los resultados obtenidos en la medición del coeficiente de absorción acústica en tubo de impedancia, según la norma UNE-EN ISO 10534-2 “Determinación del coeficiente de absorción acústica y de la impedancia acústica en tubos de impedancia. Parte 2: Método de la función de transferencia”. En primer lugar, se han estudiado los fundamentos teóricos relacionados con el método de ensayo. A continuación, se ha detallado dicho método y se ha aplicado a un caso práctico con la instrumentación disponible en el laboratorio de Acústica de la Escuela. En relación a la precisión del método se ha analizado si la preparación e instalación de la muestra son causas de imprecisión. Para ello, se han realizado varios ensayos con dos tipos de materiales acústicos, con el fin de estudiar la dispersión entre los resultados que produce tanto el corte de la muestra, realizado en el proceso de confección, como su colocación en el tubo de impedancia. Además, se ha estudiado si desviaciones en la medida de la temperatura y de la distancia entre los micrófonos influyen en los valores del coeficiente de absorción acústica medido y de su incertidumbre asociada. Puesto que el resultado de un ensayo únicamente se halla completo cuando está acompañado de una declaración acerca de la incertidumbre de dicho resultado, en el presente trabajo se ha aplicado a este método de ensayo un procedimiento para estimar la incertidumbre empleando el método de Monte Carlo. ABSTRACT The objective of this project is studying the precision of the measurements of sound absorption coefficient in impedance tube, according to standard UNE-EN ISO 10534-2 “Determination of sound absorption coefficient and impedance in impedance tubes. Part 2: Transfer-function method”. Firstly, theoretical basis related to the test method have been studied. Furthermore, this method has been defined and applied to a particular case with the instrumentation available in the Acoustics laboratory of the College. In relation to the precision of the method, the preparation and installation of the test sample have been analyzed as possible causes of imprecision. For this purpose, two types of acoustic materials have been tested in order to study the deviation between the results produced by the cutting of the test sample and the collocation in impedance tube. In addition, it has been studied if deviations in the measurement of the temperature and distance between microphones may influence the sound absorption coefficient measured and its associated uncertainty. The test result must be accompanied by a statement of the uncertainty of the result. For this reason, in this project a procedure for estimating uncertainty of the result of this test method has been applied using the Monte Carlo method.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

The paper proposes a new application of non-parametric statistical processing of signals recorded from vibration tests for damage detection and evaluation on I-section steel segments. The steel segments investigated constitute the energy dissipating part of a new type of hysteretic damper that is used for passive control of buildings and civil engineering structures subjected to earthquake-type dynamic loadings. Two I-section steel segments with different levels of damage were instrumented with piezoceramic sensors and subjected to controlled white noise random vibrations. The signals recorded during the tests were processed using two non-parametric methods (the power spectral density method and the frequency response function method) that had never previously been applied to hysteretic dampers. The appropriateness of these methods for quantifying the level of damage on the I-shape steel segments is validated experimentally. Based on the results of the random vibrations, the paper proposes a new index that predicts the level of damage and the proximity of failure of the hysteretic damper

Relevância:

40.00% 40.00%

Publicador:

Resumo:

The decision to build the Secondary Lining in Route Tunnels which are perforated using the New Austrian Tunnelling Method does not generally depend on reasons of structural strength. The paper discusses the implications of the issue as well as the pros and cons of the current alternative courses of action.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The research work as presented in this article covers the design of detached breakwaters since they constitute a type of coastal defence work with which to combat many of the erosion problems found on beaches in a stable, sustainable fashion. The main aim of this work is to formulate a functional and environmental (but not structural) design method, enabling the fundamental characteristics of a detached breakwater to be defined as a function of the effect it is wished to induce on the coast, and taking into account variables of a different nature (climate, geomorphology and geometry) influencing the changes the shoreline undergoes after its construction. With this article, it is intended to submit the final result of the investigation undertaken, applying the detached breakwater design method as developed to solving a practical case. Thus it may be shown how the method enables a detached breakwater’s geometric pre-sizing to be tackled at a place on the coast with certain climate, geomorphology and littoral dynamic characteristics, first setting the final state of equilibrium it is wanted to obtain therein after its construction.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The penalty corner is one of the most important game situations in field hockey with one third of all goals resulting from this tactical situation. The aim of this study was to develop and apply a training method, based on previous studies, to improve the drag- flick skill on a young top-class field hockey player. A young top-class player exercised three times per week using specific drills over a four week period. A VICON optoelectronic system (Oxford Metrics, Oxford, UK) was employed to capture twenty drag-flicks, with six cameras sampling at 250 Hz, prior and after the training period. In order to analyze pre- and post-test differences a dependent t-test was carried out. Angular velocities and the kinematic sequence were similar to previous studies. The player improved (albeit not significantly) the angular velocity of the stick. The player increased front foot to the ball at T1 (p < 0.01) and the drag-flick distances. The range of motion from the front leg decreased from T1 to T6 after the training period (p < 0.01). The specific training sessions conducted with the player improved some features of this particular skill. This article shows how technical knowledge can help with the design of training programs and whether some drills are more effective than others.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Eutectic rods of Al2O3–Er3Al5O12 were grown by directional solidification using the laser-heated floating zone method at rates in the range 25–1500 mm/h. Their microstructure and mechanical properties (hardness, toughness and strength) were investigated as a function of the growth rate. A homogeneous and interpenetrated microstructure was found in most cases, and interphase spacing decreased with growth rate following the Hunt–Jackson law. Hardness increased slightly as the interphase spacing decreased while toughness was low and independent of the microstructure. The rods presented very high bending strength as a result of the homogeneous microstructure, and their strength increased rapidly as the interphase spacing decreased, reaching a maximum of 2.7 GPa for the rods grown at 750 mm/h. The bending strength remained constant up to 1300 K and decreased above this temperature. The relationship between the microstructure and the mechanical properties was established from the analysis of the microstructure and of the fracture mechanisms

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Se presenta un nuevo método de diseño conceptual en Ingeniería Aeronáutica basado el uso de modelos reducidos, también llamados modelos sustitutos (‘surrogates’). Los ingredientes de la función objetivo se calculan para cada indiviudo mediante la utilización de modelos sustitutos asociados a las distintas disciplinas técnicas que se construyen mediante definiciones de descomposición en valores singulares de alto orden (HOSVD) e interpolaciones unidimensionales. Estos modelos sustitutos se obtienen a partir de un número limitado de cálculos CFD. Los modelos sustitutos pueden combinarse, bien con un método de optimización global de tipo algoritmo genético, o con un método local de tipo gradiente. El método resultate es flexible a la par que mucho más eficiente, computacionalmente hablando, que los modelos convencionales basados en el cálculo directo de la función objetivo, especialmente si aparecen un gran número de parámetros de diseño y/o de modelado. El método se ilustra considerando una versión simplificada del diseño conceptual de un avión. Abstract An optimization method for conceptual design in Aeronautics is presented that is based on the use of surrogate models. The various ingredients in the target function are calculated for each individual using surrogates of the associated technical disciplines that are constructed via high order singular value decomposition and one dimensional interpolation. These surrogates result from a limited number of CFD calculated snapshots. The surrogates are combined with an optimization method, which can be either a global optimization method such as a genetic algorithm or a local optimization method, such as a gradient-like method. The resulting method is both flexible and much more computationally efficient than the conventional method based on direct calculation of the target function, especially if a large number of free design parameters and/or tunablemodeling parameters are present. The method is illustrated considering a simplified version of the conceptual design of an aircraft empennage.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

A Digital Elevation Model (DEM) provides the information basis used for many geographic applications such as topographic and geomorphologic studies, landscape through GIS (Geographic Information Systems) among others. The DEM capacity to represent Earth?s surface depends on the surface roughness and the resolution used. Each DEM pixel depends on the scale used characterized by two variables: resolution and extension of the area studied. DEMs can vary in resolution and accuracy by the production method, although there are statistical characteristics that keep constant or very similar in a wide range of scales. Based on this property, several techniques have been applied to characterize DEM through multiscale analysis directly related to fractal geometry: multifractal spectrum and the structure function. The comparison of the results by both methods is discussed. The study area is represented by a 1024 x 1024 data matrix obtained from a DEM with a resolution of 10 x 10 m each point, which correspond with a region known as ?Monte de El Pardo? a property of Spanish National Heritage (Patrimonio Nacional Español) of 15820 Ha located to a short distance from the center of Madrid. Manzanares River goes through this area from North to South. In the southern area a reservoir is found with a capacity of 43 hm3, with an altitude of 603.3 m till 632 m when it is at the highest capacity. In the middle of the reservoir the minimum altitude of this area is achieved.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The Simultaneous Multiple Surface (SMS) method in planar geometry (2D) is applied to imaging designs, generating lenses that compare well with aplanatic designs. When the merit function utilizes image quality over the entire field (not just paraxial), the SMS strategy is superior. In fact, the traditional aplanatic approach is actually a particular case of the SMS strategy

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The Direct Boundary Element Method (DBEM) is presented to solve the elastodynamic field equations in 2D, and a complete comprehensive implementation is given. The DBEM is a useful approach to obtain reliable numerical estimates of site effects on seismic ground motion due to irregular geological configurations, both of layering and topography. The method is based on the discretization of the classical Somigliana's elastodynamic representation equation which stems from the reciprocity theorem. This equation is given in terms of the Green's function which is the full-space harmonic steady-state fundamental solution. The formulation permits the treatment of viscoelastic media, therefore site models with intrinsic attenuation can be examined. By means of this approach, the calculation of 2D scattering of seismic waves, due to the incidence of P and SV waves on irregular topographical profiles is performed. Sites such as, canyons, mountains and valleys in irregular multilayered media are computed to test the technique. The obtained transfer functions show excellent agreement with already published results.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

After the experience gained during the past years it seems clear that nonlinear analysis of bridges are very important to compute ductility demands and to localize potential hinges. This is specially true for irregular bridges in which it is not clear weather or not it is possible to use a linear computation followed by a correction using a behaviour factor. To simplify the numerical effort several approximate methods have been proposed. Among them, the so-called Dynamic Plastic Hinge Method in which an evolutionary shape function is used to reduce the structure to a single degree of freedom system seems to mantein a good balance between accuracy and simplicity. This paper presents results obtained in a parametric study conducted under the auspicies of PREC-8 european research program.