12 resultados para Bézier curves
em Universidad Politécnica de Madrid
Resumo:
An aerodynamic optimization of the train aerodynamic characteristics in term of front wind action sensitivity is carried out in this paper. In particular, a genetic algorithm (GA) is used to perform a shape optimization study of a high-speed train nose. The nose is parametrically defined via Bézier Curves, including a wider range of geometries in the design space as possible optimal solutions. Using a GA, the main disadvantage to deal with is the large number of evaluations need before finding such optimal. Here it is proposed the use of metamodels to replace Navier-Stokes solver. Among all the posibilities, Rsponse Surface Models and Artificial Neural Networks (ANN) are considered. Best results of prediction and generalization are obtained with ANN and those are applied in GA code. The paper shows the feasibility of using GA in combination with ANN for this problem, and solutions achieved are included.
Resumo:
Derive coordinate-free expressions for geometric characteristics of conics written in Bézier form in terms of their control points and weights.
Resumo:
El presente trabajo propone un método para la determinación de los valores de las tolerancias individuales de las piezas que forman un conjunto ensamblado a partir de valores de tolerancias especificados en el conjunto final, optimizando el coste total de fabricación de las piezas individuales a partir de funciones de coste-tolerancia basadas en el proceso de fabricación de cada una de ellas. Para ello se parte de los principales trabajos desarrollados en la línea de asignación de tolerancias y se realiza la propuesta del modelo de trabajo, basado en la optimización de costes a partir de la aplicación del método de los multiplicadores de Lagrange a diversas curvas de coste-tolerancia
Resumo:
Decreasing the accidents on highway and urban environments is the main motivation for the research and developing of driving assistance systems, also called ADAS (Advanced Driver Assistance Systems). In recent years, there are many applications of these systems in commercial vehicles: ABS systems, Cruise Control (CC), parking assistance and warning systems (including GPS), among others. However, the implementation of driving assistance systems on the steering wheel is more limited, because of their complexity and sensitivity. This paper is focused in the development, test and implementation of a driver assistance system for controlling the steering wheel in curve zones. This system is divided in two levels: an inner control loop which permits to execute the position and speed target, softening the action over the steering wheel, and a second control outer loop (controlling for fuzzy logic) that sends the reference to the inner loop according the environment and vehicle conditions. The tests have been done in different curves and speeds. The system has been proved in a commercial vehicle with satisfactory results.
Resumo:
In this talk we show a construction for characterising developable surfaces in the form of Bézier triangular patches. It is shown that constructions used for rectangular patches are not useful, since they provide degenerate triangular patches. Explicit constructions of non-degenerate developable triangular patches are provided.
Resumo:
Purpose: Surgical simulators are currently essential within any laparoscopic training program because they provide a low-stakes, reproducible and reliable environment to acquire basic skills. The purpose of this study is to determine the training learning curve based on different metrics corresponding to five tasks included in SINERGIA laparoscopic virtual reality simulator. Methods: Thirty medical students without surgical experience participated in the study. Five tasks of SINERGIA were included: Coordination, Navigation, Navigation and touch, Accurate grasping and Coordinated pulling. Each participant was trained in SINERGIA. This training consisted of eight sessions (R1–R8) of the five mentioned tasks and was carried out in two consecutive days with four sessions per day. A statistical analysis was made, and the results of R1, R4 and R8 were pair-wise compared with Wilcoxon signed-rank test. Significance is considered at P value <0.005. Results: In total, 84.38% of the metrics provided by SINERGIA and included in this study show significant differences when comparing R1 and R8. Metrics are mostly improved in the first session of training (75.00% when R1 and R4 are compared vs. 37.50% when R4 and R8 are compared). In tasks Coordination and Navigation and touch, all metrics are improved. On the other hand, Navigation just improves 60% of the analyzed metrics. Most learning curves show an improvement with better results in the fulfillment of the different tasks. Conclusions: Learning curves of metrics that assess the basic psychomotor laparoscopic skills acquired in SINERGIA virtual reality simulator show a faster learning rate during the first part of the training. Nevertheless, eight repetitions of the tasks are not enough to acquire all psychomotor skills that can be trained in SINERGIA. Therefore, and based on these results together with previous works, SINERGIA could be used as training tool with a properly designed training program.
Resumo:
En numerosas ocasiones a lo largo de la historia la imaginación de los creadores ha ido por delante de las posibilidades técnicas de cada momento. Así, muchas de estas nuevas ideas han requerido largos periodos de tiempo para materializarse como realidad construida, hasta que el desarrollo tecnológico e industrial hubo alcanzado un grado de madurez suficiente. En el campo de la arquitectura, estas limitaciones técnicas se han ido acotando paulatinamente hasta desembocar en la situación actual en la que cualquier planteamiento formal puede ser representado gráficamente y analizado desde un punto de vista estructural, superádose de este modo la barrera existente históricamente en el tratamiento de las formas. A lo largo del presente tesis doctoral se analiza cómo la formulación del Método de los Elementos Finitos en la década de los cincuenta y las curvas de Bézier en la década de los sesenta del siglo pasado y la posterior generalización de los ordenadores personales y de los programas informáticos asociados (C.A.D. y F.E.M. principalmente) en los estudios de arquitectura e ingeniería a partir de la década de los noventa, posibilitó el desarrollo de cualquier propuesta arquitectónica, por compleja que ésta fuese, provocando una verdadera revolución a nivel formal en el mundo de la arquitectura, especialmente en el campo de la edificación singular o icónica. Se estudia este proceso a través de ocho edificios; cuatro anteriores y otros tantos posteriores a la desaparición de la barrera anteriormente referida, establecida de forma simbólica en la década de los años ochenta del siglo XX: Frontón de Recoletos en Madrid, Edificio Seagram en Nueva York, Habitat ’67 en Montreal, Ópera de Sídney, museo Guggenheim de Bilbao, ampliación del Victoria & Albert Museum en Londres, tanatorio “Meiso no Mori” en Gifu y nueva sede de la CCTV en Pekín. De entre ellos, la Ópera de Sídney, obra del arquitecto danés Jørn Utzon, condensa gran parte de los aspectos relevantes investigados en relación a la influencia que los métodos de representación y análisis estructural ejercen en la concepción y construcción de las obras de arquitectura. Por este motivo y por considerarse un hito de la arquitectura a nivel global se toma como caso de estudio. La idea general del edificio, que data de 1956, se enmarca en una época inmediatamente anterior a la del desarrollo científico y tecnológico anteriormente referido. Esta ausencia de herramientas de diseño disponibles acordes a la complejidad formal de la propuesta planteada condicionó enormente la marcha del proyecto, dilatándose dramáticamente en el tiempo y disparándose su coste hasta el punto de que el propio arquitecto danés fue separado de las obras antes de su conclusión. Además, la solución estructural finalmente construida de las cubiertas dista mucho de la prevista por Utzon inicialmente. Donde él había imaginado unas finas láminas de hormigón flotando sobre el paisaje se materializó una estructura más pesada, formada por costillas pretensadas de hormigón con unas secciones notablemente mayores. La forma también debió ser modificada de modo ostensible respecto a la propuesta inicial. Si este edificio se pretendiese construir en la actualidad, con toda seguridad el curso de los acontecimientos se desarrollaría por senderos muy diferentes. Ante este supuesto, se plantean las siguientes cuestiones: ¿sería posible realizar un análisis estructural de la cubierta laminar planteada por Utzon inicialmente en el concurso con las herramientas disponibles en la actualidad?; ¿sería dicha propuesta viable estructuralmente?. A lo largo de las siguientes páginas se pretende dar respuesta a estas cuestiones, poniendo de relieve el impacto que los ordenadores personales y los programas informáticos asociados han tenido en la manera de concebir y construir edificios. También se han analizado variantes a la solución laminar planteada en la fase de concurso, a través de las cuales, tratando en la medida de lo posible de ajustarse a las sugerencias que Ove Arup y su equipo realizaron a Jørn Utzon a lo largo del dilatado proceso de proyecto, mejorar el comportamiento general de la estructura del edificio. Por último, se ha pretendido partir de cero y plantear, desde una perspectiva contemporánea, posibles enfoques metodológicos aplicables a la búsqueda de soluciones estructurales compatibles con la forma propuesta originalmente por Utzon para las cubiertas de la Ópera de Sídney y que nunca llegó a ser construida (ni analizada), considerando para ello los medios tecnológicos, científicos e industriales disponibles en la actualidad. Abstract On numerous occasions throughout history the imagination of creators has gone well beyond of the technical possibilities of their time. Many new ideas have required a long period to materialize, until the technological and industrial development had time to catch up. In the architecture field, these technical limitations have gradually tightened leading to the current situation in which any formal approach can be represented and analyzed from a structural point of view, thus concluding that the structural analysis and the graphical representation’s barrier in the development of architectural projects has dissappeared. Throughout the following pages it is examined how the development of the Finite Element Method in the fifties and the Bezier curves in the sixties of the last century and the subsequent spread of personal computers and specialized software in the architectural and engineering offices from the nineties, enabled the development of any architectural proposal independently of its complexity. This has caused a revolution at a formal level in architecture, especially in the field of iconic building. This process is analyzed through eight buildings, four of them before and another four after the disappearance of the above mentioned barrier, roughly established in the eighties of the last century: Fronton Recoletos in Madrid, Seagram Building in New York Habitat '67 in Montreal, Sydney Opera House, Guggenheim Museum Bilbao, Victoria & Albert Museum extension in London, Crematorium “Meiso no Mori” in Gifu and the new CCTV headquarters in Beijing. Among them, the Sydney Opera House, designed by Danish architect Jørn Utzon, condenses many of the main aspects previously investigated regarding the impact of representation methods and structural analysis on the design and construction of architectural projects. For this reason and also because it is considered a global architecture milestone, it is selected as a case study. The building’s general idea, which dates from 1956, is framed at a time immediately preceding the above mentioned scientific and technological development. This lack of available design tools in accordance with the proposal’s formal complexity conditioned enormously the project’s progress, leading to a dramatic delay and multiplying the final budget disproportionately to the point that the Danish architect himself was separated from the works before completion. Furthermore, the built structure differs dramatically from the architect’s initial vision. Where Utzon saw a thin concrete shell floating over the landscape a heavier structure was built, consisting of prestressed concrete ribs with a significantly greater size. The geometry also had to be modified. If this building were to built today, the course of events surely would walk very different paths. Given this assumption, a number of questions could then be formulated: Would it be possible to perform a structural analysis of Utzon’s initially proposed competition-free-ways roof’s geometry with the tools available nowadays?; Would this proposal be structurally feasable?. Throughout the following pages it is intended to clarify this issues, highlighting personal computers and associated software’s impact in building design and construction procedures, especially in the field of iconic building. Variants have also been analyzed for the laminar solution proposed in the competition phase, through which, trying as far as possible to comply with the suggestions that Ove Arup and his team did to Jørn Utzon along the lengthy process project, improving the overall performance of the building structure. Finally, we have started from scratch and analyzed, from a contemporary perspective, possible structural solutions compatible with Utzon’s Opera House’s original geometry and vision –proposal that was never built (nor even analyzed)-, taking into consideration the technological, scientific and industrial means currently available.
Resumo:
Synthetic derivation of closed formulae of the geometric characteristic of a conic given in Bézier form in terms of its control polygon, (P; Q; R) and weights, (1; w; 1g)
Resumo:
ABSTRACT: The comparison of the different bids in the tender for a project, with the traditional contract system based on unit rates open to and re-measurement, requires analysis tools that are able to discriminate proposals having a similar overall economic impact, but that might show a very different behaviour during the execution of the works. RESUMEN: La estimación rápida de costes en fases iniciales del proyecto por métodos paramétricos y referencias estadísticas es un tema bien estudiado, divulgado y aplicado en el sector de la construcción. Sin embargo, existe poca literatura técnica sobre sistemas de predimensionado de tiempos, que permitan realizar rápidamente una planificación con un grado de aproximación razonable. Este texto reúne dos aspectos ya conocidos, pero hasta ahora independientes, y una aportación propia: -La estimación del plazo final por referencias estadísticas (BCIS, 2000) - La estimación del reparto del coste total a lo largo de la ejecución mediante curvas "S" (diversos autores) La estimación de la duración de la ejecución de las actividades en función de su coste. El conjunto de estas tres técnicas, aplicadas a un proyecto, permite obtener una planificación con el suficiente grado de detalle y fiabilidad para tomar decisiones en fases iniciales del proyecto.
Resumo:
The problem of parameterizing approximately algebraic curves and surfaces is an active research field, with many implications in practical applications. The problem can be treated locally or globally. We formally state the problem, in its global version for the case of algebraic curves (planar or spatial), and we report on some algorithms approaching it, as well as on the associated error distance analysis.
Resumo:
The purpose of this study was to compare a number of state-of-the-art methods in airborne laser scan- ning (ALS) remote sensing with regards to their capacity to describe tree size inequality and other indi- cators related to forest structure. The indicators chosen were based on the analysis of the Lorenz curve: Gini coefficient ( GC ), Lorenz asymmetry ( LA ), the proportions of basal area ( BALM ) and stem density ( NSLM ) stocked above the mean quadratic diameter. Each method belonged to one of these estimation strategies: (A) estimating indicators directly; (B) estimating the whole Lorenz curve; or (C) estimating a complete tree list. Across these strategies, the most popular statistical methods for area-based approach (ABA) were used: regression, random forest (RF), and nearest neighbour imputation. The latter included distance metrics based on either RF (NN–RF) or most similar neighbour (MSN). In the case of tree list esti- mation, methods based on individual tree detection (ITD) and semi-ITD, both combined with MSN impu- tation, were also studied. The most accurate method was direct estimation by best subset regression, which obtained the lowest cross-validated coefficients of variation of their root mean squared error CV(RMSE) for most indicators: GC (16.80%), LA (8.76%), BALM (8.80%) and NSLM (14.60%). Similar figures [CV(RMSE) 16.09%, 10.49%, 10.93% and 14.07%, respectively] were obtained by MSN imputation of tree lists by ABA, a method that also showed a number of additional advantages, such as better distributing the residual variance along the predictive range. In light of our results, ITD approaches may be clearly inferior to ABA with regards to describing the structural properties related to tree size inequality in for- ested areas.
Resumo:
When applying computational mathematics in practical applications, even though one may be dealing with a problem that can be solved algorithmically, and even though one has good algorithms to approach the solution, it can happen, and often it is the case, that the problem has to be reformulated and analyzed from a different computational point of view. This is the case of the development of approximate algorithms. This paper frames in the research area of approximate algebraic geometry and commutative algebra and, more precisely, on the problem of the approximate parametrization.