17 resultados para ORDER ACCURACY APPROXIMATIONS

em Universidad Politécnica de Madrid


Relevância:

30.00% 30.00%

Publicador:

Resumo:

This thesis contributes to the analysis and design of printed reflectarray antennas. The main part of the work is focused on the analysis of dual offset antennas comprising two reflectarray surfaces, one of them acts as sub-reflector and the second one acts as mainreflector. These configurations introduce additional complexity in several aspects respect to conventional dual offset reflectors, however they present a lot of degrees of freedom that can be used to improve the electrical performance of the antenna. The thesis is organized in four parts: the development of an analysis technique for dualreflectarray antennas, a preliminary validation of such methodology using equivalent reflector systems as reference antennas, a more rigorous validation of the software tool by manufacturing and testing a dual-reflectarray antenna demonstrator and the practical design of dual-reflectarray systems for some applications that show the potential of these kind of configurations to scan the beam and to generate contoured beams. In the first part, a general tool has been implemented to analyze high gain antennas which are constructed of two flat reflectarray structures. The classic reflectarray analysis based on MoM under local periodicity assumption is used for both sub and main reflectarrays, taking into account the incident angle on each reflectarray element. The incident field on the main reflectarray is computed taking into account the field radiated by all the elements on the sub-reflectarray.. Two approaches have been developed, one which employs a simple approximation to reduce the computer run time, and the other which does not, but offers in many cases, improved accuracy. The approximation is based on computing the reflected field on each element on the main reflectarray only once for all the fields radiated by the sub-reflectarray elements, assuming that the response will be the same because the only difference is a small variation on the angle of incidence. This approximation is very accurate when the reflectarray elements on the main reflectarray show a relatively small sensitivity to the angle of incidence. An extension of the analysis technique has been implemented to study dual-reflectarray antennas comprising a main reflectarray printed on a parabolic surface, or in general in a curved surface. In many applications of dual-reflectarray configurations, the reflectarray elements are in the near field of the feed-horn. To consider the near field radiated by the horn, the incident field on each reflectarray element is computed using a spherical mode expansion. In this region, the angles of incidence are moderately wide, and they are considered in the analysis of the reflectarray to better calculate the actual incident field on the sub-reflectarray elements. This technique increases the accuracy for the prediction of co- and cross-polar patterns and antenna gain respect to the case of using ideal feed models. In the second part, as a preliminary validation, the proposed analysis method has been used to design a dual-reflectarray antenna that emulates previous dual-reflector antennas in Ku and W-bands including a reflectarray as subreflector. The results for the dualreflectarray antenna compare very well with those of the parabolic reflector and reflectarray subreflector; radiation patterns, antenna gain and efficiency are practically the same when the main parabolic reflector is substituted by a flat reflectarray. The results show that the gain is only reduced by a few tenths of a dB as a result of the ohmic losses in the reflectarray. The phase adjustment on two surfaces provided by the dual-reflectarray configuration can be used to improve the antenna performance in some applications requiring multiple beams, beam scanning or shaped beams. Third, a very challenging dual-reflectarray antenna demonstrator has been designed, manufactured and tested for a more rigorous validation of the analysis technique presented. The proposed antenna configuration has the feed, the sub-reflectarray and the main-reflectarray in the near field one to each other, so that the conventional far field approximations are not suitable for the analysis of such antenna. This geometry is used as benchmarking for the proposed analysis tool in very stringent conditions. Some aspects of the proposed analysis technique that allow improving the accuracy of the analysis are also discussed. These improvements include a novel method to reduce the inherent cross polarization which is introduced mainly from grounded patch arrays. It has been checked that cross polarization in offset reflectarrays can be significantly reduced by properly adjusting the patch dimensions in the reflectarray in order to produce an overall cancellation of the cross-polarization. The dimensions of the patches are adjusted in order not only to provide the required phase-distribution to shape the beam, but also to exploit the crosses by zero of the cross-polarization components. The last part of the thesis deals with direct applications of the technique described. The technique presented is directly applicable to the design of contoured beam antennas for DBS applications, where the requirements of cross-polarisation are very stringent. The beam shaping is achieved by synthesithing the phase distribution on the main reflectarray while the sub-reflectarray emulates an equivalent hyperbolic subreflector. Dual-reflectarray antennas present also the ability to scan the beam over small angles about boresight. Two possible architectures for a Ku-band antenna are also described based on a dual planar reflectarray configuration that provides electronic beam scanning in a limited angular range. In the first architecture, the beam scanning is achieved by introducing a phase-control in the elements of the sub-reflectarray and the mainreflectarray is passive. A second alternative is also studied, in which the beam scanning is produced using 1-bit control on the main reflectarray, while a passive subreflectarray is designed to provide a large focal distance within a compact configuration. The system aims to develop a solution for bi-directional satellite links for emergency communications. In both proposed architectures, the objective is to provide a compact optics and simplicity to be folded and deployed.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Some basic ideas are presented for the construction of robust, computationally efficient reduced order models amenable to be used in industrial environments, combined with somewhat rough computational fluid dynamics solvers. These ideas result from a critical review of the basic principles of proper orthogonal decomposition-based reduced order modeling of both steady and unsteady fluid flows. In particular, the extent to which some artifacts of the computational fluid dynamics solvers can be ignored is addressed, which opens up the possibility of obtaining quite flexible reduced order models. The methods are illustrated with the steady aerodynamic flow around a horizontal tail plane of a commercial aircraft in transonic conditions, and the unsteady lid-driven cavity problem. In both cases, the approximations are fairly good, thus reducing the computational cost by a significant factor.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

A method is presented to construct computationally efficient reduced-order models (ROMs) of three-dimensional aerodynamic flows around commercial aircraft components. The method is based on the proper orthogonal decomposition (POD) of a set of steady snapshots, which are calculated using an industrial solver based on some Reynolds averaged Navier-Stokes (RANS) equations. The POD-mode amplitudes are calculated by minimizing a residual defined from the Euler equations, even though the snapshots themselves are calculated from viscous equations. This makes the ROM independent of the peculiarities of the solver used to calculate the snapshots. Also, both the POD modes and the residual are calculated using points in the computational mesh that are concentrated in a close vicinity of the aircraft, which constitute a much smaller number than the total number of mesh points. Despite these simplifications, the method provides quite good approximations of the flow variables distributions in the whole computational domain, including the boundary layer attached to the aircraft surface and the wake. Thus, the method is both robust and computationally efficient, which is checked considering the aerodynamic flow around a horizontal tail plane, in the transonic range 0.4?Mach number?0.8, ?3°?angle of attack?3°.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

A unified solution framework is presented for one-, two- or three-dimensional complex non-symmetric eigenvalue problems, respectively governing linear modal instability of incompressible fluid flows in rectangular domains having two, one or no homogeneous spatial directions. The solution algorithm is based on subspace iteration in which the spatial discretization matrix is formed, stored and inverted serially. Results delivered by spectral collocation based on the Chebyshev-Gauss-Lobatto (CGL) points and a suite of high-order finite-difference methods comprising the previously employed for this type of work Dispersion-Relation-Preserving (DRP) and Padé finite-difference schemes, as well as the Summationby- parts (SBP) and the new high-order finite-difference scheme of order q (FD-q) have been compared from the point of view of accuracy and efficiency in standard validation cases of temporal local and BiGlobal linear instability. The FD-q method has been found to significantly outperform all other finite difference schemes in solving classic linear local, BiGlobal, and TriGlobal eigenvalue problems, as regards both memory and CPU time requirements. Results shown in the present study disprove the paradigm that spectral methods are superior to finite difference methods in terms of computational cost, at equal accuracy, FD-q spatial discretization delivering a speedup of ð (10 4). Consequently, accurate solutions of the three-dimensional (TriGlobal) eigenvalue problems may be solved on typical desktop computers with modest computational effort.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Accuracy in the liquid hydrocarbons custody transfer is mandatory because it has a great economic impact. By far the most accurate meter is the positive displacement (PD) meter. Increasing such an accuracy may adversely affect the cost of the custody transfer, unless simple models are developed in order to lower the cost, which is the purpose of this work. PD meter consists of a fixed volume rotating chamber. For each turn a pulse is counted, hence, the measured volume is the number of pulses times the volume of the chamber. It does not coincide with the real volume, so corrections have to be made. All the corrections are grouped by a meter factor. Among corrections highlights the slippage flow. By solving the Navier-Stokes equations one can find an analytical expression for this flow. It is neither easy nor cheap to apply straightforward the slippage correction; therefore we have made a simple model where slippage is regarded as a single parameter with dimension of time. The model has been tested for several PD meters. In our careful experiments, the meter factor grows with temperature at a constant pace of 8?10?5?ºC?1. Be warned

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The main focus of this paper is on hydrodynamic modelling of a semisubmersible platform (which can support a 1.5MW wind turbine and is composed by three buoyant columns connected by bracings) with especial emphasis on the estimation of the wave drift components and their effects on the design of the mooring system. Indeed, with natural periods of drift around 60 seconds, accurate computation of the low-frequency second-order components is not a straightforward task. As methods usually adopted when dealing with the slow-drifts of deep-water moored systems, such as Newman?s approximation, have their errors increased by the relatively low resonant periods, and as the effects of depth cannot be ignored, the wave diffraction analysis must be based on full Quadratic Transfer Functions (QTF) computations. A discussion on the numerical aspects of performing such computations is presented, making use of the second-order module available with the seakeeping software WAMIT®. Finally, the paper also provides a preliminary verification of the accuracy of the numerical predictions based on the results obtained in a series of model tests with the structure fixed in bichromatic waves.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

La segmentación de imágenes es un campo importante de la visión computacional y una de las áreas de investigación más activas, con aplicaciones en comprensión de imágenes, detección de objetos, reconocimiento facial, vigilancia de vídeo o procesamiento de imagen médica. La segmentación de imágenes es un problema difícil en general, pero especialmente en entornos científicos y biomédicos, donde las técnicas de adquisición imagen proporcionan imágenes ruidosas. Además, en muchos de estos casos se necesita una precisión casi perfecta. En esta tesis, revisamos y comparamos primero algunas de las técnicas ampliamente usadas para la segmentación de imágenes médicas. Estas técnicas usan clasificadores a nivel de pixel e introducen regularización sobre pares de píxeles que es normalmente insuficiente. Estudiamos las dificultades que presentan para capturar la información de alto nivel sobre los objetos a segmentar. Esta deficiencia da lugar a detecciones erróneas, bordes irregulares, configuraciones con topología errónea y formas inválidas. Para solucionar estos problemas, proponemos un nuevo método de regularización de alto nivel que aprende información topológica y de forma a partir de los datos de entrenamiento de una forma no paramétrica usando potenciales de orden superior. Los potenciales de orden superior se están popularizando en visión por computador, pero la representación exacta de un potencial de orden superior definido sobre muchas variables es computacionalmente inviable. Usamos una representación compacta de los potenciales basada en un conjunto finito de patrones aprendidos de los datos de entrenamiento que, a su vez, depende de las observaciones. Gracias a esta representación, los potenciales de orden superior pueden ser convertidos a potenciales de orden 2 con algunas variables auxiliares añadidas. Experimentos con imágenes reales y sintéticas confirman que nuestro modelo soluciona los errores de aproximaciones más débiles. Incluso con una regularización de alto nivel, una precisión exacta es inalcanzable, y se requeire de edición manual de los resultados de la segmentación automática. La edición manual es tediosa y pesada, y cualquier herramienta de ayuda es muy apreciada. Estas herramientas necesitan ser precisas, pero también lo suficientemente rápidas para ser usadas de forma interactiva. Los contornos activos son una buena solución: son buenos para detecciones precisas de fronteras y, en lugar de buscar una solución global, proporcionan un ajuste fino a resultados que ya existían previamente. Sin embargo, requieren una representación implícita que les permita trabajar con cambios topológicos del contorno, y esto da lugar a ecuaciones en derivadas parciales (EDP) que son costosas de resolver computacionalmente y pueden presentar problemas de estabilidad numérica. Presentamos una aproximación morfológica a la evolución de contornos basada en un nuevo operador morfológico de curvatura que es válido para superficies de cualquier dimensión. Aproximamos la solución numérica de la EDP de la evolución de contorno mediante la aplicación sucesiva de un conjunto de operadores morfológicos aplicados sobre una función de conjuntos de nivel. Estos operadores son muy rápidos, no sufren de problemas de estabilidad numérica y no degradan la función de los conjuntos de nivel, de modo que no hay necesidad de reinicializarlo. Además, su implementación es mucho más sencilla que la de las EDP, ya que no requieren usar sofisticados algoritmos numéricos. Desde un punto de vista teórico, profundizamos en las conexiones entre operadores morfológicos y diferenciales, e introducimos nuevos resultados en este área. Validamos nuestra aproximación proporcionando una implementación morfológica de los contornos geodésicos activos, los contornos activos sin bordes, y los turbopíxeles. En los experimentos realizados, las implementaciones morfológicas convergen a soluciones equivalentes a aquéllas logradas mediante soluciones numéricas tradicionales, pero con ganancias significativas en simplicidad, velocidad y estabilidad. ABSTRACT Image segmentation is an important field in computer vision and one of its most active research areas, with applications in image understanding, object detection, face recognition, video surveillance or medical image processing. Image segmentation is a challenging problem in general, but especially in the biological and medical image fields, where the imaging techniques usually produce cluttered and noisy images and near-perfect accuracy is required in many cases. In this thesis we first review and compare some standard techniques widely used for medical image segmentation. These techniques use pixel-wise classifiers and introduce weak pairwise regularization which is insufficient in many cases. We study their difficulties to capture high-level structural information about the objects to segment. This deficiency leads to many erroneous detections, ragged boundaries, incorrect topological configurations and wrong shapes. To deal with these problems, we propose a new regularization method that learns shape and topological information from training data in a nonparametric way using high-order potentials. High-order potentials are becoming increasingly popular in computer vision. However, the exact representation of a general higher order potential defined over many variables is computationally infeasible. We use a compact representation of the potentials based on a finite set of patterns learned fromtraining data that, in turn, depends on the observations. Thanks to this representation, high-order potentials can be converted into pairwise potentials with some added auxiliary variables and minimized with tree-reweighted message passing (TRW) and belief propagation (BP) techniques. Both synthetic and real experiments confirm that our model fixes the errors of weaker approaches. Even with high-level regularization, perfect accuracy is still unattainable, and human editing of the segmentation results is necessary. The manual edition is tedious and cumbersome, and tools that assist the user are greatly appreciated. These tools need to be precise, but also fast enough to be used in real-time. Active contours are a good solution: they are good for precise boundary detection and, instead of finding a global solution, they provide a fine tuning to previously existing results. However, they require an implicit representation to deal with topological changes of the contour, and this leads to PDEs that are computationally costly to solve and may present numerical stability issues. We present a morphological approach to contour evolution based on a new curvature morphological operator valid for surfaces of any dimension. We approximate the numerical solution of the contour evolution PDE by the successive application of a set of morphological operators defined on a binary level-set. These operators are very fast, do not suffer numerical stability issues, and do not degrade the level set function, so there is no need to reinitialize it. Moreover, their implementation is much easier than their PDE counterpart, since they do not require the use of sophisticated numerical algorithms. From a theoretical point of view, we delve into the connections between differential andmorphological operators, and introduce novel results in this area. We validate the approach providing amorphological implementation of the geodesic active contours, the active contours without borders, and turbopixels. In the experiments conducted, the morphological implementations converge to solutions equivalent to those achieved by traditional numerical solutions, but with significant gains in simplicity, speed, and stability.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Location-based services (LBS) highly rely on the location of the mobile user in order to provide the service tailored to that location. This location is calculated differently depending on the technology available in the used mobile device. No matter which technology is used, the location will never be calculated 100% correctly; instead there will always be a margin of error generated during the calculation, which is referred to as positional accuracy. This research has reviewed the eight most common positioning technologies available in the major current smart-phones and assessed their positional accuracy with respect to its usage by LBS applications. Given the vast majority of these applications, this research classified them into thirteen categories, and these categories were also classified depending on their level criticality as low, medium, or high critical, and whether they function indoor or outdoor. The accuracies of different positioning technologies are compared to these two criteria. Low critical outdoor and high critical indoor applications were found technologically covered; high and medium critical outdoor ones weren?t fully resolved. Finally three potential solutions are suggested to be implemented in future smartphones to resolve this technological gap: Real-Time Kinematics Global Positioning System (RTK GPS), terrestrial transmitters, and combination of Wireless Sensors Network and Radio Frequency Identification (WSN-RFID).

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This paper proposes an automatic expert system for accuracy crop row detection in maize fields based on images acquired from a vision system. Different applications in maize, particularly those based on site specific treatments, require the identification of the crop rows. The vision system is designed with a defined geometry and installed onboard a mobile agricultural vehicle, i.e. submitted to vibrations, gyros or uncontrolled movements. Crop rows can be estimated by applying geometrical parameters under image perspective projection. Because of the above undesired effects, most often, the estimation results inaccurate as compared to the real crop rows. The proposed expert system exploits the human knowledge which is mapped into two modules based on image processing techniques. The first one is intended for separating green plants (crops and weeds) from the rest (soil, stones and others). The second one is based on the system geometry where the expected crop lines are mapped onto the image and then a correction is applied through the well-tested and robust Theil–Sen estimator in order to adjust them to the real ones. Its performance is favorably compared against the classical Pearson product–moment correlation coefficient.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Hybrid Stepper Motors are widely used in open-loop position applications. They are the choice of actuation for the collimators in the Large Hadron Collider, the largest particle accelerator at CERN. In this case the positioning requirements and the highly radioactive operating environment are unique. The latter forces both the use of long cables to connect the motors to the drives which act as transmission lines and also prevents the use of standard position sensors. However, reliable and precise operation of the collimators is critical for the machine, requiring the prevention of step loss in the motors and maintenance to be foreseen in case of mechanical degradation. In order to make the above possible, an approach is proposed for the application of an Extended Kalman Filter to a sensorless stepper motor drive, when the motor is separated from its drive by long cables. When the long cables and high frequency pulse width modulated control voltage signals are used together, the electrical signals difer greatly between the motor and drive-side of the cable. Since in the considered case only drive-side data is available, it is therefore necessary to estimate the motor-side signals. Modelling the entire cable and motor system in an Extended Kalman Filter is too computationally intensive for standard embedded real-time platforms. It is, in consequence, proposed to divide the problem into an Extended Kalman Filter, based only on the motor model, and separated motor-side signal estimators, the combination of which is less demanding computationally. The efectiveness of this approach is shown in simulation. Then its validity is experimentally demonstrated via implementation in a DSP based drive. A testbench to test its performance when driving an axis of a Large Hadron Collider collimator is presented along with the results achieved. It is shown that the proposed method is capable of achieving position and load torque estimates which allow step loss to be detected and mechanical degradation to be evaluated without the need for physical sensors. These estimation algorithms often require a precise model of the motor, but the standard electrical model used for hybrid stepper motors is limited when currents, which are high enough to produce saturation of the magnetic circuit, are present. New model extensions are proposed in order to have a more precise model of the motor independently of the current level, whilst maintaining a low computational cost. It is shown that a significant improvement in the model It is achieved with these extensions, and their computational performance is compared to study the cost of model improvement versus computation cost. The applicability of the proposed model extensions is demonstrated via their use in an Extended Kalman Filter running in real-time for closed-loop current control and mechanical state estimation. An additional problem arises from the use of stepper motors. The mechanics of the collimators can wear due to the abrupt motion and torque profiles that are applied by them when used in the standard way, i.e. stepping in open-loop. Closed-loop position control, more specifically Field Oriented Control, would allow smoother profiles, more respectful to the mechanics, to be applied but requires position feedback. As mentioned already, the use of sensors in radioactive environments is very limited for reliability reasons. Sensorless control is a known option but when the speed is very low or zero, as is the case most of the time for the motors used in the LHC collimator, the loss of observability prevents its use. In order to allow the use of position sensors without reducing the long term reliability of the whole system, the possibility to switch from closed to open loop is proposed and validated, allowing the use of closed-loop control when the position sensors function correctly and open-loop when there is a sensor failure. A different approach to deal with the switched drive working with long cables is also presented. Switched mode stepper motor drives tend to have poor performance or even fail completely when the motor is fed through a long cable due to the high oscillations in the drive-side current. The design of a stepper motor output fillter which solves this problem is thus proposed. A two stage filter, one devoted to dealing with the diferential mode and the other with the common mode, is designed and validated experimentally. With this ?lter the drive performance is greatly improved, achieving a positioning repeatability even better than with the drive working without a long cable, the radiated emissions are reduced and the overvoltages at the motor terminals are eliminated.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

El objetivo de la presente investigación es el desarrollo de un modelo de cálculo rápido, eficiente y preciso, para la estimación de los costes finales de construcción, en las fases preliminares del proyecto arquitectónico. Se trata de una herramienta a utilizar durante el proceso de elaboración de estudios previos, anteproyecto y proyecto básico, no siendo por tanto preciso para calcular el “predimensionado de costes” disponer de la total definición grafica y literal del proyecto. Se parte de la hipótesis de que en la aplicación práctica del modelo no se producirán desviaciones superiores al 10 % sobre el coste final de la obra proyectada. Para ello se formulan en el modelo de predimensionado cinco niveles de estimación de costes, de menor a mayor definición conceptual y gráfica del proyecto arquitectónico. Los cinco niveles de cálculo son: dos que toman como referencia los valores “exógenos” de venta de las viviendas (promoción inicial y promoción básica) y tres basados en cálculos de costes “endógenos” de la obra proyectada (estudios previos, anteproyecto y proyecto básico). El primer nivel de estimación de carácter “exógeno” (nivel .1), se calcula en base a la valoración de mercado de la promoción inmobiliaria y a su porcentaje de repercusión de suelo sobre el valor de venta de las viviendas. El quinto nivel de valoración, también de carácter “exógeno” (nivel .5), se calcula a partir del contraste entre el valor externo básico de mercado, los costes de construcción y los gastos de promoción estimados de la obra proyectada. Este contraste entre la “repercusión del coste de construcción” y el valor de mercado, supone una innovación respecto a los modelos de predimensionado de costes existentes, como proceso metodológico de verificación y validación extrínseca, de la precisión y validez de las estimaciones resultantes de la aplicación práctica del modelo, que se denomina Pcr.5n (Predimensionado costes de referencia con .5niveles de cálculo según fase de definición proyectual / ideación arquitectónica). Los otros tres niveles de predimensionado de costes de construcción “endógenos”, se estiman mediante cálculos analíticos internos por unidades de obra y cálculos sintéticos por sistemas constructivos y espacios funcionales, lo que se lleva a cabo en las etapas iniciales del proyecto correspondientes a estudios previos (nivel .2), anteproyecto (nivel .3) y proyecto básico (nivel .4). Estos cálculos teóricos internos son finalmente evaluados y validados mediante la aplicación práctica del modelo en obras de edificación residencial, de las que se conocen sus costes reales de liquidación final de obra. Según va evolucionando y se incrementa el nivel de definición y desarrollo del proyecto, desde los estudios previos hasta el proyecto básico, el cálculo se va perfeccionando en su nivel de eficiencia y precisión de la estimación, según la metodología aplicada: [aproximaciones sucesivas en intervalos finitos], siendo la hipótesis básica como anteriormente se ha avanzado, lograr una desviación máxima de una décima parte en el cálculo estimativo del predimensionado del coste real de obra. El cálculo del coste de ejecución material de la obra, se desarrolla en base a parámetros cúbicos funcionales “tridimensionales” del espacio proyectado y parámetros métricos constructivos “bidimensionales” de la envolvente exterior de cubierta/fachada y de la huella del edificio sobre el terreno. Los costes funcionales y constructivos se ponderan en cada fase del proceso de cálculo con sus parámetros “temáticos/específicos” de gestión (Pg), proyecto (Pp) y ejecución (Pe) de la concreta obra presupuestada, para finalmente estimar el coste de construcción por contrata, como resultado de incrementar al coste de ejecución material el porcentaje correspondiente al parámetro temático/especifico de la obra proyectada. El modelo de predimensionado de costes de construcción Pcr.5n, será una herramienta de gran interés y utilidad en el ámbito profesional, para la estimación del coste correspondiente al Proyecto Básico previsto en el marco técnico y legal de aplicación. Según el Anejo I del Código Técnico de la Edificación (CTE), es de obligado cumplimiento que el proyecto básico contenga una “Valoración aproximada de la ejecución material de la obra proyectada por capítulos”, es decir , que el Proyecto Básico ha de contener al menos un “presupuesto aproximado”, por capítulos, oficios ó tecnologías. El referido cálculo aproximado del presupuesto en el Proyecto Básico, necesariamente se ha de realizar mediante la técnica del predimensionado de costes, dado que en esta fase del proyecto arquitectónico aún no se dispone de cálculos de estructura, planos de acondicionamiento e instalaciones, ni de la resolución constructiva de la envolvente, por cuanto no se han desarrollado las especificaciones propias del posterior proyecto de ejecución. Esta estimación aproximada del coste de la obra, es sencilla de calcular mediante la aplicación práctica del modelo desarrollado, y ello tanto para estudiantes como para profesionales del sector de la construcción. Como se contiene y justifica en el presente trabajo, la aplicación práctica del modelo para el cálculo de costes en las fases preliminares del proyecto, es rápida y certera, siendo de sencilla aplicación tanto en vivienda unifamiliar (aisladas y pareadas), como en viviendas colectivas (bloques y manzanas). También, el modelo es de aplicación en el ámbito de la valoración inmobiliaria, tasaciones, análisis de viabilidad económica de promociones inmobiliarias, estimación de costes de obras terminadas y en general, cuando no se dispone del proyecto de ejecución y sea preciso calcular los costes de construcción de las obras proyectadas. Además, el modelo puede ser de aplicación para el chequeo de presupuestos calculados por el método analítico tradicional (estado de mediciones pormenorizadas por sus precios unitarios y costes descompuestos), tanto en obras de iniciativa privada como en obras promovidas por las Administraciones Públicas. Por último, como líneas abiertas a futuras investigaciones, el modelo de “predimensionado costes de referencia 5 niveles de cálculo”, se podría adaptar y aplicar para otros usos y tipologías diferentes a la residencial, como edificios de equipamientos y dotaciones públicas, valoración de edificios históricos, obras de urbanización interior y exterior de parcela, proyectos de parques y jardines, etc….. Estas lineas de investigación suponen trabajos paralelos al aquí desarrollado, y que a modo de avance parcial se recogen en las comunicaciones presentadas en los Congresos internacionales Scieconf/Junio 2013, Rics‐Cobra/Septiembre 2013 y en el IV Congreso nacional de patología en la edificación‐Ucam/Abril 2014. ABSTRACT The aim of this research is to develop a fast, efficient and accurate calculation model to estimate the final costs of construction, during the preliminary stages of the architectural project. It is a tool to be used during the preliminary study process, drafting and basic project. It is not therefore necessary to have the exact, graphic definition of the project in order to be able to calculate the cost‐scaling. It is assumed that no deviation 10% higher than the final cost of the projected work will occur during the implementation. To that purpose five levels of cost estimation are formulated in the scaling model, from a lower to a higher conceptual and graphic definition of the architectural project. The five calculation levels are: two that take as point of reference the ”exogenous” values of house sales (initial development and basic development), and three based on calculation of endogenous costs (preliminary study, drafting and basic project). The first ”exogenous” estimation level (level.1) is calculated over the market valuation of real estate development and the proportion the cost of land has over the value of the houses. The fifth level of valuation, also an ”exogenous” one (level.5) is calculated from the contrast between the basic external market value, the construction costs, and the estimated development costs of the projected work. This contrast between the ”repercussions of construction costs” and the market value is an innovation regarding the existing cost‐scaling models, as a methodological process of extrinsic verification and validation, of the accuracy and validity of the estimations obtained from the implementation of the model, which is called Pcr.5n (reference cost‐scaling with .5calculation levels according to the stage of project definition/ architectural conceptualization) The other three levels of “endogenous” construction cost‐scaling are estimated from internal analytical calculations by project units and synthetic calculations by construction systems and functional spaces. This is performed during the initial stages of the project corresponding to preliminary study process (level.2), drafting (level.3) and basic project (level.4). These theoretical internal calculations are finally evaluated and validated via implementation of the model in residential buildings, whose real costs on final payment of the works are known. As the level of definition and development of the project evolves, from preliminary study to basic project, the calculation improves in its level of efficiency and estimation accuracy, following the applied methodology: [successive approximations at finite intervals]. The basic hypothesis as above has been made, achieving a maximum deviation of one tenth, in the estimated calculation of the true cost of predimensioning work. The cost calculation for material execution of the works is developed from functional “three‐dimensional” cubic parameters for the planned space and constructive “two dimensional” metric parameters for the surface that envelopes around the facade and the building’s footprint on the plot. The functional and building costs are analyzed at every stage of the process of calculation with “thematic/specific” parameters of management (Pg), project (Pp) and execution (Pe) of the estimated work in question, and finally the cost of contractual construction is estimated, as a consequence of increasing the cost of material execution with the percentage pertaining to the thematic/specific parameter of the projected work. The construction cost‐scaling Pcr.5n model will be a useful tool of great interest in the professional field to estimate the cost of the Basic Project as prescribed in the technical and legal framework of application. According to the appendix of the Technical Building Code (CTE), it is compulsory that the basic project contains an “approximate valuation of the material execution of the work, projected by chapters”, that is, that the basic project must contain at least an “approximate estimate” by chapter, trade or technology. This approximate estimate in the Basic Project is to be performed through the cost‐scaling technique, given that structural calculations, reconditioning plans and definitive contruction details of the envelope are still not available at this stage of the architectural project, insofar as specifications pertaining to the later project have not yet been developed. This approximate estimate of the cost of the works is easy to calculate through the implementation of the given model, both for students and professionals of the building sector. As explained and justified in this work, the implementation of the model for cost‐scaling during the preliminary stage is fast and accurate, as well as easy to apply both in single‐family houses (detached and semi‐detached) and collective housing (blocks). The model can also be applied in the field of the real‐estate valuation, official appraisal, analysis of the economic viability of real estate developments, estimate of the cost of finished projects and, generally, when an implementation project is not available and it is necessary to calculate the building costs of the projected works. The model can also be applied to check estimates calculated by the traditional analytical method (state of measurements broken down into price per unit cost details), both in private works and those promoted by Public Authorities. Finally, as potential lines for future research, the “five levels of calculation cost‐scaling model”, could be adapted and applied to purposes and typologies other than the residential one, such as service buildings and public facilities, valuation of historical buildings, interior and exterior development works, park and garden planning, etc… These lines of investigation are parallel to this one and, by way of a preview, can be found in the dissertations given in the International Congresses Scieconf/June 2013, Rics‐Cobra/September 2013 and in the IV Congress on building pathology ‐Ucam/April 2014.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The use of techniques such as envelope tracking (ET) and envelope elimination and restoration (EER) can improve the efficiency of radio frequency power amplifiers (RFPA). In both cases, high-bandwidth DC/DC converters called envelope amplifiers (EA) are used to modulate the supply voltage of the RFPA. This paper addresses the analysis and design of a modified two-phase Buck converter optimized to operate as EA. The effects of multiphase operation on the tracking capabilities are analyzed. The use of a fourth-order output filter is proposed to increase the attenuation of the harmonics generated by the PWM operation, thus allowing a reduction of the ratio between the switching frequency and the converter bandwidth. The design of the output filter is addressed considering envelope tracking accuracy and distortion caused by the side bands arising from the nonlinear modulation process. Finally, the proposed analysis and design methods are supported by simulation results, as well as demonstrated by experiments obtained using two 100-W, 10-MHz, two-phase Buck EAs capable of accurately tracking a 1.5-MHz bandwidth OFDM signal.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Los análisis de fiabilidad representan una herramienta adecuada para contemplar las incertidumbres inherentes que existen en los parámetros geotécnicos. En esta Tesis Doctoral se desarrolla una metodología basada en una linealización sencilla, que emplea aproximaciones de primer o segundo orden, para evaluar eficientemente la fiabilidad del sistema en los problemas geotécnicos. En primer lugar, se emplean diferentes métodos para analizar la fiabilidad de dos aspectos propios del diseño de los túneles: la estabilidad del frente y el comportamiento del sostenimiento. Se aplican varias metodologías de fiabilidad — el Método de Fiabilidad de Primer Orden (FORM), el Método de Fiabilidad de Segundo Orden (SORM) y el Muestreo por Importancia (IS). Los resultados muestran que los tipos de distribución y las estructuras de correlación consideradas para todas las variables aleatorias tienen una influencia significativa en los resultados de fiabilidad, lo cual remarca la importancia de una adecuada caracterización de las incertidumbres geotécnicas en las aplicaciones prácticas. Los resultados también muestran que tanto el FORM como el SORM pueden emplearse para estimar la fiabilidad del sostenimiento de un túnel y que el SORM puede mejorar el FORM con un esfuerzo computacional adicional aceptable. Posteriormente, se desarrolla una metodología de linealización para evaluar la fiabilidad del sistema en los problemas geotécnicos. Esta metodología solamente necesita la información proporcionada por el FORM: el vector de índices de fiabilidad de las funciones de estado límite (LSFs) que componen el sistema y su matriz de correlación. Se analizan dos problemas geotécnicos comunes —la estabilidad de un talud en un suelo estratificado y un túnel circular excavado en roca— para demostrar la sencillez, precisión y eficiencia del procedimiento propuesto. Asimismo, se reflejan las ventajas de la metodología de linealización con respecto a las herramientas computacionales alternativas. Igualmente se muestra que, en el caso de que resulte necesario, se puede emplear el SORM —que aproxima la verdadera LSF mejor que el FORM— para calcular estimaciones más precisas de la fiabilidad del sistema. Finalmente, se presenta una nueva metodología que emplea Algoritmos Genéticos para identificar, de manera precisa, las superficies de deslizamiento representativas (RSSs) de taludes en suelos estratificados, las cuales se emplean posteriormente para estimar la fiabilidad del sistema, empleando la metodología de linealización propuesta. Se adoptan tres taludes en suelos estratificados característicos para demostrar la eficiencia, precisión y robustez del procedimiento propuesto y se discuten las ventajas del mismo con respecto a otros métodos alternativos. Los resultados muestran que la metodología propuesta da estimaciones de fiabilidad que mejoran los resultados previamente publicados, enfatizando la importancia de hallar buenas RSSs —y, especialmente, adecuadas (desde un punto de vista probabilístico) superficies de deslizamiento críticas que podrían ser no-circulares— para obtener estimaciones acertadas de la fiabilidad de taludes en suelos. Reliability analyses provide an adequate tool to consider the inherent uncertainties that exist in geotechnical parameters. This dissertation develops a simple linearization-based approach, that uses first or second order approximations, to efficiently evaluate the system reliability of geotechnical problems. First, reliability methods are employed to analyze the reliability of two tunnel design aspects: face stability and performance of support systems. Several reliability approaches —the first order reliability method (FORM), the second order reliability method (SORM), the response surface method (RSM) and importance sampling (IS)— are employed, with results showing that the assumed distribution types and correlation structures for all random variables have a significant effect on the reliability results. This emphasizes the importance of an adequate characterization of geotechnical uncertainties for practical applications. Results also show that both FORM and SORM can be used to estimate the reliability of tunnel-support systems; and that SORM can outperform FORM with an acceptable additional computational effort. A linearization approach is then developed to evaluate the system reliability of series geotechnical problems. The approach only needs information provided by FORM: the vector of reliability indices of the limit state functions (LSFs) composing the system, and their correlation matrix. Two common geotechnical problems —the stability of a slope in layered soil and a circular tunnel in rock— are employed to demonstrate the simplicity, accuracy and efficiency of the suggested procedure. Advantages of the linearization approach with respect to alternative computational tools are discussed. It is also found that, if necessary, SORM —that approximates the true LSF better than FORM— can be employed to compute better estimations of the system’s reliability. Finally, a new approach using Genetic Algorithms (GAs) is presented to identify the fully specified representative slip surfaces (RSSs) of layered soil slopes, and such RSSs are then employed to estimate the system reliability of slopes, using our proposed linearization approach. Three typical benchmark-slopes with layered soils are adopted to demonstrate the efficiency, accuracy and robustness of the suggested procedure, and advantages of the proposed method with respect to alternative methods are discussed. Results show that the proposed approach provides reliability estimates that improve previously published results, emphasizing the importance of finding good RSSs —and, especially, good (probabilistic) critical slip surfaces that might be non-circular— to obtain good estimations of the reliability of soil slope systems.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

La Energía eléctrica producida mediante tecnología eólica flotante es uno de los recursos más prometedores para reducir la dependencia de energía proveniente de combustibles fósiles. Esta tecnología es de especial interés en países como España, donde la plataforma continental es estrecha y existen pocas áreas para el desarrollo de estructuras fijas. Entre los diferentes conceptos flotantes, esta tesis se ha ocupado de la tipología semisumergible. Estas plataformas pueden experimentar movimientos resonantes en largada y arfada. En largada, dado que el periodo de resonancia es largo estos puede ser inducidos por efectos de segundo orden de deriva lenta que pueden tener una influencia muy significativa en las cargas en los fondeos. En arfada las fuerzas de primer orden pueden inducir grandes movimientos y por tanto la correcta determinación del amortiguamiento es esencial para la analizar la operatividad de la plataforma. Esta tesis ha investigado estos dos efectos, para ello se ha usado como caso base el diseño de una plataforma desarrollada en el proyecto Europeo Hiprwind. La plataforma se compone de 3 columnas cilíndricas unidas mediante montantes estructurales horizontales y diagonales, Los cilindros proporcionan flotabilidad y momentos adrizante. A la base de cada columna se le ha añadido un gran “Heave Plate” o placa de cierre. El diseño es similar a otros diseños previos (Windfloat). Se ha fabricado un modelo a escala de una de las columnas para el estudio detallado del amortiguamiento mediante oscilaciones forzadas. Las dimensiones del modelo (1m diámetro en la placa de cierre) lo hacen, de los conocidos por el candidato, el mayor para el que se han publicado datos. El diseño del cilindro se ha realizado de tal manera que permite la fijación de placas de cierre planas o con refuerzo, ambos modelos se han fabricado y analizado. El modelo con refuerzos es una reproducción exacta del diseño a escala real incluyendo detalles distintivos del mismo, siendo el más importante la placa vertical perimetral. Los ensayos de oscilaciones forzadas se han realizado para un rango de frecuencias, tanto para el disco plano como el reforzado. Se han medido las fuerzas durante los ensayos y se han calculado los coeficientes de amortiguamiento y de masa añadida. Estos coeficientes son necesarios para el cálculo del fondeo mediante simulaciones en el dominio del tiempo. Los coeficientes calculados se han comparado con la literatura existente, con cálculos potenciales y por ultimo con cálculos CFD. Para disponer de información relevante para el diseño estructural de la plataforma se han medido y analizado experimentalmente las presiones en la parte superior e inferior de cada placa de cierre. Para la correcta estimación numérica de las fuerzas de deriva lenta en la plataforma se ha realizado una campaña experimental que incluye ensayos con modelo cautivo de la plataforma completa en olas bicromaticas. Pese a que estos experimentos no reproducen un escenario de oleaje realista, los mismos permiten una verificación del modelo numérico mediante la comparación de fuerzas medidas en el modelo físico y el numérico. Como resultados de esta tesis podemos enumerar las siguientes conclusiones. 1. El amortiguamiento y la masa añadida muestran una pequeña dependencia con la frecuencia pero una gran dependencia con la amplitud del movimiento. siendo coherente con investigaciones existentes. 2. Las medidas con la placa de cierre reforzada con cierre vertical en el borde, muestra un amortiguamiento significativamente menor comparada con la placa plana. Esto implica que para ensayos de canal es necesario incluir estos detalles en el modelo. 3. La masa añadida no muestra grandes variaciones comparando placa plana y placa con refuerzos. 4. Un coeficiente de amortiguamiento del 6% del crítico se puede considerar conservador para el cálculo en el dominio de la frecuencia. Este amortiguamiento es equivalente a un coeficiente de “drag” de 4 en elementos de Morison cuadráticos en las placas de cierre usadas en simulaciones en el dominio del tiempo. 5. Se han encontrado discrepancias en algunos valores de masa añadida y amortiguamiento de la placa plana al comparar con datos publicados. Se han propuesto algunas explicaciones basadas en las diferencias en la relación de espesores, en la distancia a la superficie libre y también relacionadas con efectos de escala. 6. La presión en la placa con refuerzos son similares a las de la placa plana, excepto en la zona del borde donde la placa con refuerzo vertical induce una gran diferencias de presiones entre la cara superior e inferior. 7. La máxima diferencia de presión escala coherentemente con la fuerza equivalente a la aceleración de la masa añadida distribuida sobre la placa. 8. Las masas añadidas calculadas con el código potencial (WADAM) no son suficientemente precisas, Este software no contempla el modelado de placas de pequeño espesor con dipolos, la poca precisión de los resultados aumenta la importancia de este tipo de elementos al realizar simulaciones con códigos potenciales para este tipo de plataformas que incluyen elementos de poco espesor. 9. Respecto al código CFD (Ansys CFX) la precisión de los cálculos es razonable para la placa plana, esta precisión disminuye para la placa con refuerzo vertical en el borde, como era de esperar dado la mayor complejidad del flujo. 10. Respecto al segundo orden, los resultados, en general, muestran que, aunque la tendencia en las fuerzas de segundo orden se captura bien con los códigos numéricos, se observan algunas reducciones en comparación con los datos experimentales. Las diferencias entre simulaciones y datos experimentales son mayores al usar la aproximación de Newman, que usa únicamente resultados de primer orden para el cálculo de las fuerzas de deriva media. 11. Es importante remarcar que las tendencias observadas en los resultados con modelo fijo cambiarn cuando el modelo este libre, el impacto que los errores en las estimaciones de fuerzas segundo orden tienen en el sistema de fondeo dependen de las condiciones ambientales que imponen las cargas ultimas en dichas líneas. En cualquier caso los resultados que se han obtenido en esta investigación confirman que es necesaria y deseable una detallada investigación de los métodos usados en la estimación de las fuerzas no lineales en las turbinas flotantes para que pueda servir de guía en futuros diseños de estos sistemas. Finalmente, el candidato espera que esta investigación pueda beneficiar a la industria eólica offshore en mejorar el diseño hidrodinámico del concepto semisumergible. ABSTRACT Electrical power obtained from floating offshore wind turbines is one of the promising resources which can reduce the fossil fuel energy consumption and cover worldwide energy demands. The concept is the most competitive in countries, such as Spain, where the continental shelf is narrow and does not provide space for fixed structures. Among the different floating structures concepts, this thesis has dealt with the semisubmersible one. Platforms of this kind may experience resonant motions both in surge and heave directions. In surge, since the platform natural period is long, such resonance can be excited with second order slow drift forces and may have substantial influence on mooring loads. In heave, first order forces can induce significant motion, whose damping is a crucial factor for the platform downtime. These two topics have been investigated in this thesis. To this aim, a design developed during HiPRWind EU project, has been selected as reference case study. The platform is composed of three cylindrical legs, linked together by a set of structural braces. The cylinders provide buoyancy and restoring forces and moments. Large circular heave plates have been attached to their bases. The design is similar to other documented in literature (e.g. Windfloat), which implies outcomes could have a general value. A large scale model of one of the legs has been built in order to study heave damping through forced oscillations. The final dimensions of the specimen (one meter diameter discs) make it, to the candidate’s knowledge, the largest for which data has been published. The model design allows for the fitting of either a plain solid heave plate or a flapped reinforced one; both have been built. The latter is a model scale reproduction of the prototype heave plate and includes some distinctive features, the most important being the inclusion of a vertical flap on its perimeter. The forced oscillation tests have been conducted for a range of frequencies and amplitudes, with both the solid plain model and the vertical flap one. Forces have been measured, from which added mass and damping coefficients have been obtained. These are necessary to accurately compute time-domain simulations of mooring design. The coefficients have been compared with literature, and potential flow and CFD predictions. In order to provide information for the structural design of the platform, pressure measurements on the top and bottom side of the heave discs have been recorded and pressure differences analyzed. In addition, in order to conduct a detailed investigation on the numerical estimations of the slow-drift forces of the HiPRWind platform, an experimental campaign involving captive (fixed) model tests of a model of the whole platform in bichromatic waves has been carried out. Although not reproducing the more realistic scenario, these tests allowed a preliminary verification of the numerical model based directly on the forces measured on the structure. The following outcomes can be enumerated: 1. Damping and added mass coefficients show, on one hand, a small dependence with frequency and, on the other hand, a large dependence with the motion amplitude, which is coherent with previously published research. 2. Measurements with the prototype plate, equipped with the vertical flap, show that damping drops significantly when comparing this to the plain one. This implies that, for tank tests of the whole floater and turbine, the prototype plate, equipped with the flap, should be incorporated to the model. 3. Added mass values do not suffer large alterations when comparing the plain plate and the one equipped with a vertical flap. 4. A conservative damping coefficient equal to 6% of the critical damping can be considered adequate for the prototype heave plate for frequency domain analysis. A corresponding drag coefficient equal to 4.0 can be used in time domain simulations to define Morison elements. 5. When comparing to published data, some discrepancies in added mass and damping coefficients for the solid plain plate have been found. Explanations have been suggested, focusing mainly on differences in thickness ratio and distance to the free surface, and eventual scale effects. 6. Pressures on the plate equipped with the vertical flap are similar in magnitude to those of the plain plate, even though substantial differences are present close to the edge, where the flap induces a larger pressure difference in the reinforced case. 7. The maximum pressure difference scales coherently with the force equivalent to the acceleration of the added mass, distributed over the disc surface. 8. Added mass coefficient values predicted with the potential solver (WADAM) are not accurate enough. The used solver does not contemplate modeling thin plates with doublets. The relatively low accuracy of the results highlights the importance of these elements when performing potential flow simulations of offshore platforms which include thin plates. 9. For the full CFD solver (Ansys CFX), the accuracy of the computations is found reasonable for the plain plate. Such accuracy diminishes for the disc equipped with a vertical flap, an expected result considering the greater complexity of the flow. 10. In regards to second order effects, in general, the results showed that, although the main trend in the behavior of the second-order forces is well captured by the numerical predictions, some under prediction of the experimental values is visible. The gap between experimental and numerical results is more pronounced when Newman’s approximation is considered, making use exclusively of the mean drift forces calculated in the first-order solution. 11. It should be observed that the trends observed in the fixed model test may change when the body is free to float, and the impact that eventual errors in the estimation of the second-order forces may have on the mooring system depends on the characteristics of the sea conditions that will ultimately impose the maximum loads on the mooring lines. Nevertheless, the preliminary results obtained in this research do confirm that a more detailed investigation of the methods adopted for the estimation of the nonlinear wave forces on the FOWT would be welcome and may provide some further guidance for the design of such systems. As a final remark, the candidate hopes this research can benefit the offshore wind industry in improving the hydrodynamic design of the semi-submersible concept.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

La mecanización de las labores del suelo es la causa, por su consumo energético e impacto directo sobre el medio ambiente, que más afecta a la degradación y pérdida de productividad de los suelos. Entre los factores de disminución de la productividad se deben considerar la compactación, la erosión, el encostramiento y la pérdida de estructura. Todo esto obliga a cuidar el manejo agrícola de los suelos tratando de mejorar las condiciones del suelo y elevar sus rendimientos sin comprometer aspectos económicos, ecológicos y ambientales. En el presente trabajo se adecuan los parámetros constitutivos del modelo de Drucker Prager Extendido (DPE) que definen la fricción y la dilatancia del suelo en la fase de deformación plástica, para minimizar los errores en las predicciones durante la simulación de la respuesta mecánica de un Vertisol mediante el Método de Elementos Finitos. Para lo cual inicialmente se analizaron las bases teóricas que soportan este modelo, se determinaron las propiedades y parámetros físico-mecánicos del suelo requeridos como datos de entrada por el modelo, se determinó la exactitud de este modelo en las predicciones de la respuesta mecánica del suelo, se estimaron mediante el método de aproximación de funciones de Levenberg-Marquardt los parámetros constitutivos que definen la trayectoria de la curva esfuerzo-deformación plástica. Finalmente se comprobó la exactitud de las predicciones a partir de las adecuaciones realizadas al modelo. Los resultados permitieron determinar las propiedades y parámetros del suelo, requeridos como datos de entrada por el modelo, mostrando que su magnitud está en función su estado de humedad y densidad, además se obtuvieron los modelos empíricos de estas relaciones exhibiendo un R2>94%. Se definieron las variables que provocan las inexactitudes del modelo constitutivo (ángulo de fricción y dilatancia), mostrando que las mismas están relacionadas con la etapa de falla y deformación plástica. Finalmente se estimaron los valores óptimos de estos ángulos, disminuyendo los errores en las predicciones del modelo DPE por debajo del 4,35% haciéndelo adecuado para la simulación de la respuesta mecánica del suelo investigado. ABSTRACT The mechanization using farming techniques is one of the main factors that affects the most the soil, causing its degradation and loss of productivity, because of its energy consumption and direct impact on the environment. Compaction, erosion, crusting and loss of structure should be considered among the factors that decrease productivity. All this forces the necessity to take care of the agricultural-land management trying to improve soil conditions and increase yields without compromising economic, ecological and environmental aspects. The present study was aimed to adjust the parameters of the Drucker-Prager Extended Model (DPE), defining friction and dilation of soil in plastic deformation phase, in order to minimize the error of prediction when simulating the mechanical response of a Vertisol through the fine element method. First of all the theoretic fundamentals that withstand the model were analyzed. The properties and physical-mechanical parameters of the soil needed as input data to initialize the model, were established. And the precision of the predictions for the mechanical response of the soil was assessed. Then the constitutive parameters which define the path of the plastic stress-strain curve were estimated through Levenberg-Marquardt method of function approximations. Lastly the accuracy of the predictions from the adequacies made to the model was tested. The results permitted to determine those properties and parameters of the soil, needed in order to initialize the model. It showed that their magnitude is in function of density and humidity. Moreover, the empirical models from these relations were obtained: R2>94%. The variables producing inaccuracies in the constitutive model (angle of repose and dilation) were defined, and there was showed that they are linked with the plastic deformation and rupture point. Finally the optimal values of these angles were established, obtaining thereafter error values for the DPE model under 4, 35%, and making it suitable for the simulation of the mechanical response of the soil under study.