20 resultados para over-generalization and under-generalization problems
em Universidad Politécnica de Madrid
Resumo:
En este trabajo se han analizado varios problemas en el contexto de la elasticidad no lineal basándose en modelos constitutivos representativos. En particular, se han analizado problemas relacionados con el fenómeno de perdida de estabilidad asociada con condiciones de contorno en el caso de material reforzados con fibras. Cada problema se ha formulado y se ha analizado por separado en diferentes capítulos. En primer lugar se ha mostrado el análisis del gradiente de deformación discontinuo para un material transversalmente isótropo, en particular, el modelo del material considerado consiste de una base neo-Hookeana isótropa incrustada con fibras de refuerzo direccional caracterizadas con un solo parámetro. La solución de este problema se vincula con instabilidades que dan lugar al mecanismo de fallo conocido como banda de cortante. La perdida de elipticidad de las ecuaciones diferenciales de equilibrio es una condición necesaria para que aparezca este tipo de soluciones y por tanto las inestabilidades asociadas. En segundo lugar se ha analizado una deformación combinada de extensión, inación y torsión de un tubo cilíndrico grueso donde se ha encontrado que la deformación citada anteriormente puede ser controlada solo para determinadas direcciones de las fibras refuerzo. Para entender el comportamiento elástico del tubo considerado se ha ilustrado numéricamente los resultados obtenidos para las direcciones admisibles de las fibras de refuerzo bajo la deformación considerada. En tercer lugar se ha estudiado el caso de un tubo cilíndrico grueso reforzado con dos familias de fibras sometido a cortante en la dirección azimutal para un modelo de refuerzo especial. En este problema se ha encontrado que las inestabilidades que aparecen en el material considerado están asociadas con lo que se llama soluciones múltiples de la ecuación diferencial de equilibrio. Se ha encontrado que el fenómeno de instabilidad ocurre en un estado de deformación previo al estado de deformación donde se pierde la elipticidad de la ecuación diferencial de equilibrio. También se ha demostrado que la condición de perdida de elipticidad y ^W=2 = 0 (la segunda derivada de la función de energía con respecto a la deformación) son dos condiciones necesarias para la existencia de soluciones múltiples. Finalmente, se ha analizado detalladamente en el contexto de elipticidad un problema de un tubo cilíndrico grueso sometido a una deformación combinada en las direcciones helicoidal, axial y radial para distintas geotermias de las fibras de refuerzo . In the present work four main problems have been addressed within the framework of non-linear elasticity based on representative constitutive models. Namely, problems related to the loss of stability phenomena associated with boundary value problems for fibre-reinforced materials. Each of the considered problems is formulated and analysed separately in different chapters. We first start with the analysis of discontinuous deformation gradients for a transversely isotropic material under plane deformation. In particular, the material model is an augmented neo-Hookean base with a simple unidirectional reinforcement characterised by a single parameter. The solution of this problem is related to material instabilities and it is associated with a shear band-type failure mode. The loss of ellipticity of the governing differential equations is a necessary condition for the existence of these material instabilities. The second problem involves a detailed analysis of the combined non-linear extension, inflation and torsion of a thick-walled circular cylindrical tube where it has been found that the aforementioned deformation is controllable only for certain preferred directions of transverse isotropy. Numerical results have been illustrated to understand the elastic behaviour of the tube for the admissible preferred directions under the considered deformation. The third problem deals with the analysis of a doubly fibre-reinforced thickwalled circular cylindrical tube undergoing pure azimuthal shear for a special class of the reinforcing model where multiple non-smooth solutions emerge. The associated instability phenomena are found to occur prior to the point where the nominal stress tensor changes monotonicity in a particular direction. It has been also shown that the loss of ellipticity condition that arises from the equilibrium equation and ^W=2 = 0 (the second derivative of the strain-energy function with respect to the deformation) are equivalent necessary conditions for the emergence of multiple solutions for the considered material. Finally, a detailed analysis in the basis of the loss of ellipticity of the governing differential equations for a combined helical, axial and radial elastic deformations of a fibre-reinforced circular cylindrical tube is carried out.
Resumo:
In the last decade, a large number of software repositories have been created for different purposes. In this paper we present a survey of the publicly available repositories and classify the most common ones as well as discussing the problems faced by researchers when applying machine learning or statistical techniques to them.
Resumo:
Polymers constitute a distinct class of anisotropic particles with unique dynamical, rheological and mechanical properties.
Resumo:
We review the main results from extensive Monte Carlo (MC) simulations on athermal polymer packings in the bulk and under confinement. By employing the simplest possible model of excluded volume, macromolecules are represented as freely-jointed chains of hard spheres of uniform size. Simulations are carried out in a wide concentration range: from very dilute up to very high volume fractions, reaching the maximally random jammed (MRJ) state. We study how factors like chain length, volume fraction and flexibility of bond lengths affect the structure, shape and size of polymers, their packing efficiency and their phase behaviour (disorder–order transition). In addition, we observe how these properties are affected by confinement realized by flat, impenetrable walls in one dimension. Finally, by mapping the parent polymer chains to primitive paths through direct geometrical algorithms, we analyse the characteristics of the entanglement network as a function of packing density.
Resumo:
El objetivo de la tesis es la investigación de algoritmos numéricos para el desarrollo de herramientas numéricas para la simulación de problemas tanto de comportamiento en la mar como de resistencia al avance de buques y estructuras flotantes. La primera herramienta desarrollada resuelve el problema de difracción y radiación de olas. Se basan en el método de los elementos finitos (MEF) para la resolución de la ecuación de Laplace, así como en esquemas basados en MEF, integración a lo largo de líneas de corriente, y en diferencias finitas desarrollados para la condición de superficie libre. Se han desarrollado herramientas numéricas para la resolución de la dinámica de sólido rígido en sistemas multicuerpos con ligaduras. Estas herramientas han sido integradas junto con la herramienta de resolución de olas difractadas y radiadas para la resolución de problemas de interacción de cuerpos con olas. También se han diseñado algoritmos de acoplamientos con otras herramientas numéricas para la resolución de problemas multifísica. En particular, se han realizado acoplamientos con una herramienta numérica basada de cálculo de estructuras con MEF para problemas de interacción fluido-estructura, otra de cálculo de líneas de fondeo, y con una herramienta numérica de cálculo de flujos en tanques internos para problemas acoplados de comportamiento en la mar con “sloshing”. Se han realizado simulaciones numéricas para la validación y verificación de los algoritmos desarrollados, así como para el análisis de diferentes casos de estudio con aplicaciones diversas en los campos de la ingeniería naval, oceánica, y energías renovables marinas. ABSTRACT The objective of this thesis is the research on numerical algorithms to develop numerical tools to simulate seakeeping problems as well as wave resistance problems of ships and floating structures. The first tool developed is a wave diffraction-radiation solver. It is based on the finite element method (FEM) in order to solve the Laplace equation, as well as numerical schemes based on FEM, streamline integration, and finite difference method tailored for solving the free surface boundary condition. It has been developed numerical tools to solve solid body dynamics of multibody systems with body links across them. This tool has been integrated with the wave diffraction-radiation solver to solve wave-body interaction problems. Also it has been tailored coupling algorithms with other numerical tools in order to solve multi-physics problems. In particular, it has been performed coupling with a MEF structural solver to solve fluid-structure interaction problems, with a mooring solver, and with a solver capable of simulating internal flows in tanks to solve couple seakeeping-sloshing problems. Numerical simulations have been carried out to validate and verify the developed algorithms, as well as to analyze case studies in the areas of marine engineering, offshore engineering, and offshore renewable energy.
Resumo:
Water is fundamental to human life and the availability of freshwater is often a constraint on human welfare and economic development. Consequently, the potential effects of global changes on hydrology and water resources are considered among the most severe and vital ones. Water scarcity is one of the main problems in the rural communities of Central America, as a result of an important degradation of catchment areas and the over-exploitation of aquifers. The present Thesis is focused on two critical aspects of global changes over water resources: (1) the potential effects of climate change on water quantity and (2) the impacts of land cover and land use changes on the hydrological processes and water cycle. Costa Rica is among the few developing countries that have recently achieved a land use transition with a net increase in forest cover. Osa Region in South Pacific Costa Rica is an appealing study site to assess water supply management plans and to measure the effects of deforestation, forest transitions and climate change projections reported in the region. Rural Community Water Supply systems (ASADAS) in Osa are dealing with an increasing demand of freshwater due to the growing population and the change in the way of life in the rural livelihoods. Land cover mosaics which have resulted from the above mentioned processes are characterized by the abandonment of marginal farmland with the spread over these former grasslands of high return crops and the expansion of secondary forests due to reforestation initiatives. These land use changes have a significant impact on runoff generation in priority water-supply catchments in the humid tropics, as evidenced by the analysis of the Tinoco Experimental Catchment in the Southern Pacific area of Costa Rica. The monitoring system assesses the effects of the different land uses on the runoff responses and on the general water cycle of the basin. Runoff responses at plot scale are analyzed for secondary forests, oil palm plantations, forest plantations and grasslands. The Oil palm plantation plot presented the highest runoff coefficient (mean RC=32.6%), twice that measured under grasslands (mean RC=15.3%) and 20-fold greater than in secondary forest (mean RC=1.7%). A Thornthwaite-type water balance is proposed to assess the impact of land cover and climate change scenarios over water availability for rural communities in Osa Region. Climate change projections were obtained by the downscaling of BCM2, CNCM3 and ECHAM5 models. Precipitation and temperature were averaged and conveyed by the A1B, A2 and B1 IPCC climate scenario for 2030, 2060 and 2080. Precipitation simulations exhibit a positive increase during the dry season for the three scenarios and a decrease during the rainy season, with the highest magnitude (up to 25%) by the end of the 21st century under scenario B1. Monthly mean temperature simulations increase for the three scenarios throughout the year with a maximum increase during the dry season of 5% under A1B and A2 scenarios and 4% under B1 scenario. The Thornthwaite-type Water Balance model indicates important decreases of water surplus for the three climate scenarios during the rainy season, with a maximum decrease on May, which under A1B scenario drop up to 20%, under A2 up to 40% and under B1 scenario drop up to almost 60%. Land cover scenarios were created taking into account current land cover dynamics of the region. Land cover scenario 1 projects a deforestation situation, with forests decreasing up to 15% due to urbanization of the upper catchment areas; land cover scenario 2 projects a forest recovery situation where forested areas increase due to grassland abandonment on areas with more than 30% of slope. Deforestation scenario projects an annual water surplus decrease of 15% while the reforestation scenario projects a water surplus increase of almost 25%. This water balance analysis indicates that climate scenarios are equal contributors as land cover scenarios to future water resource estimations.
Resumo:
A method to reduce truncation errors in near-field antenna measurements is presented. The method is based on the Gerchberg-Papoulis iterative algorithm used to extrapolate band-limited functions and it is able to extend the valid region of the calculated far-field pattern up to the whole forward hemisphere. The extension of the valid region is achieved by the iterative application of a transformation between two different domains. After each transformation, a filtering process that is based on known information at each domain is applied. The first domain is the spectral domain in which the plane wave spectrum (PWS) is reliable only within a known region. The second domain is the field distribution over the antenna under test (AUT) plane in which the desired field is assumed to be concentrated on the antenna aperture. The method can be applied to any scanning geometry, but in this paper, only the planar, cylindrical, and partial spherical near-field measurements are considered. Several simulation and measurement examples are presented to verify the effectiveness of the method.
Resumo:
El uso de las partículas magnéticas ha venido a más durante los últimos 10 años. Lo anterior ha estado condicionado por factores como la amplia investigación en materiales a nivel atómico, los retos propuestos por las diferentes áreas de investigación donde los niveles nanométricos cobran importancia, la inherente innovación tecnológica que pueden aportar a la industria dichas escalas y finalmente, el impacto socioeconómico que pueden tener estos avances tecnológicos. Sin embargo, la traslación a la industria de los avances tecnológicos se ha visto limitada por la falta de estrategias, ya que el 88% del conocimiento se queda en Universidades y Centros de Investigación y el resto, 12%, pasa a la industria o es obtenido en ella. En este trabajo de investigación se ha tratado de completar el círculo de idea, conceptualización, generación de conocimiento, prototipo, prueba in vitro y prueba in vivo; todo ello para dejar el menor espacio posible entre Investigación e Industria. Se ha fabricado un recinto que permite la observación del comportamiento de las partículas magnéticas cuando están inmersas en un medio con viscosidad controlada y con una fuente magnética estática. De la observación experimental se han encontrado fenómenos magnéticos como la aglomeración y la autoorganización, que han sido aprovechados, no solo para ofrecer una solución alternativa a la corrección del desprendimiento de retina, sino también para promover las mejoras de métodos de separación y/o filtrado celular, ayuda en correcciones ortopédicas o en el diagnóstico y tratamiento de enfermedades degenerativas. Particularizando en la corrección del desprendimiento de retina, se ha desarrollado y patentado un Dispositivo Ocular cuya función es la de tapón del agujero retiniano, ayudando en la cicatrización de la retina. Dicho dispositivo ha sido evaluado en conejos albinos neozelandeses obteniéndose un éxito en la reaplicación de la retina del 80% y resultando su toxicidad en condición difusa tras evaluación anatomopatológica. Así mismo, el ferrofluido diseñado y fabricado se ha localizado en la zona de interacción magnética siempre y su disposición ha sido en el mayor de los casos en forma de placa pero también se han formado anillos; lo anterior no ha afectado en ningún sentido el éxito en la reaplicación de la retina. Por otro lado, no se han encontrado indicios de siderosis en la zona de interacción magnética, sus alrededores o el resto del globo ocular; ni las células han presentado apoptosis inducida por los materiales introducidos en el ojo. Sin embargo, se ha encontrado una fuerte necrosis inducida por la presión ejercida sobre la retina por el dispositivo ocular, evidenciando la necesidad de un mejor control de dicha presión. Tras la caracterización de los copos y la presión que éstos ejercen, se ha podido identificar que el dispositivo ocular se puede sintonizar para inducir una presión entre 0.5 y 2.5 mmHg; por tanto se puede vislumbrar que es posible minimizar la muerte celular causada en las células retinianas. Finalmente, tras la experimentación con conejos se ha observado que el dispositivo ocular resuelve los problemas inducidos por las siliconas o los gases e inclusive aporta ventajas competitivas de alto valor añadido como: no requerir de posiciones incómodas en el post-operatorio, no modifica las propiedades ópticas del ojo ya que no es necesario dejar silicona o gases en el interior del ojo, se pueden sellar múltiples agujeros retinianos, corrección de desprendimiento de retina inferior sin complicaciones asociadas, etc.; en definitiva mejora la calidad de vida domiciliaria del paciente además de solventar los problemas inducidos por las soluciones actuales. The using of magnetic particles has become more important during last 10 years. This event has been conditioned by several factors like a wide researching on materials at atomic level, the challenges proposed by different research areas where nanometric levels became important, the inherent technological innovation to the industry and finally, the socioeconomic impact of these kinds of technologies. However, the nanotechnological advances have not been well addressed to the industry since 88% of the acknowledge keeps on Universities and Research Centers, and the remaining 12% goes through the industry or is obtained in it. In this research work, we have tried to complete the entire process about idea, concept, acknowledge generation, prototyping and in vitro and in vivo testing; all to leave as little space as possible between Research and Industry. We have built an experimental hall which allows us the observation of magnetic particles' behavior when are immersing in a controlled viscous medium and under a static magnetic field. The experimental observation has been useful to identify and use two magnetic phenomena: agglomeration and self-organization; no only for retinal detachment, but also to improve cellular separation and/or filtered methods, orthopedic solutions or the diagnostic and treatment of degenerative diseases. Particularizing on retinal detachment, we have developed and patented an Ocular Device which its function is acting like a tampon of the retinal hole for helping with the scarring retina. The device has been evaluated in animals and the results show that we can achieve a success of 80% before rabbit’s death and non cell apoptosis, only necrosis, over the retina attached by the ocular device after anatomopathological evaluation. Also we have identified a diffuse toxicity after anatomopathological evaluation, so more experimenting must be proposed. Otherwise, the ferrofluid has been localized where the magnetic interaction area is located and its deposition has been in most cases plaque shape like but rings have also been observed; this has not affected in any way the successful reattachment of the retina. On the other hand, we have not found any evidence of siderosis in the magnetic interaction zone, near of or some other place of the ocular globe; and we did not find any apoptosis induced by the materials inserted within the eye. However full necrosis, exactly on the magnetic interaction zone, has been found and induced by the pressure exerted over the retina by the ocular device, evidencing the needing of its better control. After flakes’ characterization and its pressure exerted, we identified that the ocular device can be tuned in order to induce a pressure between 0.5 and 2.5 mmHg; therefore we can perceive that death caused to the retinal cells can be minimized. Finally, the experimentation has shown that the ocular device solves all problems induced by silicone or gases an even contributes with some other competitive advantages of high added value like: no uncomfortable postoperative positions, it does not modify any optical property of the eye because there is no needing of leaving silicone or gases inside the eye, many holes can be sealed, the inferior retinal detachment can be solve without any complications, etc.; definitively the ocular device improves patients’ quality of life.
Resumo:
Zinc chelates have been widely used to correct deficiencies in this micronutrient in different soil types and under different moisture conditions. The aging of the metal in soil could cause a change in its availability. Over time the most labile forms of Zn could decrease in activity and extractability and change to more stable forms. Various soil parameters, such as redox conditions, time, soil type and moisture conditions, affect the aging process and modify the solubility of the metal. In general, redox conditions influence pH and also the chemical forms dissolved in the soil solution. Soil pH also affects Zn solubility; at high pH values, most of the Zn is present in forms that are not bioavailable to plants. The objective of this study was to determine the changes in Zn over time in a soil solution in a waterlogged acidic soil to which synthetic and natural chelates were applied
Resumo:
Abstract The proliferation of wireless sensor networks and the variety of envisioned applications associated with them has motivated the development of distributed algorithms for collaborative processing over networked systems. One of the applications that has attracted the attention of the researchers is that of target localization where the nodes of the network try to estimate the position of an unknown target that lies within its coverage area. Particularly challenging is the problem of estimating the target’s position when we use received signal strength indicator (RSSI) due to the nonlinear relationship between the measured signal and the true position of the target. Many of the existing approaches suffer either from high computational complexity (e.g., particle filters) or lack of accuracy. Further, many of the proposed solutions are centralized which make their application to a sensor network questionable. Depending on the application at hand and, from a practical perspective it could be convenient to find a balance between localization accuracy and complexity. Into this direction we approach the maximum likelihood location estimation problem by solving a suboptimal (and more tractable) problem. One of the main advantages of the proposed scheme is that it allows for a decentralized implementation using distributed processing tools (e.g., consensus and convex optimization) and therefore, it is very suitable to be implemented in real sensor networks. If further accuracy is needed an additional refinement step could be performed around the found solution. Under the assumption of independent noise among the nodes such local search can be done in a fully distributed way using a distributed version of the Gauss-Newton method based on consensus. Regardless of the underlying application or function of the sensor network it is al¬ways necessary to have a mechanism for data reporting. While some approaches use a special kind of nodes (called sink nodes) for data harvesting and forwarding to the outside world, there are however some scenarios where such an approach is impractical or even impossible to deploy. Further, such sink nodes become a bottleneck in terms of traffic flow and power consumption. To overcome these issues instead of using sink nodes for data reporting one could use collaborative beamforming techniques to forward directly the generated data to a base station or gateway to the outside world. In a dis-tributed environment like a sensor network nodes cooperate in order to form a virtual antenna array that can exploit the benefits of multi-antenna communications. In col-laborative beamforming nodes synchronize their phases in order to add constructively at the receiver. Some of the inconveniences associated with collaborative beamforming techniques is that there is no control over the radiation pattern since it is treated as a random quantity. This may cause interference to other coexisting systems and fast bat-tery depletion at the nodes. Since energy-efficiency is a major design issue we consider the development of a distributed collaborative beamforming scheme that maximizes the network lifetime while meeting some quality of service (QoS) requirement at the re¬ceiver side. Using local information about battery status and channel conditions we find distributed algorithms that converge to the optimal centralized beamformer. While in the first part we consider only battery depletion due to communications beamforming, we extend the model to account for more realistic scenarios by the introduction of an additional random energy consumption. It is shown how the new problem generalizes the original one and under which conditions it is easily solvable. By formulating the problem under the energy-efficiency perspective the network’s lifetime is significantly improved. Resumen La proliferación de las redes inalámbricas de sensores junto con la gran variedad de posi¬bles aplicaciones relacionadas, han motivado el desarrollo de herramientas y algoritmos necesarios para el procesado cooperativo en sistemas distribuidos. Una de las aplicaciones que suscitado mayor interés entre la comunidad científica es la de localization, donde el conjunto de nodos de la red intenta estimar la posición de un blanco localizado dentro de su área de cobertura. El problema de la localization es especialmente desafiante cuando se usan niveles de energía de la seal recibida (RSSI por sus siglas en inglés) como medida para la localization. El principal inconveniente reside en el hecho que el nivel de señal recibida no sigue una relación lineal con la posición del blanco. Muchas de las soluciones actuales al problema de localization usando RSSI se basan en complejos esquemas centralizados como filtros de partículas, mientas que en otras se basan en esquemas mucho más simples pero con menor precisión. Además, en muchos casos las estrategias son centralizadas lo que resulta poco prácticos para su implementación en redes de sensores. Desde un punto de vista práctico y de implementation, es conveniente, para ciertos escenarios y aplicaciones, el desarrollo de alternativas que ofrezcan un compromiso entre complejidad y precisión. En esta línea, en lugar de abordar directamente el problema de la estimación de la posición del blanco bajo el criterio de máxima verosimilitud, proponemos usar una formulación subóptima del problema más manejable analíticamente y que ofrece la ventaja de permitir en¬contrar la solución al problema de localization de una forma totalmente distribuida, convirtiéndola así en una solución atractiva dentro del contexto de redes inalámbricas de sensores. Para ello, se usan herramientas de procesado distribuido como los algorit¬mos de consenso y de optimización convexa en sistemas distribuidos. Para aplicaciones donde se requiera de un mayor grado de precisión se propone una estrategia que con¬siste en la optimización local de la función de verosimilitud entorno a la estimación inicialmente obtenida. Esta optimización se puede realizar de forma descentralizada usando una versión basada en consenso del método de Gauss-Newton siempre y cuando asumamos independencia de los ruidos de medida en los diferentes nodos. Independientemente de la aplicación subyacente de la red de sensores, es necesario tener un mecanismo que permita recopilar los datos provenientes de la red de sensores. Una forma de hacerlo es mediante el uso de uno o varios nodos especiales, llamados nodos “sumidero”, (sink en inglés) que actúen como centros recolectores de información y que estarán equipados con hardware adicional que les permita la interacción con el exterior de la red. La principal desventaja de esta estrategia es que dichos nodos se convierten en cuellos de botella en cuanto a tráfico y capacidad de cálculo. Como alter¬nativa se pueden usar técnicas cooperativas de conformación de haz (beamforming en inglés) de manera que el conjunto de la red puede verse como un único sistema virtual de múltiples antenas y, por tanto, que exploten los beneficios que ofrecen las comu¬nicaciones con múltiples antenas. Para ello, los distintos nodos de la red sincronizan sus transmisiones de manera que se produce una interferencia constructiva en el recep¬tor. No obstante, las actuales técnicas se basan en resultados promedios y asintóticos, cuando el número de nodos es muy grande. Para una configuración específica se pierde el control sobre el diagrama de radiación causando posibles interferencias sobre sis¬temas coexistentes o gastando más potencia de la requerida. La eficiencia energética es una cuestión capital en las redes inalámbricas de sensores ya que los nodos están equipados con baterías. Es por tanto muy importante preservar la batería evitando cambios innecesarios y el consecuente aumento de costes. Bajo estas consideraciones, se propone un esquema de conformación de haz que maximice el tiempo de vida útil de la red, entendiendo como tal el máximo tiempo que la red puede estar operativa garantizando unos requisitos de calidad de servicio (QoS por sus siglas en inglés) que permitan una decodificación fiable de la señal recibida en la estación base. Se proponen además algoritmos distribuidos que convergen a la solución centralizada. Inicialmente se considera que la única causa de consumo energético se debe a las comunicaciones con la estación base. Este modelo de consumo energético es modificado para tener en cuenta otras formas de consumo de energía derivadas de procesos inherentes al funcionamiento de la red como la adquisición y procesado de datos, las comunicaciones locales entre nodos, etc. Dicho consumo adicional de energía se modela como una variable aleatoria en cada nodo. Se cambia por tanto, a un escenario probabilístico que generaliza el caso determinista y se proporcionan condiciones bajo las cuales el problema se puede resolver de forma eficiente. Se demuestra que el tiempo de vida de la red mejora de forma significativa usando el criterio propuesto de eficiencia energética.
Resumo:
Goal-level Independent and-parallelism (IAP) is exploited by scheduling for simultaneous execution two or more goals which will not interfere with each other at run time. This can be done safely even if such goals can produce multiple answers. The most successful IAP implementations to date have used recomputation of answers and sequentially ordered backtracking. While in principle simplifying the implementation, recomputation can be very inefficient if the granularity of the parallel goals is large enough and they produce several answers, while sequentially ordered backtracking limits parallelism. And, despite the expected simplification, the implementation of the classic schemes has proved to involve complex engineering, with the consequent difficulty for system maintenance and expansion, and still frequently run into the well-known trapped goal and garbage slot problems. This work presents ideas about an alternative parallel backtracking model for IAP and a simulation studio. The model features parallel out-of-order backtracking and relies on answer memoization to reuse and combine answers. Whenever a parallel goal backtracks, its siblings also perform backtracking, but after storing the bindings generated by previous answers. The bindings are then reinstalled when combining answers. In order not to unnecessarily penalize forward execution, non-speculative and-parallel goals which have not been executed yet take precedence over sibling goals which could be backtracked over. Using a simulator, we show that this approach can bring significant performance advantages over classical approaches.
Resumo:
Higher education students demand fast feedback about their assignments and the opportunity to repeat them in case they do in a wrong way. Here a computer based trainer for Signals and Systems students is presented. An application, that automatically generates and assesses thousands of numerically different versions of several Signals and Systems problems have been developed. This applet guides the students to find the solution and automatically assesses and grades the students proposed solution. The students can use the application to practice in solving several types of Signals and Systems basic problems. After selecting the problem type, the student introduces a seed and the application generates a numerical version of the selected problem. Then the application presents a sequence of questions that the students must solve and the application automatically assess their answers. After solving a given problem, the students can repeat the same numerical variation of the problem by introducing the same seed to the application. In this way, they can review their solution with the help of the hints given by the application for wrong solutions. This application can also be used as an automatic assessment tool by the instructor. When the assessment is made in a controlled environment (examination classroom or laboratory) the instructor can use the same seed for all students. Otherwise, different seeds can be assigned to different students and in this way they solve different numerical variation of the proposed problem, so cheating becomes an arduous task. Given a problem type, the mathematical or conceptual difficulty of the problem can vary depending on the numerical values of the parameters of the problem. The application permits to easily select groups of seeds that yield to numerical variations with similar mathematical or conceptual difficulty. This represents an advantage over a randomised task assignment where students are asked to solve tasks with different difficulty.
Resumo:
Helium Brayton cycles have been studied as power cycles for both fission and fusion reactors obtaining high thermal efficiency. This paper studies several technological schemes of helium Brayton cycles applied for the HiPER reactor proposal. Since HiPER integrates technologies available at short term, its working conditions results in a very low maximum temperature of the energy sources, something that limits the thermal performance of the cycle. The aim of this work is to analyze the potential of the helium Brayton cycles as power cycles for HiPER. Several helium Brayton cycle configurations have been investigated with the purpose of raising the cycle thermal efficiency under the working conditions of HiPER. The effects of inter-cooling and reheating have specifically been studied. Sensitivity analyses of the key cycle parameters and component performances on the maximum thermal efficiency have also been carried out. The addition of several inter-cooling stages in a helium Brayton cycle has allowed obtaining a maximum thermal efficiency of over 36%, and the inclusion of a reheating process may also yield an added increase of nearly 1 percentage point to reach 37%. These results confirm that helium Brayton cycles are to be considered among the power cycle candidates for HiPER.
Resumo:
Mediterranean Dehesas are one of the European natural habitat types of Community interest (43/92/EEC Directive), associated to high diversity levels and producer of important goods and services. In this work, tree contribution and grazing influence over pasture alpha diversity in a Dehesa in Central Spain was studied. We analyzed Richness and Shannon-Wiener (SW) indexes on herbaceous layer under 16 holms oak trees (64 sampling units distributed in two directions and in two distances to the trunk) distributed in four different grazing management zones (depending on species and stocking rate). Floristic composition by species or morphospecies and species abundance were analyzed for each sample unit. Linear mixed models (LMM) and generalized linear mixed models (GLMMs) were used to study relationships between alpha diversity measures and independent factors. Edge crown influence showed the highest values of Richness and SW index. No significant differences were found between orientations under tree crown influence. Grazing management had a significant effect over Richness and SW measures, specially the grazing species (cattle or sheep). We preliminary quantify and analyze the interaction of tree stratum and grazing management over herbaceous diversity in a year of extreme climatic conditions.
Resumo:
Los fundamentos de la Teoría de la Decisión Bayesiana proporcionan un marco coherente en el que se pueden resolver los problemas de toma de decisiones. La creciente disponibilidad de ordenadores potentes está llevando a tratar problemas cada vez más complejos con numerosas fuentes de incertidumbre multidimensionales; varios objetivos conflictivos; preferencias, metas y creencias cambiantes en el tiempo y distintos grupos afectados por las decisiones. Estos factores, a su vez, exigen mejores herramientas de representación de problemas; imponen fuertes restricciones cognitivas sobre los decisores y conllevan difíciles problemas computacionales. Esta tesis tratará estos tres aspectos. En el Capítulo 1, proporcionamos una revisión crítica de los principales métodos gráficos de representación y resolución de problemas, concluyendo con algunas recomendaciones fundamentales y generalizaciones. Nuestro segundo comentario nos lleva a estudiar tales métodos cuando sólo disponemos de información parcial sobre las preferencias y creencias del decisor. En el Capítulo 2, estudiamos este problema cuando empleamos diagramas de influencia (DI). Damos un algoritmo para calcular las soluciones no dominadas en un DI y analizamos varios conceptos de solución ad hoc. El último aspecto se estudia en los Capítulos 3 y 4. Motivado por una aplicación de gestión de embalses, introducimos un método heurístico para resolver problemas de decisión secuenciales. Como muestra resultados muy buenos, extendemos la idea a problemas secuenciales generales y cuantificamos su bondad. Exploramos después en varias direcciones la aplicación de métodos de simulación al Análisis de Decisiones. Introducimos primero métodos de Monte Cario para aproximar el conjunto no dominado en problemas continuos. Después, proporcionamos un método de Monte Cario basado en cadenas de Markov para problemas con información completa con estructura general: las decisiones y las variables aleatorias pueden ser continuas, y la función de utilidad puede ser arbitraria. Nuestro esquema es aplicable a muchos problemas modelizados como DI. Finalizamos con un capítulo de conclusiones y problemas abiertos.---ABSTRACT---The foundations of Bayesian Decisión Theory provide a coherent framework in which decisión making problems may be solved. With the advent of powerful computers and given the many challenging problems we face, we are gradually attempting to solve more and more complex decisión making problems with high and multidimensional uncertainty, múltiple objectives, influence of time over decisión tasks and influence over many groups. These complexity factors demand better representation tools for decisión making problems; place strong cognitive demands on the decison maker judgements; and lead to involved computational problems. This thesis will deal with these three topics. In recent years, many representation tools have been developed for decisión making problems. In Chapter 1, we provide a critical review of most of them and conclude with recommendations and generalisations. Given our second query, we could wonder how may we deal with those representation tools when there is only partial information. In Chapter 2, we find out how to deal with such a problem when it is structured as an influence diagram (ID). We give an algorithm to compute nondominated solutions in ID's and analyse several ad hoc solution concepts.- The last issue is studied in Chapters 3 and 4. In a reservoir management case study, we have introduced a heuristic method for solving sequential decisión making problems. Since it shows very good performance, we extend the idea to general problems and quantify its goodness. We explore then in several directions the application of simulation based methods to Decisión Analysis. We first introduce Monte Cario methods to approximate the nondominated set in continuous problems. Then, we provide a Monte Cario Markov Chain method for problems under total information with general structure: decisions and random variables may be continuous, and the utility function may be arbitrary. Our scheme is applicable to many problems modeled as IDs. We conclude with discussions and several open problems.