29 resultados para the arousal theory
Resumo:
In this work, the dimensional synthesis of a spherical Parallel Manipulator (PM) with a -1S kinematic chain is presented. The goal of the synthesis is to find a set of parameters that defines the PM with the best performance in terms of workspace capabilities, dexterity and isotropy. The PM is parametrized in terms of a reference element, and a non-directed search of these parameters is carried out. First, the inverse kinematics and instantaneous kinematics of the mechanism are presented. The latter is found using the screw theory formulation. An algorithm that explores a bounded set of parameters and determines the corresponding value of global indexes is presented. The concepts of a novel global performance index and a compound index are introduced. Simulation results are shown and discussed. The best PMs found in terms of each performance index evaluated are locally analyzed in terms of its workspace and local dexterity. The relationship between the performance of the PM and its parameters is discussed, and a prototype with the best performance in terms of the compound index is presented and analyzed.
Resumo:
At present, in the University curricula in most countries, the decision theory and the mathematical models to aid decision making is not included, as in the graduate program like in Doctored and Master´s programs. In the Technical School of High Level Agronomic Engineers of the Technical University of Madrid (ETSIA-UPM), the need to offer to the future engineers training in a subject that could help them to take decisions in their profession was felt. Along the life, they will have to take a lot of decisions. Ones, will be important and others no. In the personal level, they will have to take several very important decisions, like the election of a career, professional work, or a couple, but in the professional field, the decision making is the main role of the Managers, Politicians and Leaders. They should be decision makers and will be paid for it. Therefore, nobody can understand that such a professional that is called to practice management responsibilities in the companies, does not take training in such an important matter. For it, in the year 2000, it was requested to the University Board to introduce in the curricula an optional qualified subject of the second cycle with 4,5 credits titled " Mathematical Methods for Making Decisions ". A program was elaborated, the didactic material prepared and programs as Maple, Lingo, Math Cad, etc. installed in several IT classrooms, where the course will be taught. In the course 2000-2001 this subject was offered with a great acceptance that exceeded the forecasts of capacity and had to be prepared more classrooms. This course in graduate program took place in the Department of Applied Mathematics to the Agronomic Engineering, as an extension of the credits dedicated to Mathematics in the career of Engineering.
Resumo:
The flexural vibration of a homogeneous isotropic linearly elastic cylinder of any aspect ratio is analysed in this paper. Natural frequencies of a cylinder under uniformly distributed axial loads acting on its bases are calculated numerically by the Ritz method with terms of power series in the coordinate directions as approximating functions. The effect of axial loads on the flexural vibration cannot be described by applying infinitesimal strain theory, therefore, geometrically nonlinear strain–displacement relations with second-order terms are considered here. The natural frequencies of free–free, clamped–clamped, and sliding–sliding cylinders subjected to axial loads are calculated using the proposed three-dimensional Ritz approach and are compared with those obtained with the finite element method and the Bernoulli–Euler theory. Different experiments with cylinders axially compressed by a hydraulic press are carried out and the experimental results for the lowest flexural frequency are compared with the numerical results. An approach based on the Ritz formulation is proposed for the flexural vibration of a cylinder between the platens of the press with constraints varying with the intensity of the compression. The results show that for low compressions the cylinder behaves similarly to a sliding–sliding cylinder, whereas for high compressions the cylinder vibrates as a clamped–clamped one.
Resumo:
The purpose of this study is to determine the stress distribution in the carpentry joint of halved and tabled scarf joint with the finite element method (FEM) and its comparison with the values obtained using the theory of Strength of Materials. The stress concentration areas where analyzed and the influence of mesh refinement was studied on the results in order to determine the mesh size that provides the stress values more consistent with the theory. In areas where stress concentration is lower, different mesh sizes show similar stress values. In areas where stress concentration occurs, the same values increase considerably with the refinement of the mesh. The results show a central symmetry of the isobar lines distribution where the centre of symmetry corresponds to the geometric centre of the joint. Comparison of normal stress levels obtained by the FEM and the classical theory shows small differences, except at points of stress concentration.
Resumo:
La prevalencia de las alergias está aumentando desde mediados del siglo XX, y se estima que actualmente afectan a alrededor del 2-8 % de la población, pero las causas de este aumento aún no están claras. Encontrar el origen del mecanismo por el cual una proteína inofensiva se convierte en capaz de inducir una respuesta alérgica es de vital importancia para prevenir y tratar estas enfermedades. Aunque la caracterización de alérgenos relevantes ha ayudado a mejorar el manejo clínico y a aclarar los mecanismos básicos de las reacciones alérgicas, todavía queda un largo camino para establecer el origen de la alergenicidad y reactividad cruzada. El objetivo de esta tesis ha sido caracterizar las bases moleculares de la alergenicidad tomando como modelo dos familias de panalergenos (proteínas de transferencia de lípidos –LTPs- y taumatinas –TLPs-) y estudiando los mecanismos que median la sensibilización y la reactividad cruzada para mejorar tanto el diagnóstico como el tratamiento de la alergia. Para ello, se llevaron a cabo dos estrategias: estudiar la reactividad cruzada de miembros de familias de panalérgenos; y estudiar moléculas-co-adyuvantes que pudieran favorecer la capacidad alergénica de dichas proteínas. Para estudiar la reactividad cruzada entre miembros de la misma familia de proteínas, se seleccionaron LTPs y TLPs, descritas como alergenos, tomando como modelo la alergia a frutas. Por otra parte, se estudiaron los perfiles de sensibilización a alérgenos de trigo relacionados con el asma del panadero, la enfermedad ocupacional más relevante de origen alérgico. Estos estudios se llevaron a cabo estandarizando ensayos tipo microarrays con alérgenos y analizando los resultados por la teoría de grafos. En relación al estudiar moléculas-co-adyuvantes que pudieran favorecer la capacidad alergénica de dichas proteínas, se llevaron a cabo estudios sobre la interacción de los alérgenos alimentarios con células del sistema inmune humano y murino y el epitelio de las mucosas, analizando la importancia de moléculas co-transportadas con los alérgenos en el desarrollo de una respuesta Th2. Para ello, Pru p 3(LTP y alérgeno principal del melocotón) se selección como modelo para llevarlo a cabo. Por otra parte, se analizó el papel de moléculas activadoras del sistema inmune producidas por patógenos en la inducción de alergias alimentarias seleccionando el modelo kiwi-alternaria, y el papel de Alt a 1, alérgeno mayor de dicho hongo, en la sensibilización a Act d 2, alérgeno mayor de kiwi. En resumen, el presente trabajo presenta una investigación innovadora aportando resultados de gran utilidad tanto para la mejora del diagnóstico como para nuevas investigaciones sobre la alergia y el esclarecimiento final de los mecanismos que caracterizan esta enfermedad. ABSTRACT Allergies are increasing their prevalence from mid twentieth century, and they are currently estimated to affect around 2-8% of the population but the underlying causes of this increase remain still elusive. The understanding of the mechanism by which a harmless protein becomes capable of inducing an allergic response provides us the basis to prevent and treat these diseases. Although the characterization of relevant allergens has led to improved clinical management and has helped to clarify the basic mechanisms of allergic reactions, it seems justified in aspiring to molecularly dissecting these allergens to establish the structural basis of their allergenicity and cross-reactivity. The aim of this thesis was to characterize the molecular basis of the allergenicity of model proteins belonging to different families (Lipid Transfer Proteins –LTPs-, and Thaumatin-like Proteins –TLPs-) in order to identify mechanisms that mediate sensitization and cross reactivity for developing new strategies in the management of allergy, both diagnosis and treatment, in the near future. With this purpose, two strategies have been conducted: studies of cross-reactivity among panallergen families and molecular studies of the contribution of cofactors in the induction of the allergic response by these panallergens. Following the first strategy, we studied the cross-reactivity among members of two plant panallergens (LTPs , Lipid Transfer Proteins , and TLPs , Thaumatin-like Proteins) using the peach allergy as a model. Similarly, we characterized the sensitization profiles to wheat allergens in baker's asthma development, the most relevant occupational disease. These studies were performed using allergen microarrays and the graph theory for analyzing the results. Regarding the second approach, we analyzed the interaction of plant allergens with immune and epithelial cells. To perform these studies , we examined the importance of ligands and co-transported molecules of plant allergens in the development of Th2 responses. To this end, Pru p 3, nsLTP (non-specific Lipid Transfer Protein) and peach major allergen, was selected as a model to investigate its interaction with cells of the human and murine immune systems as well as with the intestinal epithelium and the contribution of its ligand in inducing an allergic response was studied. Moreover, we analyzed the role of pathogen associated molecules in the induction of food allergy. For that, we selected the kiwi- alternaria system as a model and the role of Alt a 1 , major allergen of the fungus, in the development of Act d 2-sensitization was studied. In summary, this work presents an innovative research providing useful results for improving diagnosis and leading to further research on allergy and the final clarification of the mechanisms that characterize this disease.
Resumo:
The intermediate band solar cell (IBSC), the multiple exciton generation solar cell (MEGSC) and the hot carrier solar cell (HCSC) are three novel concepts in photovoltaics which aim to achieve high efficiency devices. In this paper we assess to what extent their physical principles of operation have been experimentally verified. It is found that there is experimental evidence supporting the underlying theory for all three.
Resumo:
The classical theory of intermittency developed for return maps assumes uniform density of points reinjected from the chaotic to laminar region. Though it works fine in some model systems, there exist a number of so-called pathological cases characterized by a significant deviation of main characteristics from the values predicted on the basis of the uniform distribution. Recently, we reported on how the reinjection probability density (RPD) can be generalized. Here, we extend this methodology and apply it to different dynamical systems exhibiting anomalous type-II and type-III intermittencies. Estimation of the universal RPD is based on fitting a linear function to experimental data and requires no a priori knowledge on the dynamical model behind. We provide special fitting procedure that enables robust estimation of the RPD from relatively short data sets (dozens of points). Thus, the method is applicable for a wide variety of data sets including numerical simulations and real-life experiments. Estimated RPD enables analytic evaluation of the length of the laminar phase of intermittent behaviors. We show that the method copes well with dynamical systems exhibiting significantly different statistics reported in the literature. We also derive and classify characteristic relations between the mean laminar length and main controlling parameter in perfect agreement with data provided by numerical simulations
Resumo:
Advanced control techniques like V2, Vout hysteresis or V2Ic can strongly reduce the required output capacitance in PowerSoC converters. Techniques to analyze power converters based on the analysis of the frequency response are not suitable for ripple-based controllers that use fast-scale dynamics to control the power stage. This paper proves that the use of discrete modeling together with Floquet theory is a very powerful tool to model the system and derive stable region diagrams for sensitivity analysis. It is applied to V 2Ic control, validating experimentally that Floquet theory predicts accurately subharmonic oscillations. This method is applied to several ripplebased controllers, providing higher accuracy when it is compared with other techniques based on the frequency response. The paper experimentally validates the usefulness of the discrete modeling and the Floquet theory on a 5 MHz Buck converter with a V 2Ic control.
Resumo:
Surfactant monolayers are of interest in a variety of phenomena, including thin film dynamics and the formation and dynamics of foams. Measurement of surface properties has received a continuous attention and requires good theoretical models to extract the relevant physico- chemical information from experimental data. A common experimental set up consists in a shallow liquid layer whose free surface is slowly com- pressed/expanded in periodic fashion by moving two slightly immersed solid barriers, which varies the free surface area and thus the surfactant concentration. The simplest theory ignores the fluid dynamics in the bulk fluid, assuming spatially uniform surfactant concentration, which requires quite small forcing frequencies and provides reversible dynamics in the compression/expansion cycles. Sometimes, it is not clear whether depar- ture from reversibility is due to non-equilibrium effects or to the ignored fluid dynamics. Here we present a long wave theory that takes the fluid dynamics and the symmetries of the problem into account. In particular, the validity of the spatially-uniform-surfactant-concentration assumption is established and a nonlinear diffusion equation is derived. This allows for calculating spatially nonuniform monolayer dynamics and uncovering the physical mechanisms involved in the surfactant behavior. Also, this analysis can be considered a good means for extracting more relevant information from each experimental run.
Resumo:
This paper is an introduction of the regret theory-based scenario building approach combining with a modified Delphi method that uses an interactive process to design and assess four different TDM measures (i.e., cordon toll, parking charge, increased bus frequency and decreased bus fare). The case study of Madrid is used to present the analysis and provide policy recommendations. The new scenario building approach incorporates expert judgement and transport models in an interactive process. It consists of a two-round modified Delphi survey, which was answeared by a group of Spanish transport experts who were the participants of the Transport Engineering Congress (CIT 2012), and an integrated land-use and transport model (LUTI) for Madrid that is called MARS (Metropolitan Activity Relocation Simulator).
Resumo:
This study suggests a theoretical framework for improving the teaching/ learning process of English employed in the Aeronautical discourse that brings together cognitive learning strategies, Genre Analysis and the Contemporary theory of Metaphor (Lakoff and Johnson 1980; Lakoff 1993). It maintains that cognitive strategies such as imagery, deduction, inference and grouping can be enhanced by means of metaphor and genre awareness in the context of content based approach to language learning. A list of image metaphors and conceptual metaphors which comes from the terminological database METACITEC is provided. The metaphorical terms from the area of Aeronautics have been taken from specialised dictionaries and have been categorised according to the conceptual metaphors they respond to, by establishing the source domains and the target domains, as well as the semantic networks found. This information makes reference to the internal mappings underlying the discourse of aeronautics reflected in five aviation accident case studies which are related to accident reports from the National Transportation Safety Board (NTSB) and provides an important source for designing language teaching tasks. La Lingüística Cognitiva y el Análisis del Género han contribuido a la mejora de la enseñanza de segundas lenguas y, en particular, al desarrollo de la competencia lingüística de los alumnos de inglés para fines específicos. Este trabajo pretende perfeccionar los procesos de enseñanza y el aprendizaje del lenguaje empleado en el discurso aeronáutico por medio de la práctica de estrategias cognitivas y prestando atención a la Teoría del análisis del género y a la Teoría contemporánea de la metáfora (Lakoff y Johnson 1980; Lakoff 1993). Con el propósito de crear recursos didácticos en los que se apliquen estrategias metafóricas, se ha elaborado un listado de metáforas de imagen y de metáforas conceptuales proveniente de la base de datos terminológica META-CITEC. Estos términos se han clasificado de acuerdo con las metáforas conceptuales y de imagen existentes en esta área de conocimiento. Para la enseñanza de este lenguaje de especialidad, se proponen las correspondencias y las proyecciones entre el dominio origen y el dominio meta que se han hallado en los informes de accidentes aéreos tomados de la Junta federal de la Seguridad en el Transporte (NTSB)
Resumo:
La investigación para el conocimiento del cerebro es una ciencia joven, su inicio se remonta a Santiago Ramón y Cajal en 1888. Desde esta fecha a nuestro tiempo la neurociencia ha avanzado mucho en el desarrollo de técnicas que permiten su estudio. Desde la neurociencia cognitiva hoy se explican muchos modelos que nos permiten acercar a nuestro entendimiento a capacidades cognitivas complejas. Aun así hablamos de una ciencia casi en pañales que tiene un lago recorrido por delante. Una de las claves del éxito en los estudios de la función cerebral ha sido convertirse en una disciplina que combina conocimientos de diversas áreas: de la física, de las matemáticas, de la estadística y de la psicología. Esta es la razón por la que a lo largo de este trabajo se entremezclan conceptos de diferentes campos con el objetivo de avanzar en el conocimiento de un tema tan complejo como el que nos ocupa: el entendimiento de la mente humana. Concretamente, esta tesis ha estado dirigida a la integración multimodal de la magnetoencefalografía (MEG) y la resonancia magnética ponderada en difusión (dMRI). Estas técnicas son sensibles, respectivamente, a los campos magnéticos emitidos por las corrientes neuronales, y a la microestructura de la materia blanca cerebral. A lo largo de este trabajo hemos visto que la combinación de estas técnicas permiten descubrir sinergias estructurofuncionales en el procesamiento de la información en el cerebro sano y en el curso de patologías neurológicas. Más específicamente en este trabajo se ha estudiado la relación entre la conectividad funcional y estructural y en cómo fusionarlas. Para ello, se ha cuantificado la conectividad funcional mediante el estudio de la sincronización de fase o la correlación de amplitudes entre series temporales, de esta forma se ha conseguido un índice que mide la similitud entre grupos neuronales o regiones cerebrales. Adicionalmente, la cuantificación de la conectividad estructural a partir de imágenes de resonancia magnética ponderadas en difusión, ha permitido hallar índices de la integridad de materia blanca o de la fuerza de las conexiones estructurales entre regiones. Estas medidas fueron combinadas en los capítulos 3, 4 y 5 de este trabajo siguiendo tres aproximaciones que iban desde el nivel más bajo al más alto de integración. Finalmente se utilizó la información fusionada de MEG y dMRI para la caracterización de grupos de sujetos con deterioro cognitivo leve, la detección de esta patología resulta relevante en la identificación precoz de la enfermedad de Alzheimer. Esta tesis está dividida en seis capítulos. En el capítulos 1 se establece un contexto para la introducción de la connectómica dentro de los campos de la neuroimagen y la neurociencia. Posteriormente en este capítulo se describen los objetivos de la tesis, y los objetivos específicos de cada una de las publicaciones científicas que resultaron de este trabajo. En el capítulo 2 se describen los métodos para cada técnica que fue empleada: conectividad estructural, conectividad funcional en resting state, redes cerebrales complejas y teoría de grafos y finalmente se describe la condición de deterioro cognitivo leve y el estado actual en la búsqueda de nuevos biomarcadores diagnósticos. En los capítulos 3, 4 y 5 se han incluido los artículos científicos que fueron producidos a lo largo de esta tesis. Estos han sido incluidos en el formato de la revista en que fueron publicados, estando divididos en introducción, materiales y métodos, resultados y discusión. Todos los métodos que fueron empleados en los artículos están descritos en el capítulo 2 de la tesis. Finalmente, en el capítulo 6 se concluyen los resultados generales de la tesis y se discuten de forma específica los resultados de cada artículo. ABSTRACT In this thesis I apply concepts from mathematics, physics and statistics to the neurosciences. This field benefits from the collaborative work of multidisciplinary teams where physicians, psychologists, engineers and other specialists fight for a common well: the understanding of the brain. Research on this field is still in its early years, being its birth attributed to the neuronal theory of Santiago Ramo´n y Cajal in 1888. In more than one hundred years only a very little percentage of the brain functioning has been discovered, and still much more needs to be explored. Isolated techniques aim at unraveling the system that supports our cognition, nevertheless in order to provide solid evidence in such a field multimodal techniques have arisen, with them we will be able to improve current knowledge about human cognition. Here we focus on the multimodal integration of magnetoencephalography (MEG) and diffusion weighted magnetic resonance imaging. These techniques are sensitive to the magnetic fields emitted by the neuronal currents and to the white matter microstructure, respectively. The combination of such techniques could bring up evidences about structural-functional synergies in the brain information processing and which part of this synergy fails in specific neurological pathologies. In particular, we are interested in the relationship between functional and structural connectivity, and how two integrate this information. We quantify the functional connectivity by studying the phase synchronization or the amplitude correlation between time series obtained by MEG, and so we get an index indicating similarity between neuronal entities, i.e. brain regions. In addition we quantify structural connectivity by performing diffusion tensor estimation from the diffusion weighted images, thus obtaining an indicator of the integrity of the white matter or, if preferred, the strength of the structural connections between regions. These quantifications are then combined following three different approaches, from the lowest to the highest level of integration, in chapters 3, 4 and 5. We finally apply the fused information to the characterization or prediction of mild cognitive impairment, a clinical entity which is considered as an early step in the continuum pathological process of dementia. The dissertation is divided in six chapters. In chapter 1 I introduce connectomics within the fields of neuroimaging and neuroscience. Later in this chapter we describe the objectives of this thesis, and the specific objectives of each of the scientific publications that were produced as result of this work. In chapter 2 I describe the methods for each of the techniques that were employed, namely structural connectivity, resting state functional connectivity, complex brain networks and graph theory, and finally, I describe the clinical condition of mild cognitive impairment and the current state of the art in the search for early biomarkers. In chapters 3, 4 and 5 I have included the scientific publications that were generated along this work. They have been included in in their original format and they contain introduction, materials and methods, results and discussion. All methods that were employed in these papers have been described in chapter 2. Finally, in chapter 6 I summarize all the results from this thesis, both locally for each of the scientific publications and globally for the whole work.
Resumo:
This article presents a new and computationally efficient method of analysis of a railway track modelled as a continuous beam of 2N spans supported by elastic vertical springs. The main feature of this method is its important reduction in computational effort with respect to standard matrix methods of structural analysis. In this article, the whole structure is considered to be a repetition of a single one. The analysis presented is applied to a simple railway track model, i.e. to a repetitive beam supported on vertical springs (sleepers). The proposed method of analysis is based on the general theory of spatially periodic structures. The main feature of this theory is the possibility to apply Discrete Fourier Transform (DFT) in order to reduce a large system of q(2N + 1) linear stiffness equilibrium equations to a set of 2N + 1 uncoupled systems of q equations each. In this way, a dramatic reduction of the computational effort of solving the large system of equations is achieved. This fact is particularly important in the analysis of railway track structures, in which N is a very large number (around several thousands), and q = 2, the vertical displacement and rotation, is very small. The proposed method allows us to easily obtain the exact solution given by Samartín [1], i.e. the continuous beam railway track response. The comparison between the proposed method and other methods of analysis of railway tracks, such as Lorente de Nó and Zimmermann-Timoshenko, clearly shows the accuracy of the obtained results for the proposed method, even for low values of N. In addition, identical results between the proposed and the Lorente methods have been found, although the proposed method seems to be of simpler application and computationally more efficient than the Lorente one. Small but significative differences occur between these two methods and the one developed by Zimmermann-Timoshenko. This article also presents a detailed sensitivity analysis of the vertical displacement of the sleepers. Although standard matrix methods of structural analysis can handle this railway model, one of the objectives of this article is to show the efficiency of DFT method with respect to standard matrix structural analysis. A comparative analysis between standard matrix structural analysis and the proposed method (DFT), in terms of computational time, input, output and also software programming, will be carried out. Finally, a URL link to a MatLab computer program list, based on the proposed method, is given
Resumo:
This paper discusses a model based on the agency theory to analyze the optimal transfer of construction risk in public works contracts. The base assumption is that of a contract between a principal (public authority) and an agent (firm), where the payment mechanism is linear and contains an incentive mechanism to enhance the effort of the agent to reduce construction costs. A theoretical model is proposed starting from a cost function with a random component and assuming that both the public authority and the firm are risk averse. The main outcome of the paper is that the optimal transfer of construction risk will be lower when the variance of errors in cost forecast, the risk aversion of the firm and the marginal cost of public funds are larger, while the optimal transfer of construction risk will grow when the variance of errors in cost monitoring and the risk aversion of the public authority are larger