959 resultados para Complexity analysis
Resumo:
Intercontinental Ballistic Missiles are capable of placing a nuclear warhead at more than 5,000 km away from its launching base. With the lethal power of a nuclear warhead a whole city could be wiped out by a single weapon causing millions of deaths. This means that the threat posed to any country from a single ICBM captured by a terrorist group or launched by a 'rogue' state is huge. This threat is increasing as more countries are achieving nuclear and advanced launcher capabilities. In order to suppress or at least reduce this threat the United States created the National Missile Defense System which involved, among other systems, the development of long-range interceptors whose aim is to destroy incoming ballistic missiles in their midcourse phase. The Ballistic Missile Defense is a high-profile topic that has been the focus of political controversy lately when the U.S. decided to expand the Ballistic Missile system to Europe, with the opposition of Russia. However the technical characteristics of this system are mostly unknown by the general public. The Interception of an ICBM using a long range Interceptor Missile as intended within the Ground-Based Missile Defense System by the American National Missile Defense (NMD) implies a series of problems of incredible complexity: - The incoming missile has to be detected almost immediately after launch. - The incoming missile has to be tracked along its trajectory with a great accuracy. - The Interceptor Missile has to implement a fast and accurate guidance algorithm in order to reach the incoming missile as soon as possible. - The Kinetic Kill Vehicle deployed by the interceptor boost vehicle has to be able to detect the reentry vehicle once it has been deployed by ICBM, when it offers a very low infrared signature, in order to perform a final rendezvous manoeuvre. - The Kinetic Kill Vehicle has to be able to discriminate the reentry vehicle from the surrounding debris and decoys. - The Kinetic Kill Vehicle has to be able to implement an accurate guidance algorithm in order to perform a kinetic interception (direct collision) of the reentry vehicle, at relative speeds of more than 10 km/s. All these problems are being dealt simultaneously by the Ground-Based Missile Defense System that is developing very complex and expensive sensors, communications and control centers and long-range interceptors (Ground-Based Interceptor Missile) including a Kinetic Kill Vehicle. Among all the technical challenges involved in this interception scenario, this thesis focuses on the algorithms required for the guidance of the Interceptor Missile and the Kinetic Kill Vehicle in order to perform the direct collision with the ICBM. The involved guidance algorithms are deeply analysed in this thesis in part III where conventional guidance strategies are reviewed and optimal guidance algorithms are developed for this interception problem. The generation of a realistic simulation of the interception scenario between an ICBM and a Ground Based Interceptor designed to destroy it was considered as necessary in order to be able to compare different guidance strategies with meaningful results. As a consequence, a highly representative simulator for an ICBM and a Kill Vehicle has been implemented, as detailed in part II, and the generation of these simulators has also become one of the purposes of this thesis. In summary, the main purposes of this thesis are: - To develop a highly representative simulator of an interception scenario between an ICBM and a Kill Vehicle launched from a Ground Based Interceptor. -To analyse the main existing guidance algorithms both for the ascent phase and the terminal phase of the missiles. Novel conclusions of these analyses are obtained. - To develop original optimal guidance algorithms for the interception problem. - To compare the results obtained using the different guidance strategies, assess the behaviour of the optimal guidance algorithms, and analyse the feasibility of the Ballistic Missile Defense system in terms of guidance (part IV). As a secondary objective, a general overview of the state of the art in terms of ballistic missiles and anti-ballistic missile defence is provided (part I).
Resumo:
La tecnología de las máquinas móviles autónomas ha sido objeto de una gran investigación y desarrollo en las últimas décadas. En muchas actividades y entornos, los robots pueden realizar operaciones que son duras, peligrosas o simplemente imposibles para los humanos. La exploración planetaria es un buen ejemplo de un entorno donde los robots son necesarios para realizar las tareas requeridas por los científicos. La reciente exploración de Marte con robots autónomos nos ha mostrado la capacidad de las nuevas tecnologías. Desde la invención de la rueda, que esta acertadamente considerado como el mayor invento en la historia del transporte humano, casi todos los vehículos para exploración planetaria han empleado las ruedas para su desplazamiento. Las nuevas misiones planetarias demandan maquinas cada vez mas complejas. En esta Tesis se propone un nuevo diseño de un robot con patas o maquina andante que ofrecerá claras ventajas en entornos extremos. Se demostrara que puede desplazarse en los terrenos donde los robots con ruedas son ineficientes, convirtiéndolo en una elección perfecta para misiones planetarias. Se presenta una reseña histórica de los principales misiones espaciales, en particular aquellos dirigidos a la exploración planetaria. A través de este estudio será posible analizar las desventajas de los robots con ruedas utilizados en misiones anteriores. El diseño propuesto de robot con patas será presentado como una alternativa para aquellas misiones donde los robots con ruedas puedan no ser la mejor opción. En esta tesis se presenta el diseño mecánico de un robot de seis patas capaz de soportar las grandes fuerzas y momentos derivadas del movimiento de avance. Una vez concluido el diseño mecánico es necesario realizar un análisis que permita entender el movimiento y comportamiento de una maquina de esta complejidad. Las ecuaciones de movimiento del robot serán validadas por dos métodos: cinemático y dinámico. Dos códigos Matlab® han sido desarrollados para resolver dichos sistemas de ecuaciones y han sido verificados por un tercer método, un modelo de elementos finitos, que también verifica el diseño mecánico. El robot con patas presentado, ha sido diseñado para la exploración planetaria en Marte. El comportamiento del robot durante sus desplazamientos será probado mediante un código de Matlab®, desarrollado para esta tesis, que permite modificar las trayectorias, el tipo de terreno, y el número y altura de los obstáculos. Estos terrenos y requisitos iniciales no han sido elegidos de forma aleatoria, si no que están basados en mi experiencia como miembro del equipo de MSL-NASA que opera un instrumento a bordo del rover Curiosity en Marte. El robot con patas desarrollado y fabricado por el Centro de Astrobiología (INTA-CSIC), esta basado en el diseño mecánico y análisis presentados en esta tesis. ABSTRACT The autonomous machines technology has undergone a major research and development during the last decades. In many activities and environments, robots can perform operations that are tought, dangerous or simply imposible to humans. Planetary exploration is a good example of such environment where robots are needed to perform the tasks required by the scientits. Recent Mars exploration based on autonomous vehicles has shown us the capacity of the new technologies. From the invention of the wheel, which is rightly regarded as the greatest invention in the history of human transportation, nearly all-planetary vehicles are based in wheeled locomotion, but new missions demand new types of machines due to the complex tasks needed to be performed. It will be proposed in this thesis a new design of a legged robot or walking machine, which may offer clear advantages in tough environments. This Thesis will show that the proposed walking machine can travel, were terrain difficulties make wheeled vehicles ineffective, making it a perfect choice for planetary mission. A historical background of the main space missions, in particular those aimed at planetary exploration will be presented. From this study the disadvantages found in the existing wheel rovers will be analysed. The legged robot designed will be introduced as an alternative were wheeled rovers could be no longer the best option for planetary exploration. This thesis introduces the mechanical design of a six-leg robot capable of withstanding high forces and moments due to the walking motion. Once the mechanical design is concluded, and in order to analyse a machine of this complexity an understanding of its movement and behaviour is mandatory. This movement equation will be validated by two methods: kinematics and dynamics. Two Matlab® codes have been developed to solve the systems of equations and validated by a third method, a finite element model, which also verifies the mechanical design. The legged robot presented has been designed for a Mars planetary exploration. The movement behaviour of the robot will be tested in a Matlab® code developed that allows to modify the trajectories, the type of terrain, number and height of obstacles. These terrains and initial requirements have not been chosen randomly, those are based on my experience as a member of the MSL NASA team, which operates an instrument on-board of the Curiosity rover in Mars. The walking robot developed and manufactured by the Center of Astrobiology (CAB) is based in the mechanical design and analysis that will be presented in this thesis.
Resumo:
PURPOSE The decision-making process plays a key role in organizations. Every decision-making process produces a final choice that may or may not prompt action. Recurrently, decision makers find themselves in the dichotomous question of following a traditional sequence decision-making process where the output of a decision is used as the input of the next stage of the decision, or following a joint decision-making approach where several decisions are taken simultaneously. The implication of the decision-making process will impact different players of the organization. The choice of the decision- making approach becomes difficult to find, even with the current literature and practitioners’ knowledge. The pursuit of better ways for making decisions has been a common goal for academics and practitioners. Management scientists use different techniques and approaches to improve different types of decisions. The purpose of this decision is to use the available resources as well as possible (data and techniques) to achieve the objectives of the organization. The developing and applying of models and concepts may be helpful to solve managerial problems faced every day in different companies. As a result of this research different decision models are presented to contribute to the body of knowledge of management science. The first models are focused on the manufacturing industry and the second part of the models on the health care industry. Despite these models being case specific, they serve the purpose of exemplifying that different approaches to the problems and could provide interesting results. Unfortunately, there is no universal recipe that could be applied to all the problems. Furthermore, the same model could deliver good results with certain data and bad results for other data. A framework to analyse the data before selecting the model to be used is presented and tested in the models developed to exemplify the ideas. METHODOLOGY As the first step of the research a systematic literature review on the joint decision is presented, as are the different opinions and suggestions of different scholars. For the next stage of the thesis, the decision-making process of more than 50 companies was analysed in companies from different sectors in the production planning area at the Job Shop level. The data was obtained using surveys and face-to-face interviews. The following part of the research into the decision-making process was held in two application fields that are highly relevant for our society; manufacturing and health care. The first step was to study the interactions and develop a mathematical model for the replenishment of the car assembly where the problem of “Vehicle routing problem and Inventory” were combined. The next step was to add the scheduling or car production (car sequencing) decision and use some metaheuristics such as ant colony and genetic algorithms to measure if the behaviour is kept up with different case size problems. A similar approach is presented in a production of semiconductors and aviation parts, where a hoist has to change from one station to another to deal with the work, and a jobs schedule has to be done. However, for this problem simulation was used for experimentation. In parallel, the scheduling of operating rooms was studied. Surgeries were allocated to surgeons and the scheduling of operating rooms was analysed. The first part of the research was done in a Teaching hospital, and for the second part the interaction of uncertainty was added. Once the previous problem had been analysed a general framework to characterize the instance was built. In the final chapter a general conclusion is presented. FINDINGS AND PRACTICAL IMPLICATIONS The first part of the contributions is an update of the decision-making literature review. Also an analysis of the possible savings resulting from a change in the decision process is made. Then, the results of the survey, which present a lack of consistency between what the managers believe and the reality of the integration of their decisions. In the next stage of the thesis, a contribution to the body of knowledge of the operation research, with the joint solution of the replenishment, sequencing and inventory problem in the assembly line is made, together with a parallel work with the operating rooms scheduling where different solutions approaches are presented. In addition to the contribution of the solving methods, with the use of different techniques, the main contribution is the framework that is proposed to pre-evaluate the problem before thinking of the techniques to solve it. However, there is no straightforward answer as to whether it is better to have joint or sequential solutions. Following the proposed framework with the evaluation of factors such as the flexibility of the answer, the number of actors, and the tightness of the data, give us important hints as to the most suitable direction to take to tackle the problem. RESEARCH LIMITATIONS AND AVENUES FOR FUTURE RESEARCH In the first part of the work it was really complicated to calculate the possible savings of different projects, since in many papers these quantities are not reported or the impact is based on non-quantifiable benefits. The other issue is the confidentiality of many projects where the data cannot be presented. For the car assembly line problem more computational power would allow us to solve bigger instances. For the operation research problem there was a lack of historical data to perform a parallel analysis in the teaching hospital. In order to keep testing the decision framework it is necessary to keep applying more case studies in order to generalize the results and make them more evident and less ambiguous. The health care field offers great opportunities since despite the recent awareness of the need to improve the decision-making process there are many opportunities to improve. Another big difference with the automotive industry is that the last improvements are not spread among all the actors. Therefore, in the future this research will focus more on the collaboration between academia and the health care sector.
Resumo:
An important aspect of Process Simulators for photovoltaics is prediction of defect evolution during device fabrication. Over the last twenty years, these tools have accelerated process optimization, and several Process Simulators for iron, a ubiquitous and deleterious impurity in silicon, have been developed. The diversity of these tools can make it difficult to build intuition about the physics governing iron behavior during processing. Thus, in one unified software environment and using self-consistent terminology, we combine and describe three of these Simulators. We vary structural defect distribution and iron precipitation equations to create eight distinct Models, which we then use to simulate different stages of processing. We find that the structural defect distribution influences the final interstitial iron concentration ([Fe-i]) more strongly than the iron precipitation equations. We identify two regimes of iron behavior: (1) diffusivity-limited, in which iron evolution is kinetically limited and bulk [Fe-i] predictions can vary by an order of magnitude or more, and (2) solubility-limited, in which iron evolution is near thermodynamic equilibrium and the Models yield similar results. This rigorous analysis provides new intuition that can inform Process Simulation, material, and process development, and it enables scientists and engineers to choose an appropriate level of Model complexity based on wafer type and quality, processing conditions, and available computation time.
Resumo:
Chemical process accidents still occur and cost billions of dollars and, what is worse, many human lives. That means that traditional hazard analysis techniques are not enough mainly owing to the increase of complexity and size of chemical plants. In the last years, a new hazard analysis technique has been developed, changing the focus from reliability to system theory and showing promising results in other industries such as aeronautical and nuclear. In this paper, we present an approach for the application of STAMP and STPA analysis developed by Leveson in 2011 to the process industry.
Resumo:
Increasingly, studies of genes and genomes are indicating that considerable horizontal transfer has occurred between prokaryotes. Extensive horizontal transfer has occurred for operational genes (those involved in housekeeping), whereas informational genes (those involved in transcription, translation, and related processes) are seldomly horizontally transferred. Through phylogenetic analysis of six complete prokaryotic genomes and the identification of 312 sets of orthologous genes present in all six genomes, we tested two theories describing the temporal flow of horizontal transfer. We show that operational genes have been horizontally transferred continuously since the divergence of the prokaryotes, rather than having been exchanged in one, or a few, massive events that occurred early in the evolution of prokaryotes. In agreement with earlier studies, we found that differences in rates of evolution between operational and informational genes are minimal, suggesting that factors other than rate of evolution are responsible for the observed differences in horizontal transfer. We propose that a major factor in the more frequent horizontal transfer of operational genes is that informational genes are typically members of large, complex systems, whereas operational genes are not, thereby making horizontal transfer of informational gene products less probable (the complexity hypothesis).
Resumo:
A rapidly growing area of genome research is the generation of expressed sequence tags (ESTs) in which large numbers of randomly selected cDNA clones are partially sequenced. The collection of ESTs reflects the level and complexity of gene expression in the sampled tissue. To date, the majority of plant ESTs are from nonwoody plants such as Arabidopsis, Brassica, maize, and rice. Here, we present a large-scale production of ESTs from the wood-forming tissues of two poplars, Populus tremula L. × tremuloides Michx. and Populus trichocarpa ‘Trichobel.’ The 5,692 ESTs analyzed represented a total of 3,719 unique transcripts for the two cDNA libraries. Putative functions could be assigned to 2,245 of these transcripts that corresponded to 820 protein functions. Of specific interest to forest biotechnology are the 4% of ESTs involved in various processes of cell wall formation, such as lignin and cellulose synthesis, 5% similar to developmental regulators and members of known signal transduction pathways, and 2% involved in hormone biosynthesis. An additional 12% of the ESTs showed no significant similarity to any other DNA or protein sequences in existing databases. The absence of these sequences from public databases may indicate a specific role for these proteins in wood formation. The cDNA libraries and the accompanying database are valuable resources for forest research directed toward understanding the genetic control of wood formation and future endeavors to modify wood and fiber properties for industrial use.
Resumo:
Analysis of previously published sets of DNA microarray gene expression data by singular value decomposition has uncovered underlying patterns or “characteristic modes” in their temporal profiles. These patterns contribute unequally to the structure of the expression profiles. Moreover, the essential features of a given set of expression profiles are captured using just a small number of characteristic modes. This leads to the striking conclusion that the transcriptional response of a genome is orchestrated in a few fundamental patterns of gene expression change. These patterns are both simple and robust, dominating the alterations in expression of genes throughout the genome. Moreover, the characteristic modes of gene expression change in response to environmental perturbations are similar in such distant organisms as yeast and human cells. This analysis reveals simple regularities in the seemingly complex transcriptional transitions of diverse cells to new states, and these provide insights into the operation of the underlying genetic networks.
Resumo:
The proteasome is a large protease complex consisting of multiple catalytic subunits that function simultaneously to digest protein substrates. This complexity has made deciphering the role each subunit plays in the generation of specific protein fragments difficult. Positional scanning libraries of peptide vinyl sulfones were generated in which the amino acid located directly at the site of hydrolysis (P1 residue) was held constant and sequences distal to that residue (P2, P3, and P4 positions) were varied across all natural amino acids (except cysteine and methionine). Binding information for each of the individual catalytic subunits was obtained for each library under a variety of different conditions. The resulting specificity profiles indicated that substrate positions distal to P1 are critical for directing substrates to active subunits in the complex. Furthermore, specificity profiles of IFN-γ-regulated subunits closely matched those of their noninducible counterparts, suggesting that subunit swapping may modulate substrate processing by a mechanism that does require a change in the primary sequence specificity of individual catalytic subunits in the complex. Finally, specificity profiles were used to design specific inhibitors of a single active site in the complex. These reagents can be used to further establish the role of each subunit in substrate processing by the proteasome.
Resumo:
Genetic analysis of plant–pathogen interactions has demonstrated that resistance to infection is often determined by the interaction of dominant plant resistance (R) genes and dominant pathogen-encoded avirulence (Avr) genes. It was postulated that R genes encode receptors for Avr determinants. A large number of R genes and their cognate Avr genes have now been analyzed at the molecular level. R gene loci are extremely polymorphic, particularly in sequences encoding amino acids of the leucine-rich repeat motif. A major challenge is to determine how Avr perception by R proteins triggers the plant defense response. Mutational analysis has identified several genes required for the function of specific R proteins. Here we report the identification of Rcr3, a tomato gene required specifically for Cf-2-mediated resistance. We propose that Avr products interact with host proteins to promote disease, and that R proteins “guard” these host components and initiate Avr-dependent plant defense responses.
Resumo:
Complexity originates from the tendency of large dynamical systems to organize themselves into a critical state, with avalanches or "punctuations" of all sizes. In the critical state, events which would otherwise be uncoupled become correlated. The apparent, historical contingency in many sciences, including geology, biology, and economics, finds a natural interpretation as a self-organized critical phenomenon. These ideas are discussed in the context of simple mathematical models of sandpiles and biological evolution. Insights are gained not only from numerical simulations but also from rigorous mathematical analysis.
Resumo:
Replication-incompetent retroviral vectors encoding histochemical reporter genes have been used for studying lineal relationships in a variety of species. A crucial element in the interpretation of data generated by this method is the identification of sibling relationships, or clonal boundaries. The use of a library of viruses in which each member is unique can greatly facilitate this aspect of the analysis. A previously reported murine retroviral library containing about 80 members demonstrated the utility of the library approach. However, the relatively low number of tags in the murine library necessitated using low infection rates in order to give confidence in clonal assignments. To obviate the need for low infection rates, a far more complex library was created and characterized. The CHAPOL library was constructed such that each member encodes a histochemical reporter gene and has a DNA tag derived from a degenerate oligonucleotide pool synthesized to have a complexity of > 1 x 10(7). The library was tested after infection of cells in vitro or in vivo. The DNA tag from each histochemically labeled cell or clone of cells was recovered by PCR and sequenced for unambiguous identification. Three hundred and twenty tags have been identified after infection, and so far no tag has been seen to result from more than one independent infection. Thus, an equal distribution of inserts is suggested, and Monte Carlo analysis predicts a complexity of > 10(4) members.
Resumo:
The Vernacular Discourse of the "Arab Spring" is a project that bridges the divide between the East and the West by offering new readings to Arab subjectivities. Through an analysis of the "Arab Spring" through the lens of vernacular discourse, it challenges the Euro-Americo-centric legacies of Orientalism in Western academia and the new wave of extremism in the Arab world by offering alternative representations of Arab bodies and subjectivities. To offer this new reading of the "Arab Spring," it explores the foundations of critical rhetoric as a theory and a practice and argues for a turn towards a critical vernacular discourse. The turn towards critical vernacular discourse is important as it urges the analyses of different artifacts produced by marginalized groups in order to understand their perspectives that have largely been foreclosed in traditional cultural studies research. Building on embodied/performative critical rhetoric, the vernacular discourses of the Arab revolutionary body examines other forms of knowledge productions that are not merely textual; more specifically, through data gathered in the Lhbib Bourguiba, Tunisia. This analysis of the political revolutionary body unveils the complexity underlining the discussion around issues of identity, agency and representation in the Middle East and North Africa, and calls for a critical study towards these issues in the region beyond the binary approach that has been practiced and applied by academics and media analysts. Hence, by analyzing vernacular discourse, this research locates a method of examining and theorizing the dialectic between agency, citizenry, and subjectivity through the study of how power structure is recreated and challenged through the use of the vernacular in revolutionary movements, as well as how marginalized groups construct their own subjectivities through the use of vernacular discourse. Therefore, highlighting the political prominence of evaluating the Arab Spring as a vernacular discourse is important in creating new ways of understanding communication in postcolonial/neocolonial settings.
Resumo:
Paper submitted to the 44th European Congress of the European Regional Science Association, Porto, 25-29 August 2004.
Resumo:
Aerobic Gymnastic is the ability to perform complex movements produced by the traditional aerobic exercises, in a continuous manner, with high intensity, perfectly integrated with soundtracks. This sport is performed in an aerobic/anaerobic lactacid condition and expects the execution of complex movements produced by the traditional aerobic exercises integrated with difficulty elements performed with a high technical level. An inaccuracy about this sport is related to the name itself “aerobic” because Aerobic Gymnastic does not use just the aerobic work during the competition, due to the fact that the exercises last among 1’30” and 1’45” at high rhythm. Agonistic Aerobics exploit the basic movements of amateur Aerobics and its coordination schemes, even though the agonistic Aerobics is so much intense than the amateur Aerobics to need a completely different mix of energetic mechanisms. Due to the complexity and the speed with which you perform the technical elements of Aerobic Gymnastic, the introduction of video analysis is essential for a qualitative and quantitative evaluation of athletes’ performance during the training. The performance analysis can allow the accurate analysis and explanation of the evolution and dynamics of a historical phenomenon and motor sports. The notational analysis is used by technicians to have an objective analysis of performance. Tactics, technique and individual movements can be analyzed to help coaches and athletes to re-evaluate their performance and gain advantage during the competition. The purpose of the following experimental work will be a starting point for analyzing the performance of the athletes in an objective way, not only during competitions, but especially during the phases of training. It is, therefore, advisable to introduce the video analysis and notational analysis for more quantitative and qualitative examination of technical movements. The goal is to lead to an improvement of the technique of the athlete and the teaching of the coach.