908 resultados para Real systems
Resumo:
Os coeficientes de difusão (D 12) são propriedades fundamentais na investigação e na indústria, mas a falta de dados experimentais e a inexistência de equações que os estimem com precisão e confiança em fases comprimidas ou condensadas constituem limitações importantes. Os objetivos principais deste trabalho compreendem: i) a compilação de uma grande base de dados para valores de D 12 de sistemas gasosos, líquidos e supercríticos; ii) o desenvolvimento e validação de novos modelos de coeficientes de difusão a diluição infinita, aplicáveis em amplas gamas de temperatura e densidade, para sistemas contendo componentes muito distintos em termos de polaridade, tamanho e simetria; iii) a montagem e teste de uma instalação experimental para medir coeficientes de difusão em líquidos e fluidos supercríticos. Relativamente à modelação, uma nova expressão para coeficientes de difusão a diluição infinita de esferas rígidas foi desenvolvida e validada usando dados de dinâmica molecular (desvio relativo absoluto médio, AARD = 4.44%) Foram também estudados os coeficientes de difusão binários de sistemas reais. Para tal, foi compilada uma extensa base de dados de difusividades de sistemas reais em gases e solventes densos (622 sistemas binários num total de 9407 pontos experimentais e 358 moléculas) e a mesma foi usada na validação dos novos modelos desenvolvidos nesta tese. Um conjunto de novos modelos foi proposto para o cálculo de coeficientes de difusão a diluição infinita usando diferentes abordagens: i) dois modelos de base molecular com um parâmetro específico para cada sistema, aplicáveis em sistemas gasosos, líquidos e supercríticos, em que natureza do solvente se encontra limitada a apolar ou fracamente polar (AARDs globais na gama 4.26-4.40%); ii) dois modelos de base molecular biparamétricos, aplicáveis em todos os estados físicos, para qualquer tipo de soluto diluído em qualquer solvente (apolar, fracamente polar e polar). Ambos os modelos dão origem a erros globais entre 2.74% e 3.65%; iii) uma correlação com um parâmetro, específica para coeficientes de difusão em dióxido de carbono supercrítico (SC-CO2) e água líquida (AARD = 3.56%); iv) nove correlações empíricas e semi-empíricas que envolvem dois parâmetros, dependentes apenas da temperatura e/ou densidade do solvente e/ou viscosidade do solvente. Estes últimos modelos são muito simples e exibem excelentes resultados (AARDs entre 2.78% e 4.44%) em sistemas líquidos e supercríticos; e v) duas equações preditivas para difusividades de solutos em SC-CO2, em que os erros globais de ambas são inferiores a 6.80%. No global, deve realçar-se o facto de os novos modelos abrangerem a grande variedade de sistemas e moléculas geralmente encontrados. Os resultados obtidos são consistentemente melhores do que os obtidos com os modelos e abordagens encontrados na literatura. No caso das correlações com um ou dois parâmetros, mostrou-se que estes mesmos parâmetros podem ser ajustados usando um conjunto muito pequeno de dados, e posteriormente serem utilizados na previsão de valores de D 12 longe do conjunto original de pontos. Uma nova instalação experimental para medir coeficientes de difusão binários por técnicas cromatográficas foi montada e testada. O equipamento, o procedimento experimental e os cálculos analíticos necessários à obtenção dos valores de D 12 pelo método de abertura do pico cromatográfico, foram avaliados através da medição de difusividades de tolueno e acetona em SC-CO2. Seguidamente, foram medidos coeficientes de difusão de eucaliptol em SC-CO2 nas gamas de 202 – 252 bar e 313.15 – 333.15 K. Os resultados experimentais foram analisados através de correlações e modelos preditivos para D12.
Resumo:
The work presented in this Ph.D thesis was developed in the context of complex network theory, from a statistical physics standpoint. We examine two distinct problems in this research field, taking a special interest in their respective critical properties. In both cases, the emergence of criticality is driven by a local optimization dynamics. Firstly, a recently introduced class of percolation problems that attracted a significant amount of attention from the scientific community, and was quickly followed up by an abundance of other works. Percolation transitions were believed to be continuous, until, recently, an 'explosive' percolation problem was reported to undergo a discontinuous transition, in [93]. The system's evolution is driven by a metropolis-like algorithm, apparently producing a discontinuous jump on the giant component's size at the percolation threshold. This finding was subsequently supported by number of other experimental studies [96, 97, 98, 99, 100, 101]. However, in [1] we have proved that the explosive percolation transition is actually continuous. The discontinuity which was observed in the evolution of the giant component's relative size is explained by the unusual smallness of the corresponding critical exponent, combined with the finiteness of the systems considered in experiments. Therefore, the size of the jump vanishes as the system's size goes to infinity. Additionally, we provide the complete theoretical description of the critical properties for a generalized version of the explosive percolation model [2], as well as a method [3] for a precise calculation of percolation's critical properties from numerical data (useful when exact results are not available). Secondly, we study a network flow optimization model, where the dynamics consists of consecutive mergings and splittings of currents flowing in the network. The current conservation constraint does not impose any particular criterion for the split of current among channels outgoing nodes, allowing us to introduce an asymmetrical rule, observed in several real systems. We solved analytically the dynamic equations describing this model in the high and low current regimes. The solutions found are compared with numerical results, for the two regimes, showing an excellent agreement. Surprisingly, in the low current regime, this model exhibits some features usually associated with continuous phase transitions.
Resumo:
This paper studies a discrete dynamical system of interacting particles that evolve by interacting among them. The computational model is an abstraction of the natural world, and real systems can range from the huge cosmological scale down to the scale of biological cell, or even molecules. Different conditions for the system evolution are tested. The emerging patterns are analysed by means of fractal dimension and entropy measures. It is observed that the population of particles evolves towards geometrical objects with a fractal nature. Moreover, the time signature of the entropy can be interpreted at the light of complex dynamical systems.
Resumo:
One of the techniques used to detect faults in dynamic systems is analytical redundancy. An important difficulty in applying this technique to real systems is dealing with the uncertainties associated with the system itself and with the measurements. In this paper, this uncertainty is taken into account by the use of intervals for the parameters of the model and for the measurements. The method that is proposed in this paper checks the consistency between the system's behavior, obtained from the measurements, and the model's behavior; if they are inconsistent, then there is a fault. The problem of detecting faults is stated as a quantified real constraint satisfaction problem, which can be solved using the modal interval analysis (MIA). MIA is used because it provides powerful tools to extend the calculations over real functions to intervals. To improve the results of the detection of the faults, the simultaneous use of several sliding time windows is proposed. The result of implementing this method is semiqualitative tracking (SQualTrack), a fault-detection tool that is robust in the sense that it does not generate false alarms, i.e., if there are false alarms, they indicate either that the interval model does not represent the system adequately or that the interval measurements do not represent the true values of the variables adequately. SQualTrack is currently being used to detect faults in real processes. Some of these applications using real data have been developed within the European project advanced decision support system for chemical/petrochemical manufacturing processes and are also described in this paper
Resumo:
A simple and practical technique for assessing the risks, that is, the potential for error, and consequent loss, in software system development, acquired during a requirements engineering phase is described. The technique uses a goal-based requirements analysis as a framework to identify and rate a set of key issues in order to arrive at estimates of the feasibility and adequacy of the requirements. The technique is illustrated and how it has been applied to a real systems development project is shown. How problems in this project could have been identified earlier is shown, thereby avoiding costly additional work and unhappy users.
Resumo:
A simple and practical technique for assessing the risks, that is, the potential for error, and consequent loss, in software system development, acquired during a requirements engineering phase is described. The technique uses a goal-based requirements analysis as a framework to identify and rate a set of key issues in order to arrive at estimates of the feasibility and adequacy of the requirements. The technique is illustrated and how it has been applied to a real systems development project is shown. How problems in this project could have been identified earlier is shown, thereby avoiding costly additional work and unhappy users.
Resumo:
Context: During development managers, analysts and designers often need to know whether enough requirements analysis work has been done and whether or not it is safe to proceed to the design stage. Objective: This paper describes a new, simple and practical method for assessing our confidence in a set of requirements. Method: We identified 4 confidence factors and used a goal oriented framework with a simple ordinal scale to develop a method for assessing confidence. We illustrate the method and show how it has been applied to a real systems development project. Results: We show how assessing confidence in the requirements could have revealed problems in this project earlier and so saved both time and money. Conclusion: Our meta-level assessment of requirements provides a practical and pragmatic method that can prove useful to managers, analysts and designers who need to know when sufficient requirements analysis has been performed.
Resumo:
Complex networks exist in many areas of science such as biology, neuroscience, engineering, and sociology. The growing development of this area has led to the introduction of several topological and dynamical measurements, which describe and quantify the structure of networks. Such characterization is essential not only for the modeling of real systems but also for the study of dynamic processes that may take place in them. However, it is not easy to use several measurements for the analysis of complex networks, due to the correlation between them and the difficulty of their visualization. To overcome these limitations, we propose an effective and comprehensive approach for the analysis of complex networks, which allows the visualization of several measurements in a few projections that contain the largest data variance and the classification of networks into three levels of detail, vertices, communities, and the global topology. We also demonstrate the efficiency and the universality of the proposed methods in a series of real-world networks in the three levels.
Resumo:
This work presents a fully operational interstate CGE model implemented for the Brazilian economy that tries to quantify both the role of barriers to trade on economic growth and foreign trade performance and how the distribution of the economic activity may change as the country opens up to foreign trade. Among the distinctive features embedded in the model, modeling of external scale economies, port efficiency and land-maritime transport costs provides an innovative way of dealing explicitly with theoretical issues related to integrated regional systems. In order to illustrate the role played by the quality of infrastructure and geography on the country‟s foreign and interregional trade performance, a set of simulations is presented where barriers to trade are significantly reduced. The relative importance of trade policy, port efficiency and land-maritime transport costs for the country trade relations and regional growth is then detailed and quantified, considering both short run as well as long run scenarios. A final set of simulations shed some light on the effects of liberal trade policies on regional inequality, where the manufacturing sector in the state of São Paulo, taken as the core of industrial activity in the country, is subjected to different levels of external economies of scale. Short-run core-periphery effects are then traced out suggesting the prevalence of agglomeration forces over diversion forces could rather exacerbate regional inequality as import barriers are removed up to a certain level. Further removals can reverse this balance in favor of diversion forces, implying de-concentration of economic activity. In the long run, factor mobility allows a better characterization of the balance between agglomeration and diversion forces among regions. Regional dispersion effects are then clearly traced-out, suggesting horizontal liberal trade policies to benefit both the poorest regions in the country as well as the state of São Paulo. This long run dispersion pattern, on one hand seems to unravel the fragility of simple theoretical results from recent New Economic Geography models, once they get confronted with more complex spatially heterogeneous (real) systems. On the other hand, it seems to capture the literature‟s main insight: the possible role of horizontal liberal trade policies as diversion forces leading to a more homogeneous pattern of interregional economic growth.
Resumo:
Este trabalho é um estudo sobre a aplicação de técnicas de visualização de informação em sistemas de software científico, i.e., sistemas de software voltados para matemática, ciências e engenharias. Enquanto sistemas dessa natureza normalmente fazem uso da visualização científica e figuram como caso de sucesso nessa área, nem sempre são projetados considerando os princípios de visualização de informação. Esse trabalho tem por objetivo avaliar a aplicação desse conceitos em alguns sistemas reais, desenvolvidos com diferentes níveis de entendimento nessa área de conhecimento e com base nessas experiências propor o desenvolvimento de componentes de software capazes de facilitar a criação de sistemas semelhantes e ao mesmo tempo promover a aplicação destes conceitos.
Resumo:
The present study aims to analyse, in different levels of demand, what is the best layout strategy to adopt for the small metallic shipbuilding. To achieve this purpose, three simulation models are developed for analyze these production strategies under the positional, cellular and linear layouts. By the use of a simulation tool for compare the scenarios, Chwif and Medina (2010) and Law (2009)´s methodologies were adapted that includes three phases: conception, implementation and analysis. In conception real systems were represented by process mapping according to time, material resources and human resources variables required for each step of the production process. All of this information has been transformed in the cost variable. Data were collected from three different production systems, two located in Natal RN with cellular and positional layouts and one located in Belém-PA with linear layout. In the implementation phase, the conceptual models were converted in computacional models through the tool Rockwell Software Arena ® 13.5 and then validated. In the analysis stage the production of 960 ships in a year vessels were simulated for each layout noting that, for a production of until 80 units positional layout is the most recommended, between 81 and 288 units the cellular layout and more than 289 units the linear layout
Resumo:
The Predictive Controller has been receiving plenty attention in the last decades, because the need to understand, to analyze, to predict and to control real systems has been quickly growing with the technological and industrial progress. The objective of this thesis is to present a contribution for the development and implementation of Nonlinear Predictive Controllers based on Hammerstein model, as well as to its make properties evaluation. In this case, in the Nonlinear Predictive Controller development the time-step linearization method is used and a compensation term is introduced in order to improve the controller performance. The main motivation of this thesis is the study and stability guarantee for the Nonlinear Predictive Controller based on Hammerstein model. In this case, was used the concepts of sections and Popov Theorem. Simulation results with literature models shows that the proposed approaches are able to control with good performance and to guarantee the systems stability
Resumo:
In this Thesis, the development of the dynamic model of multirotor unmanned aerial vehicle with vertical takeoff and landing characteristics, considering input nonlinearities and a full state robust backstepping controller are presented. The dynamic model is expressed using the Newton-Euler laws, aiming to obtain a better mathematical representation of the mechanical system for system analysis and control design, not only when it is hovering, but also when it is taking-off, or landing, or flying to perform a task. The input nonlinearities are the deadzone and saturation, where the gravitational effect and the inherent physical constrains of the rotors are related and addressed. The experimental multirotor aerial vehicle is equipped with an inertial measurement unit and a sonar sensor, which appropriately provides measurements of attitude and altitude. A real-time attitude estimation scheme based on the extended Kalman filter using quaternions was developed. Then, for robustness analysis, sensors were modeled as the ideal value with addition of an unknown bias and unknown white noise. The bounded robust attitude/altitude controller were derived based on globally uniformly practically asymptotically stable for real systems, that remains globally uniformly asymptotically stable if and only if their solutions are globally uniformly bounded, dealing with convergence and stability into a ball of the state space with non-null radius, under some assumptions. The Lyapunov analysis technique was used to prove the stability of the closed-loop system, compute bounds on control gains and guaranteeing desired bounds on attitude dynamics tracking errors in the presence of measurement disturbances. The controller laws were tested in numerical simulations and in an experimental hexarotor, developed at the UFRN Robotics Laboratory
Resumo:
Fuzzy intelligent systems are present in a variety of equipment ranging from household appliances to Fuzzy intelligent systems are present in a variety of equipment ranging from household appliances to small devices such as digital cameras and cell phones being used primarily for dealing with the uncertainties in the modeling of real systems. However, commercial implementations of Fuzzy systems are not general purpose and do not have portability to different hardware platforms. Thinking about these issues this work presents the implementation of an open source development environment that consists of a desktop system capable of generate Graphically a general purpose Fuzzy controller and export these parameters for an embedded system with a Fuzzy controller written in Java Platform Micro Edition To (J2ME), whose modular design makes it portable to any mobile device that supports J2ME. Thus, the proposed development platform is capable of generating all the parameters of a Fuzzy controller and export it in XML file, and the code responsible for the control logic that is embedded in the mobile device is able to read this file and start the controller. All the parameters of a Fuzzy controller are configurable using the desktop system, since the membership functions and rule base, even the universe of discourse of the linguistic terms of output variables. This system generates Fuzzy controllers for the interpolation model of Takagi-Sugeno. As the validation process and testing of the proposed solution the Fuzzy controller was embedded on the mobile device Sun SPOT ® and used to control a plant-level Quanser®, and to compare the Fuzzy controller generated by the system with other types of controllers was implemented and embedded in sun spot a PID controller to control the same level plant of Quanser®
Resumo:
The present work has as objective to present a method of project and implementation of controllers PID, based on industrial instrumentation. An automatic system of auto-tunning of controllers PID will be presented, for systems of first and second order. The software presented in this work is applied in controlled plants by PID controllers implemented in a CLP. Software is applied to make the auto-tunning of the parameters of controller PID of plants that need this tunning. Software presents two stages, the first one is the stage of identification of the system using the least square recursive algorithm and the second is the stage of project of the parameters of controller PID using the root locus algorithm. An important fact of this work is the use of industrial instrumentation for the accomplishment of the experiments. The experiments had been carried through in controlled real plants for controllers PID implemented in the CLP. Thus has not only one resulted obtained with theoreticians experiments made with computational programs, and yes resulted obtained of real systems. The experiments had shown good results gotten with developed software