987 resultados para Dynamic problems


Relevância:

30.00% 30.00%

Publicador:

Resumo:

Technical analysis of Low Voltage Direct Current (LVDC) distribution systems shows that in LVDC transmission the customer voltage quality is higher. One of the problems in LVDC distribution networks that converters both ends of the DC line are required. Because of the converters produce not pure DC voltage, but some fluctuations as well, the huge electrolytic capacitors are required to reduce voltage distortions in the DC-side. This thesis master’s thesis is focused on calculating required DC-link capacitance for LVDC transmission and estimation of the influence of different parameters on the voltage quality. The goal is to investigate the methods of the DC-link capacitance estimation and location in the transmission line.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Today´s organizations must have the ability to react to rapid changes in the market. These rapid changes cause pressure to continuously find new efficient ways to organize work practices. Increased competition requires businesses to become more effective and to pay attention to quality of management and to make people to understand their work's impact on the final result. The fundamentals in continmuois improvement are systematic and agile tackling of indentified individual process constraints and the fact tha nothin finally improves without changes. Successful continuous improvement requires management commitment, education, implementation, measurement, recognition and regeneration. These ingredients form the foundation, both for breakthrough projects and small step ongoing improvement activities. One part of the organization's management system are the quality tools, which provide systematic methodologies for identifying problems, defining their root causes, finding solutions, gathering and sorting of data, supporting decision making and implementing the changes, and many other management tasks. Organizational change management includes processes and tools for managing the people in an organizational level change. These tools include a structured approach, which can be used for effective transition of organizations through change. When combined with the understanding of change management of individuals, these tools provide a framework for managing people in change,

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The objective of this dissertation is to improve the dynamic simulation of fluid power circuits. A fluid power circuit is a typical way to implement power transmission in mobile working machines, e.g. cranes, excavators etc. Dynamic simulation is an essential tool in developing controllability and energy-efficient solutions for mobile machines. Efficient dynamic simulation is the basic requirement for the real-time simulation. In the real-time simulation of fluid power circuits there exist numerical problems due to the software and methods used for modelling and integration. A simulation model of a fluid power circuit is typically created using differential and algebraic equations. Efficient numerical methods are required since differential equations must be solved in real time. Unfortunately, simulation software packages offer only a limited selection of numerical solvers. Numerical problems cause noise to the results, which in many cases leads the simulation run to fail. Mathematically the fluid power circuit models are stiff systems of ordinary differential equations. Numerical solution of the stiff systems can be improved by two alternative approaches. The first is to develop numerical solvers suitable for solving stiff systems. The second is to decrease the model stiffness itself by introducing models and algorithms that either decrease the highest eigenvalues or neglect them by introducing steady-state solutions of the stiff parts of the models. The thesis proposes novel methods using the latter approach. The study aims to develop practical methods usable in dynamic simulation of fluid power circuits using explicit fixed-step integration algorithms. In this thesis, twomechanisms whichmake the systemstiff are studied. These are the pressure drop approaching zero in the turbulent orifice model and the volume approaching zero in the equation of pressure build-up. These are the critical areas to which alternative methods for modelling and numerical simulation are proposed. Generally, in hydraulic power transmission systems the orifice flow is clearly in the turbulent area. The flow becomes laminar as the pressure drop over the orifice approaches zero only in rare situations. These are e.g. when a valve is closed, or an actuator is driven against an end stopper, or external force makes actuator to switch its direction during operation. This means that in terms of accuracy, the description of laminar flow is not necessary. But, unfortunately, when a purely turbulent description of the orifice is used, numerical problems occur when the pressure drop comes close to zero since the first derivative of flow with respect to the pressure drop approaches infinity when the pressure drop approaches zero. Furthermore, the second derivative becomes discontinuous, which causes numerical noise and an infinitely small integration step when a variable step integrator is used. A numerically efficient model for the orifice flow is proposed using a cubic spline function to describe the flow in the laminar and transition areas. Parameters for the cubic spline function are selected such that its first derivative is equal to the first derivative of the pure turbulent orifice flow model in the boundary condition. In the dynamic simulation of fluid power circuits, a tradeoff exists between accuracy and calculation speed. This investigation is made for the two-regime flow orifice model. Especially inside of many types of valves, as well as between them, there exist very small volumes. The integration of pressures in small fluid volumes causes numerical problems in fluid power circuit simulation. Particularly in realtime simulation, these numerical problems are a great weakness. The system stiffness approaches infinity as the fluid volume approaches zero. If fixed step explicit algorithms for solving ordinary differential equations (ODE) are used, the system stability would easily be lost when integrating pressures in small volumes. To solve the problem caused by small fluid volumes, a pseudo-dynamic solver is proposed. Instead of integration of the pressure in a small volume, the pressure is solved as a steady-state pressure created in a separate cascade loop by numerical integration. The hydraulic capacitance V/Be of the parts of the circuit whose pressures are solved by the pseudo-dynamic method should be orders of magnitude smaller than that of those partswhose pressures are integrated. The key advantage of this novel method is that the numerical problems caused by the small volumes are completely avoided. Also, the method is freely applicable regardless of the integration routine applied. The superiority of both above-mentioned methods is that they are suited for use together with the semi-empirical modelling method which necessarily does not require any geometrical data of the valves and actuators to be modelled. In this modelling method, most of the needed component information can be taken from the manufacturer’s nominal graphs. This thesis introduces the methods and shows several numerical examples to demonstrate how the proposed methods improve the dynamic simulation of various hydraulic circuits.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The aim of the study was to create and evaluate an intervention programme for Tanzanian children from a low-income area who are at risk of reading and writing difficulties. The learning difficulties, including reading and writing difficulties, are likely to be behind many of the common school problems in Tanzania, but they are not well understood, and research is needed. The design of the study included an identification and intervention phase with follow-up. A group based dynamic assessment approach was used in identifying children at risk of difficulties in reading and writing. The same approach was used in the intervention. The study was a randomized experiment with one experimental and two control groups. For the experimental and the control groups, a total of 96 (46 girls and 50 boys) children from grade one were screened out of 301 children from two schools in a low income urban area of Dar-es-Salaam. One third of the children, the experimental group, participated in an intensive training programme in literacy skills for five weeks, six hours per week, aimed at promoting reading and writing ability, while the children in the control groups had a mathematics and art programme. Follow-up was performed five months after the intervention. The intervention programme and the tests were based on the Zambian BASAT (Basic Skill Assessment Tool, Ketonen & Mulenga, 2003), but the content was drawn from the Kiswahili school curriculum in Tanzania. The main components of the training and testing programme were the same, only differing in content. The training process was different from traditional training in Tanzanian schools in that principles of teaching and training in dynamic assessment were followed. Feedback was the cornerstone of the training and the focus was on supporting the children in exploring knowledge and strategies in performing the tasks. The experimental group improved significantly more (p = .000) than the control groups during the intervention from pre-test to follow-up (repeated measures ANOVA). No differences between the control groups were noticed. The effect was significant on all the measures: phonological awareness, reading skills, writing skills and overall literacy skills. A transfer effect on school marks in Kiswahili and English was found. Following a discussion of the results, suggestions for further research and adaptation of the programme are presented.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Rapid ongoing evolution of multiprocessors will lead to systems with hundreds of processing cores integrated in a single chip. An emerging challenge is the implementation of reliable and efficient interconnection between these cores as well as other components in the systems. Network-on-Chip is an interconnection approach which is intended to solve the performance bottleneck caused by traditional, poorly scalable communication structures such as buses. However, a large on-chip network involves issues related to congestion problems and system control, for instance. Additionally, faults can cause problems in multiprocessor systems. These faults can be transient faults, permanent manufacturing faults, or they can appear due to aging. To solve the emerging traffic management, controllability issues and to maintain system operation regardless of faults a monitoring system is needed. The monitoring system should be dynamically applicable to various purposes and it should fully cover the system under observation. In a large multiprocessor the distances between components can be relatively long. Therefore, the system should be designed so that the amount of energy-inefficient long-distance communication is minimized. This thesis presents a dynamically clustered distributed monitoring structure. The monitoring is distributed so that no centralized control is required for basic tasks such as traffic management and task mapping. To enable extensive analysis of different Network-on-Chip architectures, an in-house SystemC based simulation environment was implemented. It allows transaction level analysis without time consuming circuit level implementations during early design phases of novel architectures and features. The presented analysis shows that the dynamically clustered monitoring structure can be efficiently utilized for traffic management in faulty and congested Network-on-Chip-based multiprocessor systems. The monitoring structure can be also successfully applied for task mapping purposes. Furthermore, the analysis shows that the presented in-house simulation environment is flexible and practical tool for extensive Network-on-Chip architecture analysis.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

With the shift towards many-core computer architectures, dataflow programming has been proposed as one potential solution for producing software that scales to a varying number of processor cores. Programming for parallel architectures is considered difficult as the current popular programming languages are inherently sequential and introducing parallelism is typically up to the programmer. Dataflow, however, is inherently parallel, describing an application as a directed graph, where nodes represent calculations and edges represent a data dependency in form of a queue. These queues are the only allowed communication between the nodes, making the dependencies between the nodes explicit and thereby also the parallelism. Once a node have the su cient inputs available, the node can, independently of any other node, perform calculations, consume inputs, and produce outputs. Data ow models have existed for several decades and have become popular for describing signal processing applications as the graph representation is a very natural representation within this eld. Digital lters are typically described with boxes and arrows also in textbooks. Data ow is also becoming more interesting in other domains, and in principle, any application working on an information stream ts the dataflow paradigm. Such applications are, among others, network protocols, cryptography, and multimedia applications. As an example, the MPEG group standardized a dataflow language called RVC-CAL to be use within reconfigurable video coding. Describing a video coder as a data ow network instead of with conventional programming languages, makes the coder more readable as it describes how the video dataflows through the different coding tools. While dataflow provides an intuitive representation for many applications, it also introduces some new problems that need to be solved in order for data ow to be more widely used. The explicit parallelism of a dataflow program is descriptive and enables an improved utilization of available processing units, however, the independent nodes also implies that some kind of scheduling is required. The need for efficient scheduling becomes even more evident when the number of nodes is larger than the number of processing units and several nodes are running concurrently on one processor core. There exist several data ow models of computation, with different trade-offs between expressiveness and analyzability. These vary from rather restricted but statically schedulable, with minimal scheduling overhead, to dynamic where each ring requires a ring rule to evaluated. The model used in this work, namely RVC-CAL, is a very expressive language, and in the general case it requires dynamic scheduling, however, the strong encapsulation of dataflow nodes enables analysis and the scheduling overhead can be reduced by using quasi-static, or piecewise static, scheduling techniques. The scheduling problem is concerned with nding the few scheduling decisions that must be run-time, while most decisions are pre-calculated. The result is then an, as small as possible, set of static schedules that are dynamically scheduled. To identify these dynamic decisions and to find the concrete schedules, this thesis shows how quasi-static scheduling can be represented as a model checking problem. This involves identifying the relevant information to generate a minimal but complete model to be used for model checking. The model must describe everything that may affect scheduling of the application while omitting everything else in order to avoid state space explosion. This kind of simplification is necessary to make the state space analysis feasible. For the model checker to nd the actual schedules, a set of scheduling strategies are de ned which are able to produce quasi-static schedulers for a wide range of applications. The results of this work show that actor composition with quasi-static scheduling can be used to transform data ow programs to t many different computer architecture with different type and number of cores. This in turn, enables dataflow to provide a more platform independent representation as one application can be fitted to a specific processor architecture without changing the actual program representation. Instead, the program representation is in the context of design space exploration optimized by the development tools to fit the target platform. This work focuses on representing the dataflow scheduling problem as a model checking problem and is implemented as part of a compiler infrastructure. The thesis also presents experimental results as evidence of the usefulness of the approach.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Guided by the social-ecological conceptualization of bullying, this thesis examines the implications of classroom and school contexts—that is, students’ shared microsystems—for peer-to-peer bullying and antibullying practices. Included are four original publications, three of which are empirical studies utilizing data from a large Finnish sample of students in the upper grade levels of elementary school. Both self- and peer reports of bullying and victimization are utilized, and the hierarchical nature of the data collected from students nested within school ecologies is accounted for by multilevel modeling techniques. The first objective of the thesis is to simultaneously examine risk factors for victimization at individual, classroom, and school levels (Study I). The second objective is to uncover the individual- and classroom-level working mechanisms of the KiVa antibullying program which has been shown to be effective in reducing bullying problems in Finnish schools (Study II). Thirdly, an overview of the extant literature on classroom- and school-level contributions to bullying and victimization is provided (Study III). Finally, attention is paid to the assessment of victimization and, more specifically, to how the classroom context influences the concordance between self- and peer reports of victimization (Study IV). Findings demonstrate the multiple ways in which contextual factors, and importantly students’ perceptions thereof, contribute to the bullying dynamic and efforts to counteract it. Whereas certain popular beliefs regarding the implications of classroom and school contexts do not receive support, the role of peer contextual factors and the significance of students’ perceptions of teachers’ attitudes toward bullying are highlighted. Directions for future research and school-based antibullying practices are suggested.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The challenge the community college faces in helping meet the needs of the living open system of society is examined in this study. It is postulated that internalization student outcomes are required by society to reduce entropy and remain self-renewing. Such behavior is characterized as having an intrinsically motivated energy source and displays the seeking and conquering of challenge, the development of reflective knowledge and skill, full use of all capabilities, internal control, growth orientation, high self-esteem, relativistic thinking and competence. The development of a conceptual systems model that suggests how transactions among students, faculty and administration might occur to best meet the needs of internalization outcomes in students, and intrinsic motivation in faculty is a major purpose of this study. It is a speculative model that is based on a synthesis of a wide variety of variables. Empirical evidence, theoretical considerations, and speculative ideas are gathered together from researchers and theoretici.ans who are working on separate answers to questions of intrinsic motivation, internal control and environments that encourage their development. The model considers the effect administrators·have on faculty anq the corresponding effect faculty may have on students. The major concentration is on the administrator--teacher interface.For administrators the model may serve as a guide in planning effective transactions, and establishing system goals. The teacher is offered a means to coordinate actions toward a specific overall objective, and the administrator, teacher and researcher are invited to use the model to experiment, innovate, verify the assumptions on which the model is based, and raise additional hypotheses. Goals and history of the community colleges in Ontario are examined against current problems, previous progress and open system thinking. The nature of the person as a five part system is explored with emphasis on intrinsic motivation. The nature, operation, conceptualization, and value of this internal energy source is reviewed in detail. The current state of society, education and management theory are considered and the value of intrinsically motivating teaching tasks together with "system four" leadership style are featured. Evidence is reviewed that suggests intrinsically motivated faculty are needed, and "system four" leadership style is the kind of interaction-influence system needed to nurture intrinsic motivation in faculty.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Population-based metaheuristics, such as particle swarm optimization (PSO), have been employed to solve many real-world optimization problems. Although it is of- ten sufficient to find a single solution to these problems, there does exist those cases where identifying multiple, diverse solutions can be beneficial or even required. Some of these problems are further complicated by a change in their objective function over time. This type of optimization is referred to as dynamic, multi-modal optimization. Algorithms which exploit multiple optima in a search space are identified as niching algorithms. Although numerous dynamic, niching algorithms have been developed, their performance is often measured solely on their ability to find a single, global optimum. Furthermore, the comparisons often use synthetic benchmarks whose landscape characteristics are generally limited and unknown. This thesis provides a landscape analysis of the dynamic benchmark functions commonly developed for multi-modal optimization. The benchmark analysis results reveal that the mechanisms responsible for dynamism in the current dynamic bench- marks do not significantly affect landscape features, thus suggesting a lack of representation for problems whose landscape features vary over time. This analysis is used in a comparison of current niching algorithms to identify the effects that specific landscape features have on niching performance. Two performance metrics are proposed to measure both the scalability and accuracy of the niching algorithms. The algorithm comparison results demonstrate the algorithms best suited for a variety of dynamic environments. This comparison also examines each of the algorithms in terms of their niching behaviours and analyzing the range and trade-off between scalability and accuracy when tuning the algorithms respective parameters. These results contribute to the understanding of current niching techniques as well as the problem features that ultimately dictate their success.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Les décisions de localisation sont souvent soumises à des aspects dynamiques comme des changements dans la demande des clients. Pour y répondre, la solution consiste à considérer une flexibilité accrue concernant l’emplacement et la capacité des installations. Même lorsque la demande est prévisible, trouver le planning optimal pour le déploiement et l'ajustement dynamique des capacités reste un défi. Dans cette thèse, nous nous concentrons sur des problèmes de localisation avec périodes multiples, et permettant l'ajustement dynamique des capacités, en particulier ceux avec des structures de coûts complexes. Nous étudions ces problèmes sous différents points de vue de recherche opérationnelle, en présentant et en comparant plusieurs modèles de programmation linéaire en nombres entiers (PLNE), l'évaluation de leur utilisation dans la pratique et en développant des algorithmes de résolution efficaces. Cette thèse est divisée en quatre parties. Tout d’abord, nous présentons le contexte industriel à l’origine de nos travaux: une compagnie forestière qui a besoin de localiser des campements pour accueillir les travailleurs forestiers. Nous présentons un modèle PLNE permettant la construction de nouveaux campements, l’extension, le déplacement et la fermeture temporaire partielle des campements existants. Ce modèle utilise des contraintes de capacité particulières, ainsi qu’une structure de coût à économie d’échelle sur plusieurs niveaux. L'utilité du modèle est évaluée par deux études de cas. La deuxième partie introduit le problème dynamique de localisation avec des capacités modulaires généralisées. Le modèle généralise plusieurs problèmes dynamiques de localisation et fournit de meilleures bornes de la relaxation linéaire que leurs formulations spécialisées. Le modèle peut résoudre des problèmes de localisation où les coûts pour les changements de capacité sont définis pour toutes les paires de niveaux de capacité, comme c'est le cas dans le problème industriel mentionnée ci-dessus. Il est appliqué à trois cas particuliers: l'expansion et la réduction des capacités, la fermeture temporaire des installations, et la combinaison des deux. Nous démontrons des relations de dominance entre notre formulation et les modèles existants pour les cas particuliers. Des expériences de calcul sur un grand nombre d’instances générées aléatoirement jusqu’à 100 installations et 1000 clients, montrent que notre modèle peut obtenir des solutions optimales plus rapidement que les formulations spécialisées existantes. Compte tenu de la complexité des modèles précédents pour les grandes instances, la troisième partie de la thèse propose des heuristiques lagrangiennes. Basées sur les méthodes du sous-gradient et des faisceaux, elles trouvent des solutions de bonne qualité même pour les instances de grande taille comportant jusqu’à 250 installations et 1000 clients. Nous améliorons ensuite la qualité de la solution obtenue en résolvent un modèle PLNE restreint qui tire parti des informations recueillies lors de la résolution du dual lagrangien. Les résultats des calculs montrent que les heuristiques donnent rapidement des solutions de bonne qualité, même pour les instances où les solveurs génériques ne trouvent pas de solutions réalisables. Finalement, nous adaptons les heuristiques précédentes pour résoudre le problème industriel. Deux relaxations différentes sont proposées et comparées. Des extensions des concepts précédents sont présentées afin d'assurer une résolution fiable en un temps raisonnable.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Atmospheric Boundary layer (ABL) is the layer just above the earth surface and is influenced by the surface forcing within a short period of an hour or less. In this thesis, characteristics of the boundary layer over ocean, coastal and inland areas of the atmosphere, especially over the monsoon regime are thoroughly studied. The study of the coastal zone is important due to its high vulnerability mainly due to sea breeze circulation and associated changes in the atmospheric boundary layer. The major scientific problems addressed in this thesis are diurnal and seasonal variation of coastal meteorological properties, the characteristic difference in the ABL during active and weak monsoons, features of ABL over marine environment and the variation of the boundary layer structure over an inland station. The thesis describes the various features in the ABL associated with the active and weak monsoons and, the surface boundary layer properties associated with the active and weak epochs. The study provides knowledge on MABL and can be used as the estimated values of boundary layer parameters over the marine atmosphere and to know the values and variabilities of the ABL parameters such as surface wind, surface friction, drag coefficient, wind stress and wind stress curl.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Recently, cumulative residual entropy (CRE) has been found to be a new measure of information that parallels Shannon’s entropy (see Rao et al. [Cumulative residual entropy: A new measure of information, IEEE Trans. Inform. Theory. 50(6) (2004), pp. 1220–1228] and Asadi and Zohrevand [On the dynamic cumulative residual entropy, J. Stat. Plann. Inference 137 (2007), pp. 1931–1941]). Motivated by this finding, in this paper, we introduce a generalized measure of it, namely cumulative residual Renyi’s entropy, and study its properties.We also examine it in relation to some applied problems such as weighted and equilibrium models. Finally, we extend this measure into the bivariate set-up and prove certain characterizing relationships to identify different bivariate lifetime models

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Brazil has been increasing its importance in agricultural markets. The reasons are well known to be the relative abundance of land, the increasing technology used in crops, and the development of the agribusiness sector which allow for a fast response to price stimuli. The elasticity of acreage response to increases in expected return is estimated for Soybeans in a dynamic (long term) error correction model. Regarding yield patterns, a large variation in the yearly rates of growth in yield is observed, climate being probably the main source of this variation which result in ‘good’ and ‘bad’ years. In South America, special attention should be given to the El Niño and La Niña phenomena, both said to have important effects on rainfalls patterns and consequently in yield. The influence on El Niño and La Niña in historical data is examined and some ways of estimating the impact of climate on yield of Soybean and Corn markets are proposed. Possible implications of climate change may apply.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This paper deals with fault detection and isolation problems for nonlinear dynamic systems. Both problems are stated as constraint satisfaction problems (CSP) and solved using consistency techniques. The main contribution is the isolation method based on consistency techniques and uncertainty space refining of interval parameters. The major advantage of this method is that the isolation speed is fast even taking into account uncertainty in parameters, measurements, and model errors. Interval calculations bring independence from the assumption of monotony considered by several approaches for fault isolation which are based on observers. An application to a well known alcoholic fermentation process model is presented

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The performance of a model-based diagnosis system could be affected by several uncertainty sources, such as,model errors,uncertainty in measurements, and disturbances. This uncertainty can be handled by mean of interval models.The aim of this thesis is to propose a methodology for fault detection, isolation and identification based on interval models. The methodology includes some algorithms to obtain in an automatic way the symbolic expression of the residual generators enhancing the structural isolability of the faults, in order to design the fault detection tests. These algorithms are based on the structural model of the system. The stages of fault detection, isolation, and identification are stated as constraint satisfaction problems in continuous domains and solved by means of interval based consistency techniques. The qualitative fault isolation is enhanced by a reasoning in which the signs of the symptoms are derived from analytical redundancy relations or bond graph models of the system. An initial and empirical analysis regarding the differences between interval-based and statistical-based techniques is presented in this thesis. The performance and efficiency of the contributions are illustrated through several application examples, covering different levels of complexity.