917 resultados para power engineering computing
Resumo:
Aristotle made the observation that man is a political animal. Engineers often like to think they are above the fray when it comes to organizational politics, but most organizational theorists believe politics is a fundamental dynamic in any group. This paper examines the various ways that people use power within organizations to negotiate the political interactions in the work place.
Resumo:
Photovoltaic power has become one of the most popular research area in new energy field. In this report, the case of household solar power system is presented. Based on the Matlab environment, the simulation is built by using Simulink and SimPowerSystem. There are four parts in a household solar system, solar cell, MPPT system, battery and power consumer. Solar cell and MPPT system are been studied and analyzed individually. The system with MPPT generates 30% more energy than the system without MPPT. After simulating the household system, it is can be seen that the power which generated by the system is 40.392 kWh per sunny day. By combining the power generated by the system and the price of the electric power, 8.42 years are need for the system to achieve a balance of income and expenditure when weather condition is considered.
Resumo:
In power electronic basedmicrogrids, the computational requirements needed to implement an optimized online control strategy can be prohibitive. The work presented in this dissertation proposes a generalized method of derivation of geometric manifolds in a dc microgrid that is based on the a-priori computation of the optimal reactions and trajectories for classes of events in a dc microgrid. The proposed states are the stored energies in all the energy storage elements of the dc microgrid and power flowing into them. It is anticipated that calculating a large enough set of dissimilar transient scenarios will also span many scenarios not specifically used to develop the surface. These geometric manifolds will then be used as reference surfaces in any type of controller, such as a sliding mode hysteretic controller. The presence of switched power converters in microgrids involve different control actions for different system events. The control of the switch states of the converters is essential for steady state and transient operations. A digital memory look-up based controller that uses a hysteretic sliding mode control strategy is an effective technique to generate the proper switch states for the converters. An example dcmicrogrid with three dc-dc boost converters and resistive loads is considered for this work. The geometric manifolds are successfully generated for transient events, such as step changes in the loads and the sources. The surfaces corresponding to a specific case of step change in the loads are then used as reference surfaces in an EEPROM for experimentally validating the control strategy. The required switch states corresponding to this specific transient scenario are programmed in the EEPROM as a memory table. This controls the switching of the dc-dc boost converters and drives the system states to the reference manifold. In this work, it is shown that this strategy effectively controls the system for a transient condition such as step changes in the loads for the example case.
Resumo:
We present a technique for online compression of ECG signals using the Golomb-Rice encoding algorithm. This is facilitated by a novel time encoding asynchronous analog-to-digital converter targeted for low-power, implantable, long-term bio-medical sensing applications. In contrast to capturing the actual signal (voltage) values the asynchronous time encoder captures and encodes the time information at which predefined changes occur in the signal thereby minimizing the sensor's energy use and the number of bits we store to represent the information by not capturing unnecessary samples. The time encoder transforms the ECG signal data to pure time information that has a geometric distribution such that the Golomb-Rice encoding algorithm can be used to further compress the data. An overall online compression rate of about 6 times is achievable without the usual computations associated with most compression methods.
Resumo:
We present a power-scalable approach for yellow laser-light generation based on standard Ytterbium (Yb) doped fibers. To force the cavity to lase at 1154 nm, far above the gain-maximum, measures must be taken to fulfill lasing condition and to suppress competing amplified spontaneous emission (ASE) in the high-gain region. To prove the principle we built a fiber-laser cavity and a fiber-amplifier both at 1154 nm. In between cavity and amplifier we suppressed the ASE by 70 dB using a fiber Bragg grating (FBG) based filter. Finally we demonstrated efficient single pass frequency doubling to 577 nm with a periodically poled lithium niobate crystal (PPLN). With our linearly polarized 1154 nm master oscillator power fiber amplifier (MOFA) system we achieved slope efficiencies of more than 15 % inside the cavity and 24 % with the fiber-amplifier. The frequency doubling followed the predicted optimal efficiency achievable with a PPLN crystal. So far we generated 1.5 W at 1154nm and 90 mW at 577 nm. Our MOFA approach for generation of 1154 nm laser radiation is power-scalable by using multi-stage amplifiers and large mode-area fibers and is therefore very promising for building a high power yellow laser-light source of several tens of Watt.
Resumo:
Nowadays computing platforms consist of a very large number of components that require to be supplied with diferent voltage levels and power requirements. Even a very small platform, like a handheld computer, may contain more than twenty diferent loads and voltage regulators. The power delivery designers of these systems are required to provide, in a very short time, the right power architecture that optimizes the performance, meets electrical specifications plus cost and size targets. The appropriate selection of the architecture and converters directly defines the performance of a given solution. Therefore, the designer needs to be able to evaluate a significant number of options in order to know with good certainty whether the selected solutions meet the size, energy eficiency and cost targets. The design dificulties of selecting the right solution arise due to the wide range of power conversion products provided by diferent manufacturers. These products range from discrete components (to build converters) to complete power conversion modules that employ diferent manufacturing technologies. Consequently, in most cases it is not possible to analyze all the alternatives (combinations of power architectures and converters) that can be built. The designer has to select a limited number of converters in order to simplify the analysis. In this thesis, in order to overcome the mentioned dificulties, a new design methodology for power supply systems is proposed. This methodology integrates evolutionary computation techniques in order to make possible analyzing a large number of possibilities. This exhaustive analysis helps the designer to quickly define a set of feasible solutions and select the best trade-off in performance according to each application. The proposed approach consists of two key steps, one for the automatic generation of architectures and other for the optimized selection of components. In this thesis are detailed the implementation of these two steps. The usefulness of the methodology is corroborated by contrasting the results using real problems and experiments designed to test the limits of the algorithms.
Resumo:
All meta-analyses should include a heterogeneity analysis. Even so, it is not easy to decide whether a set of studies are homogeneous or heterogeneous because of the low statistical power of the statistics used (usually the Q test). Objective: Determine a set of rules enabling SE researchers to find out, based on the characteristics of the experiments to be aggregated, whether or not it is feasible to accurately detect heterogeneity. Method: Evaluate the statistical power of heterogeneity detection methods using a Monte Carlo simulation process. Results: The Q test is not powerful when the meta-analysis contains up to a total of about 200 experimental subjects and the effect size difference is less than 1. Conclusions: The Q test cannot be used as a decision-making criterion for meta-analysis in small sample settings like SE. Random effects models should be used instead of fixed effects models. Caution should be exercised when applying Q test-mediated decomposition into subgroups.
Resumo:
Nowadays, computer simulators are becoming basic tools for education and training in many engineering fields. In the nuclear industry, the role of simulation for training of operators of nuclear power plants is also recognized of the utmost relevance. As an example, the International Atomic Energy Agency sponsors the development of nuclear reactor simulators for education, and arranges the supply of such simulation programs. Aware of this, in 2008 Gas Natural Fenosa, a Spanish gas and electric utility that owns and operate nuclear power plants and promotes university education in the nuclear technology field, provided the Department of Nuclear Engineering of Universidad Politécnica de Madrid with the Interactive Graphic Simulator (IGS) of “José Cabrera” (Zorita) nuclear power plant, an industrial facility whose commercial operation ceased definitively in April 2006. It is a state-of-the-art full-scope real-time simulator that was used for training and qualification of the operators of the plant control room, as well as to understand and analyses the plant dynamics, and to develop, qualify and validate its emergency operating procedures.
Resumo:
In this paper, the dynamic response of a hydro power plant for providing secondary regulation reserve is studied in detail. Special emphasis is given to the elastic water column effects both in the penstock and the tailrace tunnel. For this purpose, a nonlinear model based on the analogy between mass and momentum conservation equations of a water conduit and those of wave propagation in transmission lines is used. The influence of the plant configuration and design parameters on the fulfilment of the Spanish Electrical System Operator requirements is analysed
Resumo:
The European Credit Transfer and Accumulation System (ECTS) is the credit system for higher education used in the European Higher Education Area (EHEA), which involves all the countries engaged in the Bologna Process. This paper describes a study which is part of the project of the Bologna Experts Team-Spain and was carried out with the following aims: 1) designing some procedures for the assessment of transferable competences; and 2) testing some basic psychometric features that an assessment device with some consequences for the subjects being evaluated needs to prove. We will focus on the degrees of Computing. The sample of students (20) includes first year students from the Technical University of Madrid. In this paper, we will report some results of data analyses carried out to this moment on reliability and validity of the task designed to measure problem solving.
Resumo:
In this paper, calculus of variations and combined blade element and momentum theory (BEMT) are used to demonstrate that, in hover, when neither root nor tip losses are considered; the rotor, which minimizes the total power (MPR), generates an induced velocity that varies linearly along the blade span. The angle of attack of every blade element is constant and equal to its optimum value. The traditional ideal twist (ITR) and optimum (OR) rotors are revisited in the context of this variational framework. Two more optimum rotors are obtained considering root and tip losses, the ORL, and the MPRL. A comparison between these five rotors is presented and discussed. The MPR and MPRL present a remarkable saving of power for low values of both thrust coefficient and maximum aerodynamic efficiency. The result obtained can be exploited to improve the aerodynamic behaviour of rotary wing micro air vehicles (MAV). A comparison with experimental results obtained from the literature is presented.
Resumo:
Ubiquitous sensor network deployments, such as the ones found in Smart cities and Ambient intelligence applications, require constantly increasing high computational demands in order to process data and offer services to users. The nature of these applications imply the usage of data centers. Research has paid much attention to the energy consumption of the sensor nodes in WSNs infrastructures. However, supercomputing facilities are the ones presenting a higher economic and environmental impact due to their very high power consumption. The latter problem, however, has been disregarded in the field of smart environment services. This paper proposes an energy-minimization workload assignment technique, based on heterogeneity and application-awareness, that redistributes low-demand computational tasks from high-performance facilities to idle nodes with low and medium resources in the WSN infrastructure. These non-optimal allocation policies reduce the energy consumed by the whole infrastructure and the total execution time.
Resumo:
In the present competitive environment, companies are wondering how to reduce their IT costs while increasing their efficiency and agility to react when changes in the business processes are required. Cloud Computing is the latest paradigm to optimize the use of IT resources considering ?everything as a service? and receiving these services from the Cloud (Internet) instead of owning and managing hardware and software assets. The benefits from the model are clear. However, there are also concerns and issues to be solved before Cloud Computing spreads across the different industries. This model will allow a pay-per-use model for the IT services and many benefits like cost savings, agility to react when business demands changes and simplicity because there will not be any infrastructure to operate and administrate. It will be comparable to the well known utilities like electricity, water or gas companies. However, this paper underlines several risk factors of the model. Leading technology companies should research on solutions to minimize the risks described in this article. Keywords - Cloud Computing, Utility Computing, Elastic Computing, Enterprise Agility
Resumo:
The European strategies on energy have been searching for years to reduce the dependency of Europe from fossil fuels. Underlying this effort, there exist geopolitical, economic, environmental reasons and the reality that oil reservoirs will dry out some day. Renewable energies have become a milestone of this strategy because their huge potential has emerged after years of uncertainty. One of the better developed renewable sources, which is nearer to commercial maturity is solar-thermal energy. In this paper, the current state of this technology will be described as well as the developments that may be expected in the short and mid terms, including the thermoelectric solar megaproject DESERTEC, a German proposal to ensure energy resources to the mayor areas of the EU-MENA countries. The reader will acquire a picture of the current state of the market, of the technical challenges already achieved and of the remaining ones.