743 resultados para Grid Computing
Resumo:
This thesis addresses the problem of computing the minimal and maximal diameter of the Cayley graph of Coxeter groups. We first present and assert relevant parts of polytope theory and related Coxeter theory. After this, a method of contracting the orthogonal projections of a polytope from Rd onto R2 and R3, d ¸ 3 is presented. This method is the Equality Set Projection algorithm that requires a constant number of linearprogramming problems per facet of the projection in the absence of degeneracy. The ESP algorithm allows us to compute also projected geometric diameters of high-dimensional polytopes. A representation set of projected polytopes is presented to illustrate the methods adopted in this thesis.
Resumo:
Laser scanning is becoming an increasingly popular method for measuring 3D objects in industrial design. Laser scanners produce a cloud of 3D points. For CAD software to be able to use such data, however, this point cloud needs to be turned into a vector format. A popular way to do this is to triangulate the assumed surface of the point cloud using alpha shapes. Alpha shapes start from the convex hull of the point cloud and gradually refine it towards the true surface of the object. Often it is nontrivial to decide when to stop this refinement. One criterion for this is to do so when the homology of the object stops changing. This is known as the persistent homology of the object. The goal of this thesis is to develop a way to compute the homology of a given point cloud when processed with alpha shapes, and to infer from it when the persistent homology has been achieved. Practically, the computation of such a characteristic of the target might be applied to power line tower span analysis.
Resumo:
Nowadays the Finnish-Russian electric energy interaction is carried out through the back-to-back DC Vyborg substation and several power plants working synchronously with Finnish power system. Constant amount of energy flows in one direction — from Russia to Finland. But the process of electricity market development in Russian energy system makes the new possibilities of electrical cooperation available. The goal of master's thesis is to analyze the current state and possible evolution trends of North-West Russian system in relation with future possible change in power flow between Russia and Finland. The research is done by modelling the market of North-West Russia and examination of technical grid restrictions. The operational market models of North-West region of Russia for the years 2008 and 2015 were created during the research process. The description of prepared market models together with modelling results and their analysis are shown in the work. The description of power flow study process and results are also presented.
Resumo:
In the Thesis main focus is done on power flow development paths around the Baltic States as well as on market-based requirements for creation of the common Baltic electricity market. Current market regulations between the countries are presented; barriers for creating competitive common Baltic power market and for electricity trading with third countries are clarified; solutions are offered and corresponding road map is developed. Future power development paths around the Baltic States are analysed. For this purpose the 330 kV transmission grid of Estonia, Latvia and Lithuania is modelled in a power flow tool. Power flow calculations are carried out for winter and summer peak and off-peak load periods in 2020 with different combinations of interconnections. While carrying out power balance experiments several power flow patterns in the Baltic States are revealed. Conclusions are made about security of supply, grid congestion and transmission capacity availability for different scenarios.
Resumo:
Quickremovalofbiosolidsinaquaculturefacilities,andspeciallyinrecirculatingaquaculturesystems(RAS),isoneofthemostimportantstepinwastemanagement.Sedimentationdynamicsofbiosolidsinanaquaculturetankwilldeterminetheiraccumulationatthebottomofthetank.
Resumo:
The use of two-dimensional spectral analysis applied to terrain heights in order to determine characteristic terrain spatial scales and its subsequent use for the objective definition of an adequate grid size required to resolve terrain forcing are presented in this paper. In order to illustrate the influence of grid size, atmospheric flow in a complex terrain area of the Spanish east coast is simulated by the Regional Atmospheric Modeling System (RAMS) mesoscale numerical model using different horizontal grid resolutions. In this area, a grid size of 2 km is required to account for 95% of terrain variance. Comparison among results of the different simulations shows that, although the main wind behavior does not change dramatically, some small-scale features appear when using a resolution of 2 km or finer. Horizontal flow pattern differences are significant both in the nighttime, when terrain forcing is more relevant, and in the daytime, when thermal forcing is dominant. Vertical structures also are investigated, and results show that vertical advection is influenced highly by the horizontal grid size during the daytime period. The turbulent kinetic energy and potential temperature vertical cross sections show substantial differences in the structure of the planetary boundary layer for each model configuration
Resumo:
ABSTRACT The successful in the implementation of wind turbines depends on several factors, including: the wind resource at the installation site, the equipment used, project acquisition and operational costs. In this paper, the production of electricity from two small wind turbines was compared through simulation using the computer software HOMER - a national model of 6kW and an imported one of 5kW. The wind resources in three different cities were considered: Campinas (SP/BR), Cubatão (São Paulo/BR) and Roscoe (Texas/ USA). A wind power system connected to the grid and a wind isolated system - batteries were evaluated. The results showed that the energy cost ($/kWh) is strongly dependent on the windmill characteristics and local wind resource. Regarding the isolated wind system – batteries, the full supply guarantee to the simulated electrical load is only achieved with a battery bank with many units and high number of wind turbines, due to the intermittency of wind power.
Resumo:
Memristive computing refers to the utilization of the memristor, the fourth fundamental passive circuit element, in computational tasks. The existence of the memristor was theoretically predicted in 1971 by Leon O. Chua, but experimentally validated only in 2008 by HP Labs. A memristor is essentially a nonvolatile nanoscale programmable resistor — indeed, memory resistor — whose resistance, or memristance to be precise, is changed by applying a voltage across, or current through, the device. Memristive computing is a new area of research, and many of its fundamental questions still remain open. For example, it is yet unclear which applications would benefit the most from the inherent nonlinear dynamics of memristors. In any case, these dynamics should be exploited to allow memristors to perform computation in a natural way instead of attempting to emulate existing technologies such as CMOS logic. Examples of such methods of computation presented in this thesis are memristive stateful logic operations, memristive multiplication based on the translinear principle, and the exploitation of nonlinear dynamics to construct chaotic memristive circuits. This thesis considers memristive computing at various levels of abstraction. The first part of the thesis analyses the physical properties and the current-voltage behaviour of a single device. The middle part presents memristor programming methods, and describes microcircuits for logic and analog operations. The final chapters discuss memristive computing in largescale applications. In particular, cellular neural networks, and associative memory architectures are proposed as applications that significantly benefit from memristive implementation. The work presents several new results on memristor modeling and programming, memristive logic, analog arithmetic operations on memristors, and applications of memristors. The main conclusion of this thesis is that memristive computing will be advantageous in large-scale, highly parallel mixed-mode processing architectures. This can be justified by the following two arguments. First, since processing can be performed directly within memristive memory architectures, the required circuitry, processing time, and possibly also power consumption can be reduced compared to a conventional CMOS implementation. Second, intrachip communication can be naturally implemented by a memristive crossbar structure.
Resumo:
Valmistustekniikoiden kehittyessä IC-piireille saadaan mahtumaan yhä enemmän transistoreja. Monimutkaisemmat piirit mahdollistavat suurempien laskutoimitusmäärien suorittamisen aikayksikössä. Piirien aktiivisuuden lisääntyessä myös niiden energiankulutus lisääntyy, ja tämä puolestaan lisää piirin lämmöntuotantoa. Liiallinen lämpö rajoittaa piirien toimintaa. Tämän takia tarvitaan tekniikoita, joilla piirien energiankulutusta saadaan pienennettyä. Uudeksi tutkimuskohteeksi ovat tulleet pienet laitteet, jotka seuraavat esimerkiksi ihmiskehon toimintaa, rakennuksia tai siltoja. Tällaisten laitteiden on oltava energiankulutukseltaan pieniä, jotta ne voivat toimia pitkiä aikoja ilman akkujen lataamista. Near-Threshold Computing on tekniikka, jolla pyritään pienentämään integroitujen piirien energiankulutusta. Periaatteena on käyttää piireillä pienempää käyttöjännitettä kuin piirivalmistaja on niille alunperin suunnitellut. Tämä hidastaa ja haittaa piirin toimintaa. Jos kuitenkin laitteen toiminnassa pystyään hyväksymään huonompi laskentateho ja pienentynyt toimintavarmuus, voidaan saavuttaa säästöä energiankulutuksessa. Tässä diplomityössä tarkastellaan Near-Threshold Computing -tekniikkaa eri näkökulmista: aluksi perustuen kirjallisuudesta löytyviin aikaisempiin tutkimuksiin, ja myöhemmin tutkimalla Near-Threshold Computing -tekniikan soveltamista kahden tapaustutkimuksen kautta. Tapaustutkimuksissa tarkastellaan FO4-invertteriä sekä 6T SRAM -solua piirisimulaatioiden avulla. Näiden komponenttien käyttäytymisen Near-Threshold Computing –jännitteillä voidaan tulkita antavan kattavan kuvan suuresta osasta tavanomaisen IC-piirin pinta-alaa ja energiankulusta. Tapaustutkimuksissa käytetään 130 nm teknologiaa, ja niissä mallinnetaan todellisia piirivalmistusprosessin tuotteita ajamalla useita Monte Carlo -simulaatioita. Tämä valmistuskustannuksiltaan huokea teknologia yhdistettynä Near-Threshold Computing -tekniikkaan mahdollistaa matalan energiankulutuksen piirien valmistaminen järkevään hintaan. Tämän diplomityön tulokset näyttävät, että Near-Threshold Computing pienentää piirien energiankulutusta merkittävästi. Toisaalta, piirien nopeus heikkenee, ja yleisesti käytetty 6T SRAM -muistisolu muuttuu epäluotettavaksi. Pidemmät polut logiikkapiireissä sekä transistorien kasvattaminen muistisoluissa osoitetaan tehokkaiksi vastatoimiksi Near- Threshold Computing -tekniikan huonoja puolia vastaan. Tulokset antavat perusteita matalan energiankulutuksen IC-piirien suunnittelussa sille, kannattaako käyttää normaalia käyttöjännitettä, vai laskea sitä, jolloin piirin hidastuminen ja epävarmempi käyttäytyminen pitää ottaa huomioon.
Resumo:
In accordance with the Moore's law, the increasing number of on-chip integrated transistors has enabled modern computing platforms with not only higher processing power but also more affordable prices. As a result, these platforms, including portable devices, work stations and data centres, are becoming an inevitable part of the human society. However, with the demand for portability and raising cost of power, energy efficiency has emerged to be a major concern for modern computing platforms. As the complexity of on-chip systems increases, Network-on-Chip (NoC) has been proved as an efficient communication architecture which can further improve system performances and scalability while reducing the design cost. Therefore, in this thesis, we study and propose energy optimization approaches based on NoC architecture, with special focuses on the following aspects. As the architectural trend of future computing platforms, 3D systems have many bene ts including higher integration density, smaller footprint, heterogeneous integration, etc. Moreover, 3D technology can signi cantly improve the network communication and effectively avoid long wirings, and therefore, provide higher system performance and energy efficiency. With the dynamic nature of on-chip communication in large scale NoC based systems, run-time system optimization is of crucial importance in order to achieve higher system reliability and essentially energy efficiency. In this thesis, we propose an agent based system design approach where agents are on-chip components which monitor and control system parameters such as supply voltage, operating frequency, etc. With this approach, we have analysed the implementation alternatives for dynamic voltage and frequency scaling and power gating techniques at different granularity, which reduce both dynamic and leakage energy consumption. Topologies, being one of the key factors for NoCs, are also explored for energy saving purpose. A Honeycomb NoC architecture is proposed in this thesis with turn-model based deadlock-free routing algorithms. Our analysis and simulation based evaluation show that Honeycomb NoCs outperform their Mesh based counterparts in terms of network cost, system performance as well as energy efficiency.
On the development of an unstructured grid solver for inert and reactive high speed flow simulations
Resumo:
An unstructured grid Euler solver for reactive compressible flow applications is presented. The method is implemented in a cell centered, finite volume context for unstructured triangular grids. Three different schemes for spatial discretization are implemented and analyzed. Time march is implemented in a time-split fashion with independent integrators for the flow and chemistry equations. The capability implemented is tested for inert flows in a hypersonic inlet and for inert and reactive supersonic flows over a 2-D wedge. The results of the different schemes are compared with each other and with independent calculations using a structured grid code. The strengths and the possible weaknesses of the proposed methods are discussed.
Resumo:
This thesis discusses the opportunities and challenges of the cloud computing technology in healthcare information systems by reviewing the existing literature on cloud computing and healthcare information system and the impact of cloud computing technology to healthcare industry. The review shows that if problems related to security of data are solved then cloud computing will positively transform the healthcare institutions by giving advantage to the healthcare IT infrastructure as well as improving and giving benefit to healthcare services. Therefore, this thesis will explore the opportunities and challenges that are associated with cloud computing in the context of Finland in order to help the healthcare organizations and stakeholders to determine its direction when it decides to adopt cloud technology on their information systems.