52 resultados para Construction. Indicators System. Performance. Ergonomics. Validation
Resumo:
The increasing power demand and emerging applications drive the design of electrical power converters into modularization. Despite the wide use of modularized power stage structures, the control schemes that are used are often traditional, in other words, centralized. The flexibility and re-usability of these controllers are typically poor. With a dedicated distributed control scheme, the flexibility and re-usability of the system parts, building blocks, can be increased. Only a few distributed control schemes have been introduced for this purpose, but their breakthrough has not yet taken place. A demand for the further development offlexible control schemes for building-block-based applications clearly exists. The control topology, communication, synchronization, and functionality allocationaspects of building-block-based converters are studied in this doctoral thesis. A distributed control scheme that can be easily adapted to building-block-based power converter designs is developed. The example applications are a parallel and series connection of building blocks. The building block that is used in the implementations of both the applications is a commercial off-the-shelf two-level three-phase frequency converter with a custom-designed controller card. The major challenge with the parallel connection of power stages is the synchronization of the building blocks. The effect of synchronization accuracy on the system performance is studied. The functionality allocation and control scheme design are challenging in the seriesconnected multilevel converters, mainly because of the large number of modules. Various multilevel modulation schemes are analyzed with respect to the implementation, and this information is used to develop a flexible control scheme for modular multilevel inverters.
Resumo:
In accordance with the Moore's law, the increasing number of on-chip integrated transistors has enabled modern computing platforms with not only higher processing power but also more affordable prices. As a result, these platforms, including portable devices, work stations and data centres, are becoming an inevitable part of the human society. However, with the demand for portability and raising cost of power, energy efficiency has emerged to be a major concern for modern computing platforms. As the complexity of on-chip systems increases, Network-on-Chip (NoC) has been proved as an efficient communication architecture which can further improve system performances and scalability while reducing the design cost. Therefore, in this thesis, we study and propose energy optimization approaches based on NoC architecture, with special focuses on the following aspects. As the architectural trend of future computing platforms, 3D systems have many bene ts including higher integration density, smaller footprint, heterogeneous integration, etc. Moreover, 3D technology can signi cantly improve the network communication and effectively avoid long wirings, and therefore, provide higher system performance and energy efficiency. With the dynamic nature of on-chip communication in large scale NoC based systems, run-time system optimization is of crucial importance in order to achieve higher system reliability and essentially energy efficiency. In this thesis, we propose an agent based system design approach where agents are on-chip components which monitor and control system parameters such as supply voltage, operating frequency, etc. With this approach, we have analysed the implementation alternatives for dynamic voltage and frequency scaling and power gating techniques at different granularity, which reduce both dynamic and leakage energy consumption. Topologies, being one of the key factors for NoCs, are also explored for energy saving purpose. A Honeycomb NoC architecture is proposed in this thesis with turn-model based deadlock-free routing algorithms. Our analysis and simulation based evaluation show that Honeycomb NoCs outperform their Mesh based counterparts in terms of network cost, system performance as well as energy efficiency.
Resumo:
Today's networked systems are becoming increasingly complex and diverse. The current simulation and runtime verification techniques do not provide support for developing such systems efficiently; moreover, the reliability of the simulated/verified systems is not thoroughly ensured. To address these challenges, the use of formal techniques to reason about network system development is growing, while at the same time, the mathematical background necessary for using formal techniques is a barrier for network designers to efficiently employ them. Thus, these techniques are not vastly used for developing networked systems. The objective of this thesis is to propose formal approaches for the development of reliable networked systems, by taking efficiency into account. With respect to reliability, we propose the architectural development of correct-by-construction networked system models. With respect to efficiency, we propose reusable network architectures as well as network development. At the core of our development methodology, we employ the abstraction and refinement techniques for the development and analysis of networked systems. We evaluate our proposal by employing the proposed architectures to a pervasive class of dynamic networks, i.e., wireless sensor network architectures as well as to a pervasive class of static networks, i.e., network-on-chip architectures. The ultimate goal of our research is to put forward the idea of building libraries of pre-proved rules for the efficient modelling, development, and analysis of networked systems. We take into account both qualitative and quantitative analysis of networks via varied formal tool support, using a theorem prover the Rodin platform and a statistical model checker the SMC-Uppaal.
Resumo:
Multiprocessor system-on-chip (MPSoC) designs utilize the available technology and communication architectures to meet the requirements of the upcoming applications. In MPSoC, the communication platform is both the key enabler, as well as the key differentiator for realizing efficient MPSoCs. It provides product differentiation to meet a diverse, multi-dimensional set of design constraints, including performance, power, energy, reconfigurability, scalability, cost, reliability and time-to-market. The communication resources of a single interconnection platform cannot be fully utilized by all kind of applications, such as the availability of higher communication bandwidth for computation but not data intensive applications is often unfeasible in the practical implementation. This thesis aims to perform the architecture-level design space exploration towards efficient and scalable resource utilization for MPSoC communication architecture. In order to meet the performance requirements within the design constraints, careful selection of MPSoC communication platform, resource aware partitioning and mapping of the application play important role. To enhance the utilization of communication resources, variety of techniques such as resource sharing, multicast to avoid re-transmission of identical data, and adaptive routing can be used. For implementation, these techniques should be customized according to the platform architecture. To address the resource utilization of MPSoC communication platforms, variety of architectures with different design parameters and performance levels, namely Segmented bus (SegBus), Network-on-Chip (NoC) and Three-Dimensional NoC (3D-NoC), are selected. Average packet latency and power consumption are the evaluation parameters for the proposed techniques. In conventional computing architectures, fault on a component makes the connected fault-free components inoperative. Resource sharing approach can utilize the fault-free components to retain the system performance by reducing the impact of faults. Design space exploration also guides to narrow down the selection of MPSoC architecture, which can meet the performance requirements with design constraints.
Resumo:
This thesis researches automatic traffic sign inventory and condition analysis using machine vision and pattern recognition methods. Automatic traffic sign inventory and condition analysis can be used to more efficient road maintenance, improving the maintenance processes, and to enable intelligent driving systems. Automatic traffic sign detection and classification has been researched before from the viewpoint of self-driving vehicles, driver assistance systems, and the use of signs in mapping services. Machine vision based inventory of traffic signs consists of detection, classification, localization, and condition analysis of traffic signs. The produced machine vision system performance is estimated with three datasets, from which two of have been been collected for this thesis. Based on the experiments almost all traffic signs can be detected, classified, and located and their condition analysed. In future, the inventory system performance has to be verified in challenging conditions and the system has to be pilot tested.
Resumo:
In recent years, technological advancements in microelectronics and sensor technologies have revolutionized the field of electrical engineering. New manufacturing techniques have enabled a higher level of integration that has combined sensors and electronics into compact and inexpensive systems. Previously, the challenge in measurements was to understand the operation of the electronics and sensors, but this has now changed. Nowadays, the challenge in measurement instrumentation lies in mastering the whole system, not just the electronics. To address this issue, this doctoral dissertation studies whether it would be beneficial to consider a measurement system as a whole from the physical phenomena to the digital recording device, where each piece of the measurement system affects the system performance, rather than as a system consisting of small independent parts such as a sensor or an amplifier that could be designed separately. The objective of this doctoral dissertation is to describe in depth the development of the measurement system taking into account the challenges caused by the electrical and mechanical requirements and the measurement environment. The work is done as an empirical case study in two example applications that are both intended for scientific studies. The cases are a light sensitive biological sensor used in imaging and a gas electron multiplier detector for particle physics. The study showed that in these two cases there were a number of different parts of the measurement system that interacted with each other. Without considering these interactions, the reliability of the measurement may be compromised, which may lead to wrong conclusions about the measurement. For this reason it is beneficial to conceptualize the measurement system as a whole from the physical phenomena to the digital recording device where each piece of the measurement system affects the system performance. The results work as examples of how a measurement system can be successfully constructed to support a study of sensors and electronics.
Resumo:
This thesis was produced for the Technology Marketing unit at the Nokia Research Center. Technology marketing was a new function at Nokia Research Center, and needed an established framework with the capacity to take into account multiple aspects for measuring the team performance. Technology marketing functions had existed in other parts of Nokia, yet no single method had been agreed upon for measuring their performance. The purpose of this study was to develop a performance measurement system for Nokia Research Center Technology Marketing. The target was that Nokia Research Center Technology Marketing had a framework for separate metrics; including benchmarking for starting level and target values in the future planning (numeric values were kept confidential within the company). As a result of this research, the Balanced Scorecard model of Kaplan and Norton, was chosen for the performance measurement system for Nokia Research Center Technology Marketing. This research selected the indicators, which were utilized in the chosen performance measurement system. Furthermore, performance measurement system was defined to guide the Head of Marketing in managing Nokia Research Center Technology Marketing team. During the research process the team mission, vision, strategy and critical success factors were outlined.
Resumo:
Tutkimuksen tavoite oli selvittää suorituskyvyn mittaamista, mittareita ja niiden suunnittelua tukku- ja jakeluliiketoiminnassa. Kriittisten menestystekijöiden mittarit auttavat yritystä kohti yhteistä päämäärää. Kriittisten menestystekijöiden mittarit ovat usein yhdistetty strategiseen suunnitteluun ja implementointiin ja niillä on yhtäläisyyksiä monien strategisten työkalujen kun Balanced scorecardin kanssa. Tutkimus ongelma voidaan esittää kysymyksen muodossa. •Mitkä ovat Oriola KD:n pitkänaikavälin tavoitteita tukevat kriittisten menestystekijöiden mittarit (KPIs) toimittajan ja tuotevalikoiman mittaamisessa? Tutkimus on jaettu kirjalliseen ja empiiriseen osaan. Kirjallisuus katsaus käsittelee aikaisempaa tutkimusta strategian, toimitusketjun hallinnan, toimittajan arvioinnin ja erilaisten suorituskyvyn mittaamisjärjestelmien osalta. Empiirinen osuus etenee nykytila-analyysista ehdotettuihin kriittisten menestystekijöiden mittareihin, jotka ovat kehitetty kirjallisuudesta löydetyn mallin avulla. Tutkimuksen lopputuloksena ovat case yrityksen tarpeisiin kehitetyt kriittisten menestystekijöiden mittarit toimittajan ja tuotevalikoiman arvioinnissa.
Resumo:
The purpose of this study is to examine macroeconomic indicators‟ and technical analysis‟ ability to signal market crashes. Indicators examined were Yield Spread, The Purchasing Managers Index and the Consumer Confidence Index. Technical Analysis indicators were moving average, Moving Average Convergence-Divergence and Relative Strength Index. We studied if commonly used macroeconomic indicators can be used as a warning system for a stock market crashes as well. The hypothesis is that the signals of recession can be used as signals of stock market crash and that way a basis for a hedging strategy. The data is collected from the U.S. markets from the years 1983-2010. Empirical studies show that macroeconomic indicators have been able to explain the future GDP development in the U.S. in research period and they were statistically significant. A hedging strategy that combined the signals of yield spread and Consumer Confidence Index gave most useful results as a basis of a hedging strategy in selected time period. It was able to outperform buy-and-hold strategy as well as all of the technical indicator based hedging strategies.
Resumo:
Metadata in increasing levels of sophistication has been the most powerful concept used in management of unstructured information ever since the first librarian used the Dewey decimal system for library classifications. It remains to be seen, however, what the best approach is to implementing metadata to manage huge volumes of unstructured information in a large organization. Also, once implemented, how is it possible to track whether it is adding value to the company, and whether the implementation has been successful? Existing literature on metadata seems to either focus too much on technical and quality aspects or describe issues with respect to adoption for general information management initiatives. This research therefore, strives to contribute to these gaps: to give a consolidated framework for striving to understand the value added by implementing metadata. The basic methodology used is that of case study, which incorporates aspects of design science, surveys, and interviews in order to provide a holistic approach to quantitative and qualitative analysis of the case. The research identifies the various approaches to implementing metadata, particularly studying the one followed by the unit of analysis of case study, a large company in the Oil and Gas Sector. Of the three approaches identified, the selected company already follows an approach that appears to be superior. The researcher further explores its shortcomings, and proposes a slightly modified approach that can handle them. The research categorically and thoroughly (in context) identifies the top effectiveness criteria, and corresponding key performance indicators(KPIs) that can be measured to understand the level of advancement of the metadata management initiative in the company. In an effort to contrast and have a basis of comparison for the findings, the research also includes views from information managers dealing with core structured data stored in ERPs and other databases. In addition, the results include the basic criteria that can be used to evaluate metrics, in order to classify a metric as a KPI.
Resumo:
The performance measurement produces information about the operation of the business process. On the basis of this information performance of the company can be followed and improved. Balanced performance measurement system can monitor performance of several perspectives and business processes can be led according to company strategy. Major part of the costs of a company is originated from purchased goods or services are an output of the buying process emphasising the importance of a reliable performance measurement of purchasing process. In the study, theory of balanced performance measurement is orientated and framework of purchasing process performance measurement system is designed. The designed balanced performance measurement system of purchasing process is tested in case company paying attention to the available data and to other environmental enablers. The balanced purchasing performance measurement system is tested and improved during the test period and attention is paid to the definition and scaling of objectives. Found development initiatives are carried out especially in the scaling of indicators. Finally results of the study are evaluated, conclusions and additional research areas proposed.
Resumo:
Tämä diplomityö on tehty osana HumanICT-projektia, jonka tavoitteena on kehittää uusi, virtuaalitekniikoita hyödyntävä, työkoneiden käyttäjäliityntöjen suunnittelumenetelmä. Työn tarkoituksena oli kehittää VTT:n Tuotteet ja tuotanto tutkimusyksikköön kuluvan Ihminen-kone-turvallisuus ryhmän nykyistä virtuaalitodellisuuslaboratoriota siten, että sitä voidaan käyttää työkoneiden suunnittelussa sekä monipuolisissa ergonomiatarkasteluissa. Itse ympäristön kehittäminen pitää sisällään uuden ohjainjärjestelmän suunnittelun sekä sen implementoinnin nykyisin käytössä olevaan virtuaaliympäristöön. Perinteisesti ohjaamosimulaattorit ovat olleet sovelluskohteisiin räätälöityjä, joten ne ovat kalliita ja niiden konfiguroinnin muuttaminen on vaikeaa, joskus jopa mahdotonta. Tämän työntarkoituksena oli kehittää PC-tietokoneeseen ja yleiseen käyttöjärjestelmään perustuva ohjainjärjestelmä, joka on nopeasti kytkettävissä erilaisiin virtuaaliympäristön sovelluksiin, kuten ohjaamomalleihin. Työssä tarkasteltiin myös tapoja mallintaa fysikaalisia ilmiöitä reaaliaikasovelluksissa, eli on-line simuloinnissa. Tämän tarkastelun perusteella etsittiin ja valittiin jatkokäsittelyyn ohjelmistoja, joiden reaaliaikaisen dynamiikan simulointialgoritmitolivat kaikkein kehittyneimpiä ja monipuolisia.
Resumo:
Container Handling Equipment Monitoring System (CHEMS) is a system developed by Savcor One Oy. CHEMS measures important information for container ports performance and produces performance indicators. The aim of this thesis was to clarify performance measurement contents to Savcor and to develop, as an example, performance measures to Steveco Oy's container operations. The theoretical part of the thesis clarifies performance measurement and which of its components are important to container port. Performance measurement and measures are presented from the operational level's point of view, in which CHEMS is planned to aim. The theory of development process of performance measures is introduced at the end of the theoretical part. To make sure that performance measures are efficiently used, Steveco Oy's performance measures are developed in cooperation with the users. The measurement in operational level is continuous and the results must be reacted asquickly as possible. CHEMS is very suitable to continuous measurement and to produce real time-measures of container operations which are hard to get any otherway.
Resumo:
Puhdastilojen suunnittelussa pyritään saamaan hallittu ja valvottu ilmanpuhtaus luokiteltuun tilaan.Luokittelu tapahtuu puhdastilastandardeilla, lisäksi lääkevalmisteita valmistettavassa tilassa GMP -säädösten mukaisin luokituksin. Puhdastilastandardi ISO 14644 käsittää seitsemän osaa, jossa on käsitelty puhdastilaa koskevia määräyksiä suunnittelusta käyttöön ja testaukseen. GMP-säädökset sisältävät yhdeksän kappaletta, joista kappale 3: 'Tilat ja laitteet' on keskeinen osa lääkeainevalmistuksen puhdastilasuunnittelua. Puhtaan ilman aikaansaamiseksi puhdastilaan merkittävimmät roolit ovat ilmanvaihdolla, puhdastilarakenteilla ja rakennusautomaatiolla. Ilma voidaan tuoda tilaan kolmella eri periaatteella. Ilmaa tuodaan tilaan yhdensuuntaisesti, turbulenttisesti tai sekavirtauksena HEPA -suodattimien kautta, joilla varmistetaan epäpuhtauksien korkea suodatusaste. Ilmapoistetaan rei'itettyjen, korotettujen lattioiden kautta tai tilan alaosassa olevien poistoilmasäleikköjen kautta, josta se johdetaan noin 75-90%:sti kierrätettynä takaisin tilaan. Lääketeollisuudessa rei'itettyjä, korotettuja lattioita eivoida käyttää kontaminaatiovaaran, vuoksi. Tilaan suunniteltuja olosuhteita ylläpidetään rakennusautomaation avulla ja monitorointijärjestelmällä valvotaan tilassa olevan ilman laatua. Kaikki GMP-luokituksen mukaiset puhdastilat tulee validoida. Validointiin kuuluu teknisten järjestelmien kvalifiointi ja koko prosessin validointi. Teknisten järjestel-mien kvalifiointi käsittää suunnitelmien tarkastuksen (DQ), asennus - ja käyttöönotto tarkastukset (IQ), toiminnan testauksen (OQ) ja suorituksen testauksen (PQ). Kvali-fiointi kuuluu yhtenä osa-alueena validointiin. Prosessin validointi on osa yrityksen laadunvarmistusta. Validoinnilla hankitaan dokumentoidut todisteet siitä, että tila tai prosessi todella täyttää annetut vaatimukset. Tässä työssä laadittiin esimerkinomainen kvalifiointisuunnitelma puhdastilan tekni-sille järjestelmille. Suunnitelma sisältää asennus- ja käyttöönoton mukaiset tarkastukset (IQ)ja toiminnan aikaiset testaukset (OQ).