990 resultados para SOFTWARE-RELIABILITY
Resumo:
The application of VSC-HVDC technology throughout the world has turned out to be an efficient solution regarding a large share of wind power in different power systems. This technology enhances the overall reliability of the grid by utilization of the active and reactive power control schemes which allows to maintain frequency and voltage on busbars of the end-consumers at the required level stated by the network operator. This master’s thesis is focused on the existing and planned wind farms as well as electric power system of the Åland Islands. The goal is to analyze the wind conditions of the islands and appropriately predict a possible production of the existing and planned wind farms with a help of WAsP software program. Further, to investigate the influence of increased wind power it is necessary to develop a simulation model of the electric grid and VSC-HVDC system in PSCAD and examine grid response to different wind power production cases with respect to the grid code requirements and ensure the stability of the power system.
Resumo:
This study discusses the formation phase of Chinese-Finnish joint ventures in China. The purpose of this thesis is to create best practices for Finnish software companies in forming a joint venture with a local Chinese company in China. Therefore, the main research question, in what are the best practices for forming Sino-Finnish joint ventures in China for Finnish software firms, is examined through four different themes within the joint venture formation phase; the motives, the partner se- lection, the choice of a joint venture type and joint venture negotiations. The theoretical background of the study consists of literature relating to the establishment process of Sino-Western joint ventures in China. The empirical research conducted for this study is based on the expert interviews. The empirical data was gathered via nine semi-structured interviews with both Chinese and Finnish experts in software and technology industry, who have experience or knowledge in establishing Sino-Finnish joint ventures in China. Thematic analysis was used to cat- egorize and interpret the interview data. In addition, a thematic network was built to act as a basis of the analysis. According to the main findings, the main motives for Finnish software companies to establish a joint venture in China are lack of skills or experience, little resources to enter on their own, and China’s large market. The main motives for Chinese companies are to gain new technology or man- agerial skills, and expand internationally. The intellectual property rights (IPR) have recently im- proved a lot in China, but the Finnish companies’ knowledge on IPR is inadequate. The Finnish software companies should conduct a market and industry research in order to understand their po- sition in the market and to find a suitable location and potential joint venture partners. It is essential to define partner selection criteria and partner attributes. In addition, it is important to build the joint venture around complementary motives and a win-win situation between the joint venture partners. The Finnish companies should be prepared that the joint venture negotiations will be challenging and they will take a long time. The challenges can be overcome by gaining understanding about the Chinese culture and business environment. The findings of this study enhance understanding of the joint venture formation phase in China. This study provides guidelines for Finnish software companies to establish a joint venture in China. In addition, this study brings new insights to the Sino-Western joint venture literature with its soft- ware industry context. Future research is, however, necessary in order to gain an understanding of the advantages and disadvantages of a joint venture as an entry mode into China for Finnish soft- ware companies
Resumo:
The objective of this study was to analyze retinol equivalent and iron content in different food composition tables and nutritional evaluation software programs. A literature search was conduct to identify tables and software available in Brazil containing information about retinol equivalent and iron content that are currently used by nutritionists. Ten tables and five software programs were selected for this study. The methodology used to present the retinol equivalent and iron content was evaluated and no pattern to obtain such content was found in the tables and software programs analyzed. Only one of the tables had enough information for the calculation of retinol equivalents; this table is recommended to all Latin America As for the iron content, three of the tables analyzed stand out and therefore should be used; two of them are based on national foods and the other is recommended for use in all Latin America countries. None of the software programs evaluated use the conversion factors suggested by IVACG to assess the vitamin A content in foods. Special attention should be given to the content of iron provided in the software programs since they use tables as international sources and fortified foods.
Resumo:
Assessing fish consumption is complex and involves several factors; however, the use of questionnaires in surveys and the use of the Internet as tool to collect data have been considered promising approaches. Therefore, the objective of this research was to design a data collection technique using a questionnaire to assess fish consumption by making it available on a specific home page on the Internet. A bibliographical survey or review was carried out to identify the features of the instrument, and therefore pre-tests were conducted with previous instruments, followed by the Focus Group technique. Specialists then performed an analysis and conducted an online pre-test. Multivariate data analysis was applied using the SmartPLS software. The results indicate that 1.966 participants belonging to the University of São Paulo (USP) community participated in the test, and after the exclusion of some variables, a statistically significant results were obtained. The final constructs comprised consumption, quality, and general characteristics. The instrument consisted of behavioral statements in a 5-point Likert scale and multiple-choice questions. The Cronbach's alpha reliability coefficient was 0.66 for general characteristics, 0.98 for quality, and 0.91 for consumption, which indicate good reliability of the instrument. In conclusion, the results proved that the Internet assessment is efficient. The instrument of analysis allowed us to better understand the process of buying and consuming fish in the country, and it can be used as base for further research.
Resumo:
Human beings have always strived to preserve their memories and spread their ideas. In the beginning this was always done through human interpretations, such as telling stories and creating sculptures. Later, technological progress made it possible to create a recording of a phenomenon; first as an analogue recording onto a physical object, and later digitally, as a sequence of bits to be interpreted by a computer. By the end of the 20th century technological advances had made it feasible to distribute media content over a computer network instead of on physical objects, thus enabling the concept of digital media distribution. Many digital media distribution systems already exist, and their continued, and in many cases increasing, usage is an indicator for the high interest in their future enhancements and enriching. By looking at these digital media distribution systems, we have identified three main areas of possible improvement: network structure and coordination, transport of content over the network, and the encoding used for the content. In this thesis, our aim is to show that improvements in performance, efficiency and availability can be done in conjunction with improvements in software quality and reliability through the use of formal methods: mathematical approaches to reasoning about software so that we can prove its correctness, together with the desirable properties. We envision a complete media distribution system based on a distributed architecture, such as peer-to-peer networking, in which different parts of the system have been formally modelled and verified. Starting with the network itself, we show how it can be formally constructed and modularised in the Event-B formalism, such that we can separate the modelling of one node from the modelling of the network itself. We also show how the piece selection algorithm in the BitTorrent peer-to-peer transfer protocol can be adapted for on-demand media streaming, and how this can be modelled in Event-B. Furthermore, we show how modelling one peer in Event-B can give results similar to simulating an entire network of peers. Going further, we introduce a formal specification language for content transfer algorithms, and show that having such a language can make these algorithms easier to understand. We also show how generating Event-B code from this language can result in less complexity compared to creating the models from written specifications. We also consider the decoding part of a media distribution system by showing how video decoding can be done in parallel. This is based on formally defined dependencies between frames and blocks in a video sequence; we have shown that also this step can be performed in a way that is mathematically proven correct. Our modelling and proving in this thesis is, in its majority, tool-based. This provides a demonstration of the advance of formal methods as well as their increased reliability, and thus, advocates for their more wide-spread usage in the future.
Resumo:
Tämä diplomityö arvioi hitsauksen laadunhallintaohjelmistomarkkinoiden kilpailijoita. Kilpailukenttä on uusi ja ei ole tarkkaa tietoa siitä minkälaisia kilpailijoita on markkinoilla. Hitsauksen laadunhallintaohjelmisto auttaa yrityksiä takaamaan korkean laadun. Ohjelmisto takaa korkean laadun varmistamalla, että hitsaaja on pätevä, hän noudattaa hitsausohjeita ja annettuja parametreja. Sen lisäksi ohjelmisto kerää kaiken tiedon hitsausprosessista ja luo siitä vaadittavat dokumentit. Diplomityön teoriaosuus muodostuu kirjallisuuskatsauksesta ratkaisuliike-toimintaan, kilpailija-analyysin ja kilpailuvoimien teoriaan sekä hitsauksen laadunhallintaan. Työn empiriaosuus on laadullinen tutkimus, jossa tutkitaan kilpailevia hitsauksen laadunhallintaohjelmistoja ja haastatellaan ohjelmistojen käyttäjiä. Diplomityön tuloksena saadaan uusi kilpailija-analyysimalli hitsauksen laadunhallintaohjelmistoille. Mallin avulla voidaan arvostella ohjelmistot niiden tarjoamien primääri- ja sekundääriominaisuuksien perusteella. Toiseksi tässä diplomityössä analysoidaan nykyinen kilpailijatilanne hyödyntämällä juuri kehitettyä kilpailija-analyysimallia.
Resumo:
Many-core systems provide a great potential in application performance with the massively parallel structure. Such systems are currently being integrated into most parts of daily life from high-end server farms to desktop systems, laptops and mobile devices. Yet, these systems are facing increasing challenges such as high temperature causing physical damage, high electrical bills both for servers and individual users, unpleasant noise levels due to active cooling and unrealistic battery drainage in mobile devices; factors caused directly by poor energy efficiency. Power management has traditionally been an area of research providing hardware solutions or runtime power management in the operating system in form of frequency governors. Energy awareness in application software is currently non-existent. This means that applications are not involved in the power management decisions, nor does any interface between the applications and the runtime system to provide such facilities exist. Power management in the operating system is therefore performed purely based on indirect implications of software execution, usually referred to as the workload. It often results in over-allocation of resources, hence power waste. This thesis discusses power management strategies in many-core systems in the form of increasing application software awareness of energy efficiency. The presented approach allows meta-data descriptions in the applications and is manifested in two design recommendations: 1) Energy-aware mapping 2) Energy-aware execution which allow the applications to directly influence the power management decisions. The recommendations eliminate over-allocation of resources and increase the energy efficiency of the computing system. Both recommendations are fully supported in a provided interface in combination with a novel power management runtime system called Bricktop. The work presented in this thesis allows both new- and legacy software to execute with the most energy efficient mapping on a many-core CPU and with the most energy efficient performance level. A set of case study examples demonstrate realworld energy savings in a wide range of applications without performance degradation.