22 resultados para source analysis

em Doria (National Library of Finland DSpace Services) - National Library of Finland, Finland


Relevância:

60.00% 60.00%

Publicador:

Resumo:

The research assesses the skills of upper comprehensive school pupils in history. The focus is on locating personal motives, assessing wider reasons hidden in historical sources and evaluating source reliability. The research also questions how a wide use of multiple sources affects pupils’ holistic understanding of historical phenomena. The participants were a multicultural group of pupils. The origins of their cultures can be traced to the Balkan, the Middle East, Asia and Europe. The number of native Finnish speakers and pupils speaking Finnish as their second language was almost equal. The multicultural composition provides opportunities to assess how culturally responsive learning history from sources is. The intercultural approach to learning in a multicultural setting emphasizes equality as a precondition for learning. In order to set assignments at least to some extent match with all participants only those answers were taken into account which were produced by pupils who had studied history for a similar period of time in the Finnish comprehensive school system. Due to the small number of participants (41), the study avoids wide generalizations. Nevertheless, possible cultural blueprints in pupils’ way of thinking are noted. The first test examined the skills of pupils to find motives for emigration. The results showed that for 7th graders finding reasons is not a problematic task. However, the number of reasons noticed and justifications varied. In addition, the way the pupils explained their choices was a distinguishing factor. Some pupils interpreted source material making use of previous knowledge on the issue, while other pupils based their analysis solely on the text handed and did not try to add their own knowledge. Answers were divided into three categories: historical, explanatory and stating. Historical answers combined smoothly previously learned historical knowledge to one’s own source analysis; explanatory answers often ignored a wider frame, although they were effective when explaining e.g. historical concepts. The stating answers only noticed motives from the sources and made no attempts to explain them historically. Was the first test culturally responsive? All pupils representing different cultures tackled the first source exam successfully, but there were some signs of how historical concepts are understood in a slightly different way if the pupil’s personal history has no linkage to the concepts under scrutiny. The second test focused on the history of Native Americans. The test first required pupils to recognize whether short source extracts (5) were written by Indians or Caucasians. Based on what they had already learned from North American history, the pupils did not find it hard to distinguish between the sources. The analysis of multiphase causes and consequences of the disputes between Native Americans and white Americans caused dispersion among pupils. Using two historical sources and combining historical knowledge from both of them simultaneously was cumbersome for many. The explanations of consequences can be divided into two groups: the ones emphasizing short term consequences and those placing emphasis on long term consequences. The short term approach was mainly followed by boys in every group. The girls mainly paid attention to long term consequences. The result suggests that historical knowledge in sources is at least to some extent read through role and gender lenses. The third test required pupils to explain in their own words how the three sources given differed in their account of living conditions in Nazi Germany, which turned out to be demanding for many pupils. The pupils’ stronghold was rather the assessment of source reliability and accounts why the sources approached the same events differently. All participants wrote critical and justified comments on reliability and aspects that might have affected the content of the sources. The pupils felt that the main reasons that affected source reliability were the authors’ ethnic background, nationality and profession. The assessment showed that pupils were well aware that position in a historical situation has an impact on historical accounts, but in certain cases the victim’s account was seen as a historical truth. The account of events by a historian was chosen most often as the most reliable source, but it was often justified leniently with an indication to professionalism rather than with clear ideas of how historians conduct accounts based on sources. In brief, the last source test demonstrates that pupils have a strong idea that the ethnicity or nationalism determines how people explained events of the past. It is also an implication that pupils understand how historical knowledge is interpretative. The results also imply that history can be analyzed from a neutral perspective. One’s own membership in an ethnical or religious group does not automatically mean that a person’s cultural identity excludes historical explanations if something in them contradicts with his or her identity. The second method of extracting knowledge of pupils’ historical thinking was an essay analysis. The analysis shows that an analytical account of complicated political issues, which often include a great number of complicated political concepts, leads more likely to an inconsistent structure in the written work of pupils. The material also demonstrates that pupils have a strong tendency to take a critical stance when assessing history. Historical empathy in particular is shown if history somehow has a linkage to young people, children or minorities. Some topics can also awake strong feelings, especially among pupils with emigrant background, if there is a linkage between one’s own personal history and that of the school; and occasionally a student’s historical experience or thoughts replaced school history. Using sources during history lessons at school seems to have many advantages. It enhances the reasoning skills of pupils and their skills to assess the nature of historical knowledge. Thus one of the main aims and a great benefit of source work is to encourage pupils to express their own ideas and opinions. To conclude, when assessing the skills of adolescents in history - their work with sources, comments on history, historical knowledge and finally their historical thinking - one should be cautious and avoid cut off score evaluations. One purpose of pursuing history with sources is to encourage pupils to think independently, which is a useful tool for further identity construction. The idea that pupils have the right to conduct their own interpretations of history can be partially understood as part of a wider learning process, justification to study history comes from extrinsic reasons. The intrinsic reason is history itself; in order to understand history one should have a basic understanding of history as a specific domain of knowledge. Using sources does not mean that knowing history is of secondary importance. Only a balance between knowing the contextual history, understanding basic key concepts and working with sources is a solid base to improve pupils’ historical understanding.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Credit risk assessment is an integral part of banking. Credit risk means that the return will not materialise in case the customer fails to fulfil its obligations. Thus a key component of banking is setting acceptance criteria for granting loans. Theoretical part of the study focuses on key components of credit assessment methods of Banks in the literature when extending credits to large corporations. Main component is Basel II Accord, which sets regulatory requirement for credit risk assessment methods of banks. Empirical part comprises, as primary source, analysis of major Nordic banks’ annual reports and risk management reports. As secondary source complimentary interviews were carried out with senior credit risk assessment personnel. The findings indicate that all major Nordic banks are using combination of quantitative and qualitative information in credit risk assessment model when extending credits to large corporations. The relative input of qualitative information depends on the selected approach to the credit rating, i.e. point-in-time or through-the-cycle.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Pulsewidth-modulated (PWM) rectifier technology is increasingly used in industrial applications like variable-speed motor drives, since it offers several desired features such as sinusoidal input currents, controllable power factor, bidirectional power flow and high quality DC output voltage. To achieve these features,however, an effective control system with fast and accurate current and DC voltage responses is required. From various control strategies proposed to meet these control objectives, in most cases the commonly known principle of the synchronous-frame current vector control along with some space-vector PWM scheme have been applied. Recently, however, new control approaches analogous to the well-established direct torque control (DTC) method for electrical machines have also emerged to implement a high-performance PWM rectifier. In this thesis the concepts of classical synchronous-frame current control and DTC-based PWM rectifier control are combined and a new converter-flux-based current control (CFCC) scheme is introduced. To achieve sufficient dynamic performance and to ensure a stable operation, the proposed control system is thoroughly analysed and simple rules for the controller design are suggested. Special attention is paid to the estimationof the converter flux, which is the key element of converter-flux-based control. Discrete-time implementation is also discussed. Line-voltage-sensorless reactive reactive power control methods for the L- and LCL-type line filters are presented. For the L-filter an open-loop control law for the d-axis current referenceis proposed. In the case of the LCL-filter the combined open-loop control and feedback control is proposed. The influence of the erroneous filter parameter estimates on the accuracy of the developed control schemes is also discussed. A newzero vector selection rule for suppressing the zero-sequence current in parallel-connected PWM rectifiers is proposed. With this method a truly standalone and independent control of the converter units is allowed and traditional transformer isolation and synchronised-control-based solutions are avoided. The implementation requires only one additional current sensor. The proposed schemes are evaluated by the simulations and laboratory experiments. A satisfactory performance and good agreement between the theory and practice are demonstrated.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

IIn electric drives, frequency converters are used to generatefor the electric motor the AC voltage with variable frequency and amplitude. When considering the annual sale of drives in values of money and units sold, the use of low-performance drives appears to be in predominant. These drives have tobe very cost effective to manufacture and use, while they are also expected to fulfill the harmonic distortion standards. One of the objectives has also been to extend the lifetime of the frequency converter. In a traditional frequency converter, a relatively large electrolytic DC-link capacitor is used. Electrolytic capacitors are large, heavy and rather expensive components. In many cases, the lifetime of the electrolytic capacitor is the main factor limiting the lifetime of the frequency converter. To overcome the problem, the electrolytic capacitor is replaced with a metallized polypropylene film capacitor (MPPF). The MPPF has improved properties when compared to the electrolytic capacitor. By replacing the electrolytic capacitor with a film capacitor the energy storage of the DC-linkwill be decreased. Thus, the instantaneous power supplied to the motor correlates with the instantaneous power taken from the network. This yields a continuousDC-link current fed by the diode rectifier bridge. As a consequence, the line current harmonics clearly decrease. Because of the decreased energy storage, the DC-link voltage fluctuates. This sets additional conditions to the controllers of the frequency converter to compensate the fluctuation from the supplied motor phase voltages. In this work three-phase and single-phase frequency converters with small DC-link capacitor are analyzed. The evaluation is obtained with simulations and laboratory measurements.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

The maximum realizable power throughput of power electronic converters may be limited or constrained by technical or economical considerations. One solution to this problemis to connect several power converter units in parallel. The parallel connection can be used to increase the current carrying capacity of the overall system beyond the ratings of individual power converter units. Thus, it is possible to use several lower-power converter units, produced in large quantities, as building blocks to construct high-power converters in a modular manner. High-power converters realized by using parallel connection are needed for example in multimegawatt wind power generation systems. Parallel connection of power converter units is also required in emerging applications such as photovoltaic and fuel cell power conversion. The parallel operation of power converter units is not, however, problem free. This is because parallel-operating units are subject to overcurrent stresses, which are caused by unequal load current sharing or currents that flow between the units. Commonly, the term ’circulatingcurrent’ is used to describe both the unequal load current sharing and the currents flowing between the units. Circulating currents, again, are caused by component tolerances and asynchronous operation of the parallel units. Parallel-operating units are also subject to stresses caused by unequal thermal stress distribution. Both of these problemscan, nevertheless, be handled with a proper circulating current control. To design an effective circulating current control system, we need information about circulating current dynamics. The dynamics of the circulating currents can be investigated by developing appropriate mathematical models. In this dissertation, circulating current models aredeveloped for two different types of parallel two-level three-phase inverter configurations. Themodels, which are developed for an arbitrary number of parallel units, provide a framework for analyzing circulating current generation mechanisms and developing circulating current control systems. In addition to developing circulating current models, modulation of parallel inverters is considered. It is illustrated that depending on the parallel inverter configuration and the modulation method applied, common-mode circulating currents may be excited as a consequence of the differential-mode circulating current control. To prevent the common-mode circulating currents that are caused by the modulation, a dual modulator method is introduced. The dual modulator basically consists of two independently operating modulators, the outputs of which eventually constitute the switching commands of the inverter. The two independently operating modulators are referred to as primary and secondary modulators. In its intended usage, the same voltage vector is fed to the primary modulators of each parallel unit, and the inputs of the secondary modulators are obtained from the circulating current controllers. To ensure that voltage commands obtained from the circulating current controllers are realizable, it must be guaranteed that the inverter is not driven into saturation by the primary modulator. The inverter saturation can be prevented by limiting the inputs of the primary and secondary modulators. Because of this, also a limitation algorithm is proposed. The operation of both the proposed dual modulator and the limitation algorithm is verified experimentally.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Tässä pro gradu -tutkielmassa käsittelen lähde- ja kohdetekstikeskeisyyttä näytelmäkääntämisessä. Tutkimuskohteina olivat käännösten sanasto, syntaksi, näyttämötekniikka, kielikuvat, sanaleikit, runomitta ja tyyli. Tutkimuksen tarkoituksena oli selvittää, näkyykö teoreettinen painopisteen siirtyminen lähdetekstikeskeisyydestä kohdetekstikeskeisyyteen suomenkielisessä näytelmäkääntämisessä. Oletuksena oli, että siirtyminen näkyy käytetyissä käännösstrategioissa. Tutkimuksen teoriaosuudessa käsitellään ensin lähde- ja kohdetekstikeskeisiä käännösteorioita. Ensin esitellään kaksi lähdetekstikeskeistä teoriaa, jotka ovat Catfordin (1965) muodollinen vastaavuus ja Nidan (1964) dynaaminen ekvivalenssi. Kohdetekstikeskeisistä teorioista käsitellään Touryn (1980) ja Newmarkin (1981) teoreettisia näkemyksiä sekä Reiss ja Vermeerin (1986) esittelemää skopos-teoriaa. Vieraannuttamisen ja kotouttamisen periaatteet esitellään lyhyesti. Teoriaosuudessa käsitellään myös näytelmäkääntämistä, William Shakespearen kieltä ja siihen liittyviä käännösongelmia. Lisäksi esittelen lyhyesti Shakespearen kääntämistä Suomessa ja Julius Caesarin neljä kääntäjää. Tutkimuksen materiaalina oli neljä Shakespearen Julius Caesar –näytelmän suomennosta, joista Paavo Cajanderin käännös on julkaistu vuonna 1883, Eeva-Liisa Mannerin vuonna 1983, Lauri Siparin vuonna 2006 ja Jarkko Laineen vuonna 2007. Analyysissa käännöksiä verrattiin lähdetekstiin ja toisiinsa ja vertailtiin kääntäjien tekemiä käännösratkaisuja. Tulokset olivat oletuksen mukaisia. Lähdetekstikeskeisiä käännösstrategioita oli käytetty uusissa käännöksissä vähemmän kuin vanhemmissa. Kohdetekstikeskeiset strategiat erosivat huomattavasti toisistaan ja uusinta käännöstä voi sanoa adaptaatioksi. Jatkotutkimuksissa tulisi materiaali laajentaa koskemaan muitakin Shakespearen näytelmien suomennoksia. Eri aikakausien käännöksiä tulisi verrata keskenään ja toisiinsa, jotta voitaisiin luotettavasti kuvata muutosta lähde- ja kohdetekstikeskeisten käännösstrategioiden käytössä ja eri aikakausien tyypillisten strategioiden kartoittamiseksi.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Työn tarkoituksena oli tutkia sisältö- ja diskurssianalyysin avulla kuinka yritykset viestivät asiakasreferenssejä verkkosivuillaan. Työssä keskityttiin tutkimaan yritysten referenssikuvausten teemoja ja diskursseja, sekä sitä kuinka referenssisuhde rakentuu diskursiivisesti referenssikuvauksissa. Tutkimukseen valittiin kolme suomalaista ICT-alan yritystä: Nokia, TietoEnator ja F-Secure. Aineisto koostuu 140:stä yritysten WWW-sivuilta kerätystä referenssikuvauksesta. Sisältöanalyysin tuloksena havaittiin, että referenssikuvaukset keskittyvät kuvaamaan yksittäisiä tuote- tai projektitoimituksia referenssiasiakkaille kyseisten asiakassuhteiden valossa. Analyysin tuloksena tunnistettiin kolme diskurssia: hyötydiskurssi, sitoutumisen diskurssi sekä teknologisen eksperttiyden diskurssi. Diskurssit paljastavat referenssikuvausten retoriset keinot ja konstruoivat referenssisuhteen ja toimittajan subjektiposition eri näkökulmista. Pääpaino referenssikuvauksissa on toimittajan ratkaisun tuomissa hyödyissä. Diskurssit tuottavat referenssisuhteesta kuvan hyötyjä tuovana ja läheisenä asiakassuhteena, joka tarjoaa väylän ulkopuolisiin kyvykkyyksiin ja teknologioihin. Toimittaja esitetään referenssikuvauksissa diskurssista riippuen hyötyjen tuojana, luotettavana partnerina sekä kokeneena eksperttinä. Referenssiasiakas sen sijaan esitetään vain yhdestä näkökulmasta stereotyyppisesti tärkeänä ja tyytyväisenä asiakkaana.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The diffusion of mobile telephony began in 1971 in Finland, when the first car phones, called ARP1 were taken to use. Technologies changed from ARP to NMT and later to GSM. The main application of the technology, however, was voice transfer. The birth of the Internet created an open public data network and easy access to other types of computer-based services over networks. Telephones had been used as modems, but the development of the cellular technologies enabled automatic access from mobile phones to Internet. Also other wireless technologies, for instance Wireless LANs, were also introduced. Telephony had developed from analog to digital in fixed networks and allowed easy integration of fixed and mobile networks. This development opened a completely new functionality to computers and mobile phones. It also initiated the merger of the information technology (IT) and telecommunication (TC) industries. Despite the arising opportunity for firms' new competition the applications based on the new functionality were rare. Furthermore, technology development combined with innovation can be disruptive to industries. This research focuses on the new technology's impact on competition in the ICT industry through understanding the strategic needs and alternative futures of the industry's customers. The change speed inthe ICT industry is high and therefore it was valuable to integrate the DynamicCapability view of the firm in this research. Dynamic capabilities are an application of the Resource-Based View (RBV) of the firm. As is stated in the literature, strategic positioning complements RBV. This theoretical framework leads theresearch to focus on three areas: customer strategic innovation and business model development, external future analysis, and process development combining these two. The theoretical contribution of the research is in the development of methodology integrating theories of the RBV, dynamic capabilities and strategic positioning. The research approach has been constructive due to the actual managerial problems initiating the study. The requirement for iterative and innovative progress in the research supported the chosen research approach. The study applies known methods in product development, for instance, innovation process in theGroup Decision Support Systems (GDSS) laboratory and Quality Function Deployment (QFD), and combines them with known strategy analysis tools like industry analysis and scenario method. As the main result, the thesis presents the strategic innovation process, where new business concepts are used to describe the alternative resource configurations and scenarios as alternative competitive environments, which can be a new way for firms to achieve competitive advantage in high-velocity markets. In addition to the strategic innovation process as a result, thestudy has also resulted in approximately 250 new innovations for the participating firms, reduced technology uncertainty and helped strategic infrastructural decisions in the firms, and produced a knowledge-bank including data from 43 ICT and 19 paper industry firms between the years 1999 - 2004. The methods presentedin this research are also applicable to other industries.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Belt-drive systems have been and still are the most commonly used power transmission form in various applications of different scale and use. The peculiar features of the dynamics of the belt-drives include highly nonlinear deformation,large rigid body motion, a dynamical contact through a dry friction interface between the belt and pulleys with sticking and slipping zones, cyclic tension of the belt during the operation and creeping of the belt against the pulleys. The life of the belt-drive is critically related on these features, and therefore, amodel which can be used to study the correlations between the initial values and the responses of the belt-drives is a valuable source of information for the development process of the belt-drives. Traditionally, the finite element models of the belt-drives consist of a large number of elements thatmay lead to computational inefficiency. In this research, the beneficial features of the absolute nodal coordinate formulation are utilized in the modeling of the belt-drives in order to fulfill the following requirements for the successful and efficient analysis of the belt-drive systems: the exact modeling of the rigid body inertia during an arbitrary rigid body motion, the consideration of theeffect of the shear deformation, the exact description of the highly nonlinear deformations and a simple and realistic description of the contact. The use of distributed contact forces and high order beam and plate elements based on the absolute nodal coordinate formulation are applied to the modeling of the belt-drives in two- and three-dimensional cases. According to the numerical results, a realistic behavior of the belt-drives can be obtained with a significantly smaller number of elements and degrees of freedom in comparison to the previously published finite element models of belt-drives. The results of theexamples demonstrate the functionality and suitability of the absolute nodal coordinate formulation for the computationally efficient and realistic modeling ofbelt-drives. This study also introduces an approach to avoid the problems related to the use of the continuum mechanics approach in the definition of elastic forces on the absolute nodal coordinate formulation. This approach is applied to a new computationally efficient two-dimensional shear deformable beam element based on the absolute nodal coordinate formulation. The proposed beam element uses a linear displacement field neglecting higher-order terms and a reduced number of nodal coordinates, which leads to fewer degrees of freedom in a finite element.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Diplomityössä on käsitelty uudenlaisia menetelmiä riippumattomien komponenttien analyysiin(ICA): Menetelmät perustuvat colligaatioon ja cross-momenttiin. Colligaatio menetelmä perustuu painojen colligaatioon. Menetelmässä on käytetty kahden tyyppisiä todennäköisyysjakaumia yhden sijasta joka perustuu yleiseen itsenäisyyden kriteeriin. Työssä on käytetty colligaatio lähestymistapaa kahdella asymptoottisella esityksellä. Gram-Charlie ja Edgeworth laajennuksia käytetty arvioimaan todennäköisyyksiä näissä menetelmissä. Työssä on myös käytetty cross-momentti menetelmää joka perustuu neljännen asteen cross-momenttiin. Menetelmä on hyvin samankaltainen FastICA algoritmin kanssa. Molempia menetelmiä on tarkasteltu lineaarisella kahden itsenäisen muuttajan sekoituksella. Lähtö signaalit ja sekoitetut matriisit ovattuntemattomia signaali lähteiden määrää lukuunottamatta. Työssä on vertailtu colligaatio menetelmään ja sen modifikaatioita FastICA:an ja JADE:en. Työssä on myös tehty vertailu analyysi suorituskyvyn ja keskusprosessori ajan suhteen cross-momenttiin perustuvien menetelmien, FastICA:n ja JADE:n useiden sekoitettujen parien kanssa.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Tämä diplomityö liittyy Spektrikuvien tutkimiseen tilastollisen kuvamallin näkökulmasta. Diplomityön ensimmäisessä osassa tarkastellaan tilastollisten parametrien jakaumien vaikutusta väreihin ja korostumiin erilaisissa valaistusolosuhteissa. Havaittiin, että tilastollisten parametrien väliset suhteet eivät riipu valaistusolosuhteista, mutta riippuvat kuvan häiriöttömyydestä. Ilmeni myös, että korkea huipukkuus saattaa aiheutua värikylläisyydestä. Lisäksi työssä kehitettiin tilastolliseen spektrimalliin perustuvaa tekstuurinyhdistämisalgoritmia. Sillä saavutettiin hyviä tuloksia, kun tilastollisten parametrien väliset riippuvuussuhteet olivat voimassa. Työn toisessa osassa erilaisia spektrikuvia tutkittiin käyttäen itsenäistä komponenttien analyysia (ICA). Seuraavia itsenäiseen komponenttien analyysiin tarkoitettuja algoritmia tarkasteltiin: JADE, kiinteän pisteen ICA ja momenttikeskeinen ICA. Tutkimuksissa painotettiin erottelun laatua. Paras erottelu saavutettiin JADE- algoritmilla, joskin erot muiden algoritmien välillä eivät olleet merkittäviä. Algoritmi jakoi kuvan kahteen itsenäiseen, joko korostuneeseen ja korostumattomaan tai kromaattiseen ja akromaattiseen, komponenttiin. Lopuksi pohditaan huipukkuuden suhdetta kuvan ominaisuuksiin, kuten korostuneisuuteen ja värikylläisyyteen. Työn viimeisessä osassa ehdotetaan mahdollisia jatkotutkimuskohteita.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

A total of over 200 different samples of bark and wood of Silver birch, Norway spruce and Scots pine were analysed. Samples were taken from several areas in western Finland, some with known sources of atmospheric heavy metal emission (Harjavalta, Ykspihlaja). Also analytical data for pine needles from some sites are reported. The chemical analyses were performed by thick-target particle-induced X-ray emission (PIXE) spectrometry after preconcentration by dry ashing of samples at 550oC. The following elements were quantified in most of the samples: P, S, K, Ca, Mn, Fe, Ni, Cu, Zn, Rb, Sr, Cd, Ba and Pb. The ash percentage and the chemical composition of ashes of different wood materials were also obtained, as dry ashing was used in the analytical procedure. The variations in elemental concentrations in wood and bark of an individual tree, expressed as RSDs, were mostly in the range 10 – 20 %. For several trees of the same species sampled from small areas (< 1 ha), the variations in elemental concentrations were surprisingly high (RSDs 20 – 50 %). In the vicinity of metal plants, effects of strong atmospheric heavy metal pollution (pollution factor above 100) were observed in pine bark. The increase of heavy metal content in wood samples from the same sites was quite small. Elemental concentrations in ashes of bark and wood, from areas with no local source of atmospheric pollution, were relatively uniform. Based on this observation an alternative way of demonstrating atmospheric pollution of tree bark is discussed.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The amount of installed wind power has been growing exponentially during the past ten years. As wind turbines have become a significant source of electrical energy, the interactions between the turbines and the electric power network need to be studied more thoroughly than before. Especially, the behavior of the turbines in fault situations is of prime importance; simply disconnecting all wind turbines from the network during a voltage drop is no longer acceptable, since this would contribute to a total network collapse. These requirements have been a contributor to the increased role of simulations in the study and design of the electric drive train of a wind turbine. When planning a wind power investment, the selection of the site and the turbine are crucial for the economic feasibility of the installation. Economic feasibility, on the other hand, is the factor that determines whether or not investment in wind power will continue, contributing to green electricity production and reduction of emissions. In the selection of the installation site and the turbine (siting and site matching), the properties of the electric drive train of the planned turbine have so far been generally not been taken into account. Additionally, although the loss minimization of some of the individual components of the drive train has been studied, the drive train as a whole has received less attention. Furthermore, as a wind turbine will typically operate at a power level lower than the nominal most of the time, efficiency analysis in the nominal operating point is not sufficient. This doctoral dissertation attempts to combine the two aforementioned areas of interest by studying the applicability of time domain simulations in the analysis of the economicfeasibility of a wind turbine. The utilization of a general-purpose time domain simulator, otherwise applied to the study of network interactions and control systems, in the economic analysis of the wind energy conversion system is studied. The main benefits of the simulation-based method over traditional methods based on analytic calculation of losses include the ability to reuse and recombine existing models, the ability to analyze interactions between the components and subsystems in the electric drive train (something which is impossible when considering different subsystems as independent blocks, as is commonly done in theanalytical calculation of efficiencies), the ability to analyze in a rather straightforward manner the effect of selections other than physical components, for example control algorithms, and the ability to verify assumptions of the effects of a particular design change on the efficiency of the whole system. Based on the work, it can be concluded that differences between two configurations can be seen in the economic performance with only minor modifications to the simulation models used in the network interaction and control method study. This eliminates the need ofdeveloping analytic expressions for losses and enables the study of the system as a whole instead of modeling it as series connection of independent blocks with no lossinterdependencies. Three example cases (site matching, component selection, control principle selection) are provided to illustrate the usage of the approach and analyze its performance.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Cutin and suberin are structural and protective polymers of plant surfaces. The epidermal cells of the aerial parts of plants are covered with an extracellular cuticular layer, which consists of polyester cutin, highly resistant cutan, cuticular waxes and polysaccharides which link the layer to the epidermal cells. A similar protective layer is formed by a polyaromatic-polyaliphatic biopolymer suberin, which is present particularly in the cell walls of the phellem layer of periderm of the underground parts of plants (e.g. roots and tubers) and the bark of trees. In addition, suberization is also a major factor in wound healing and wound periderm formation regardless of the plants’ tissue. Knowledge of the composition and functions of cuticular and suberin polymers is important for understanding the physiological properties for the plants and for nutritional quality when these plants are consumed as foods. The aims of the practical work were to assess the chemical composition of cuticular polymers of several northern berries and seeds and suberin of two varieties of potatoes. Cutin and suberin were studied as isolated polymers and further after depolymerization as soluble monomers and solid residues. Chemical and enzymatic depolymerization techniques were compared and a new chemical depolymerization method was developed. Gas chromatographic analysis with mass spectrometric detection (GC-MS) was used to assess the monomer compositions. Polymer investigations were conducted with solid state carbon-13 cross polarization magic angle spinning nuclear magnetic resonance spectroscopy (13C CP-MAS NMR), Fourier transform infrared spectroscopy (FTIR) and microscopic analysis. Furthermore, the development of suberin over one year of post-harvest storage was investigated and the cuticular layers from berries grown in the North and South of Finland were compared. The results show that the amounts of isolated cuticular layers and cutin monomers, as well as monomeric compositions vary greatly between the berries. The monomer composition of seeds was found to differ from the corresponding berry peel monomers. The berry cutin monomers were composed mostly of long-chain aliphatic ω-hydroxy acids, with various mid-chain functionalities (double-bonds, epoxy, hydroxy and keto groups). Substituted α,ω-diacids predominated over ω-hydroxy acids in potato suberin monomers and slight differences were found between the varieties. The newly-developed closed tube chemical method was found to be suitable for cutin and suberin analysis and preferred over the solvent-consuming and laborious reflux method. Enzymatic hydrolysis with cutinase was less effective than chemical methanolysis and showed specificity towards α,ω-diacid bonds. According to 13C CP-MAS NMR and FTIR, the depolymerization residues contained significant amounts of aromatic structures, polysaccharides and possible cutan-type aliphatic moieties. Cultivation location seems to have effect on cuticular composition. The materials studied contained significant amounts of different types of biopolymers that could be utilized for several purposes with or without further processing. The importance of the so-called waste material from industrial processes of berries and potatoes as a source of either dietary fiber or specialty chemicals should be further investigated in detail. The evident impact of cuticular and suberin polymers, among other fiber components, on human health should be investigated in clinical trials. These by-product materials may be used as value-added fiber fractions in the food industry and as raw materials for specialty chemicals such as lubricants and emulsifiers, or as building blocks for novel polymers.