845 resultados para exploratory design methods


Relevância:

30.00% 30.00%

Publicador:

Resumo:

Soil consolidation and erosion caused by roadway runoff have exposed the upper portions of steel piles at the abutments of numerous bridges, leaving them susceptible to accelerated corrosion rates due to the abundance of moisture, oxygen, and chlorides at these locations. This problem is compounded by the relative inaccessibility of abutment piles for close-up inspection and repair. The objective of this study was to provide bridge owners with recommendations for effective methods of addressing corrosion of steel abutment piles in existing and future bridges A review of available literature on the performance and protection of steel piles exposed to a variety of environments was performed. Eight potential coating systems for use in protecting existing and/or new piles were selected and subjected to accelerated corrosion conditions in the laboratory. Two surface preparation methods were evaluated in the field and three coating systems were installed on three piles at an existing bridge where abutment piles had been exposed by erosion. In addition, a passive cathodic protection (CP) system using sacrificial zinc anodes was tested in the laboratory. Several trial flowable mortar mixes were evaluated for use in conjunction with the CP system. For existing abutment piles, application of a protective coating system is a promising method of mitigating corrosion. Based on its excellent performance in accelerated corrosion conditions in the laboratory on steel test specimens with SSPC-SP3, -SP6, and -SP10 surface preparations, glass flake polyester is recommended for use on existing piles. An alternative is epoxy over organic zinc rich primer. Surface preparation of existing piles should include abrasive blast cleaning to SSPC-SP6. Although additional field testing is needed, based on the results of the laboratory testing, a passive CP system could provide an effective means of protecting piles in existing bridges when combined with a pumped mortar used to fill voids between the abutment footing and soil. The addition of a corrosion inhibitor to the mortar appears to be beneficial. For new construction, shop application of thermally sprayed aluminum or glass flake polyester to the upper portion of the piles is recommended.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

3 Summary 3. 1 English The pharmaceutical industry has been facing several challenges during the last years, and the optimization of their drug discovery pipeline is believed to be the only viable solution. High-throughput techniques do participate actively to this optimization, especially when complemented by computational approaches aiming at rationalizing the enormous amount of information that they can produce. In siiico techniques, such as virtual screening or rational drug design, are now routinely used to guide drug discovery. Both heavily rely on the prediction of the molecular interaction (docking) occurring between drug-like molecules and a therapeutically relevant target. Several softwares are available to this end, but despite the very promising picture drawn in most benchmarks, they still hold several hidden weaknesses. As pointed out in several recent reviews, the docking problem is far from being solved, and there is now a need for methods able to identify binding modes with a high accuracy, which is essential to reliably compute the binding free energy of the ligand. This quantity is directly linked to its affinity and can be related to its biological activity. Accurate docking algorithms are thus critical for both the discovery and the rational optimization of new drugs. In this thesis, a new docking software aiming at this goal is presented, EADock. It uses a hybrid evolutionary algorithm with two fitness functions, in combination with a sophisticated management of the diversity. EADock is interfaced with .the CHARMM package for energy calculations and coordinate handling. A validation was carried out on 37 crystallized protein-ligand complexes featuring 11 different proteins. The search space was defined as a sphere of 15 R around the center of mass of the ligand position in the crystal structure, and conversely to other benchmarks, our algorithms was fed with optimized ligand positions up to 10 A root mean square deviation 2MSD) from the crystal structure. This validation illustrates the efficiency of our sampling heuristic, as correct binding modes, defined by a RMSD to the crystal structure lower than 2 A, were identified and ranked first for 68% of the complexes. The success rate increases to 78% when considering the five best-ranked clusters, and 92% when all clusters present in the last generation are taken into account. Most failures in this benchmark could be explained by the presence of crystal contacts in the experimental structure. EADock has been used to understand molecular interactions involved in the regulation of the Na,K ATPase, and in the activation of the nuclear hormone peroxisome proliferatoractivated receptors a (PPARa). It also helped to understand the action of common pollutants (phthalates) on PPARy, and the impact of biotransformations of the anticancer drug Imatinib (Gleevec®) on its binding mode to the Bcr-Abl tyrosine kinase. Finally, a fragment-based rational drug design approach using EADock was developed, and led to the successful design of new peptidic ligands for the a5ß1 integrin, and for the human PPARa. In both cases, the designed peptides presented activities comparable to that of well-established ligands such as the anticancer drug Cilengitide and Wy14,643, respectively. 3.2 French Les récentes difficultés de l'industrie pharmaceutique ne semblent pouvoir se résoudre que par l'optimisation de leur processus de développement de médicaments. Cette dernière implique de plus en plus. de techniques dites "haut-débit", particulièrement efficaces lorsqu'elles sont couplées aux outils informatiques permettant de gérer la masse de données produite. Désormais, les approches in silico telles que le criblage virtuel ou la conception rationnelle de nouvelles molécules sont utilisées couramment. Toutes deux reposent sur la capacité à prédire les détails de l'interaction moléculaire entre une molécule ressemblant à un principe actif (PA) et une protéine cible ayant un intérêt thérapeutique. Les comparatifs de logiciels s'attaquant à cette prédiction sont flatteurs, mais plusieurs problèmes subsistent. La littérature récente tend à remettre en cause leur fiabilité, affirmant l'émergence .d'un besoin pour des approches plus précises du mode d'interaction. Cette précision est essentielle au calcul de l'énergie libre de liaison, qui est directement liée à l'affinité du PA potentiel pour la protéine cible, et indirectement liée à son activité biologique. Une prédiction précise est d'une importance toute particulière pour la découverte et l'optimisation de nouvelles molécules actives. Cette thèse présente un nouveau logiciel, EADock, mettant en avant une telle précision. Cet algorithme évolutionnaire hybride utilise deux pressions de sélections, combinées à une gestion de la diversité sophistiquée. EADock repose sur CHARMM pour les calculs d'énergie et la gestion des coordonnées atomiques. Sa validation a été effectuée sur 37 complexes protéine-ligand cristallisés, incluant 11 protéines différentes. L'espace de recherche a été étendu à une sphère de 151 de rayon autour du centre de masse du ligand cristallisé, et contrairement aux comparatifs habituels, l'algorithme est parti de solutions optimisées présentant un RMSD jusqu'à 10 R par rapport à la structure cristalline. Cette validation a permis de mettre en évidence l'efficacité de notre heuristique de recherche car des modes d'interactions présentant un RMSD inférieur à 2 R par rapport à la structure cristalline ont été classés premier pour 68% des complexes. Lorsque les cinq meilleures solutions sont prises en compte, le taux de succès grimpe à 78%, et 92% lorsque la totalité de la dernière génération est prise en compte. La plupart des erreurs de prédiction sont imputables à la présence de contacts cristallins. Depuis, EADock a été utilisé pour comprendre les mécanismes moléculaires impliqués dans la régulation de la Na,K ATPase et dans l'activation du peroxisome proliferatoractivated receptor a (PPARa). Il a également permis de décrire l'interaction de polluants couramment rencontrés sur PPARy, ainsi que l'influence de la métabolisation de l'Imatinib (PA anticancéreux) sur la fixation à la kinase Bcr-Abl. Une approche basée sur la prédiction des interactions de fragments moléculaires avec protéine cible est également proposée. Elle a permis la découverte de nouveaux ligands peptidiques de PPARa et de l'intégrine a5ß1. Dans les deux cas, l'activité de ces nouveaux peptides est comparable à celles de ligands bien établis, comme le Wy14,643 pour le premier, et le Cilengitide (PA anticancéreux) pour la seconde.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

BACKGROUND: Ischemic stroke is the leading cause of mortality worldwide and a major contributor to neurological disability and dementia. Terutroban is a specific TP receptor antagonist with antithrombotic, antivasoconstrictive, and antiatherosclerotic properties, which may be of interest for the secondary prevention of ischemic stroke. This article describes the rationale and design of the Prevention of cerebrovascular and cardiovascular Events of ischemic origin with teRutroban in patients with a history oF ischemic strOke or tRansient ischeMic Attack (PERFORM) Study, which aims to demonstrate the superiority of the efficacy of terutroban versus aspirin in secondary prevention of cerebrovascular and cardiovascular events. METHODS AND RESULTS: The PERFORM Study is a multicenter, randomized, double-blind, parallel-group study being carried out in 802 centers in 46 countries. The study population includes patients aged > or =55 years, having suffered an ischemic stroke (< or =3 months) or a transient ischemic attack (< or =8 days). Participants are randomly allocated to terutroban (30 mg/day) or aspirin (100 mg/day). The primary efficacy endpoint is a composite of ischemic stroke (fatal or nonfatal), myocardial infarction (fatal or nonfatal), or other vascular death (excluding hemorrhagic death of any origin). Safety is being evaluated by assessing hemorrhagic events. Follow-up is expected to last for 2-4 years. Assuming a relative risk reduction of 13%, the expected number of primary events is 2,340. To obtain statistical power of 90%, this requires inclusion of at least 18,000 patients in this event-driven trial. The first patient was randomized in February 2006. CONCLUSIONS: The PERFORM Study will explore the benefits and safety of terutroban in secondary cardiovascular prevention after a cerebral ischemic event.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

BACKGROUND: Health professionals and policymakers aspire to make healthcare decisions based on the entire relevant research evidence. This, however, can rarely be achieved because a considerable amount of research findings are not published, especially in case of 'negative' results - a phenomenon widely recognized as publication bias. Different methods of detecting, quantifying and adjusting for publication bias in meta-analyses have been described in the literature, such as graphical approaches and formal statistical tests to detect publication bias, and statistical approaches to modify effect sizes to adjust a pooled estimate when the presence of publication bias is suspected. An up-to-date systematic review of the existing methods is lacking. METHODS/DESIGN: The objectives of this systematic review are as follows:âeuro¢ To systematically review methodological articles which focus on non-publication of studies and to describe methods of detecting and/or quantifying and/or adjusting for publication bias in meta-analyses.âeuro¢ To appraise strengths and weaknesses of methods, the resources they require, and the conditions under which the method could be used, based on findings of included studies.We will systematically search Web of Science, Medline, and the Cochrane Library for methodological articles that describe at least one method of detecting and/or quantifying and/or adjusting for publication bias in meta-analyses. A dedicated data extraction form is developed and pilot-tested. Working in teams of two, we will independently extract relevant information from each eligible article. As this will be a qualitative systematic review, data reporting will involve a descriptive summary. DISCUSSION: Results are expected to be publicly available in mid 2013. This systematic review together with the results of other systematic reviews of the OPEN project (To Overcome Failure to Publish Negative Findings) will serve as a basis for the development of future policies and guidelines regarding the assessment and handling of publication bias in meta-analyses.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

OBJECTIVE: Bench evaluation of the hydrodynamic behavior of venous cannulas is a valuable technique for the analysis of their performance during cardiopulmonary bypass (CPB). The aim of this study was to investigate the effect of the internal diameter of the extracorporeal connecting tube of venous cannulas on flow rate (Q), pressure drop (delta P), and cannula resistance (delta P/Q²) values, using a computer assisted test bench.¦METHODS: An in vitro circuit was set up with silicone tubing between the test cannula encased in a movable reservoir, and a static reservoir. The delta P, defined as the difference between the drainage pressure and the preload pressure, was measured using high-fidelity Millar pressure transducers. Q was measured using an ultrasonic flowmeter. Data display and data recording were controlled using virtual instruments in a stepwise fashion.¦RESULTS: The 27 F smartcanula® with a 9 mm connecting tube diameter showed 17% less resistance compared to that with an 8 mm connecting tube diameter. Q values were 7.22±0.1 and 7.81±0.04 L/min for cannulas with 8 mm and 9 mm connecting tube diameters, respectively. The delta P/Q² ratio values were 72% lower for the Medtronic cannula with a 9 mm connecting tube diameter compared to that with an 8 mm connecting tube diameter. Q values for the Medtronic cannula were 3.94±0.23 and 6.58±0.04 L/min with 8 mm and 9 mm connecting tube diameters, respectively. The 27 F smartcanula® showed 13% more flow rate compared to the 28 F Medtronic cannula using the unpaired Student t-test (p<0.0001).¦CONCLUSIONS: Our results demonstrated that Q was increased but delta P and delta P/Q² values were significantly decreased when the connecting tube diameter was increased for venous cannulas. The connecting tube diameter significantly affected the resistance to liquid flow through the cannula. Smartcanulas® outperform Medtronic cannulas.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

BACKGROUND: Migration is one of the major causes of tuberculosis in developed countries. Undocumented patients are usually not screened at the border and are not covered by a health insurance increasing their risk of developing the disease unnoticed. Urban health centres could help identify this population at risk. The objective of this study is to assess the prevalence of latent tuberculosis infection (LTBI) and adherence to preventive treatment in a population of undocumented immigrant patients. METHODS: All consecutive undocumented patients that visited two urban healthcare centres for vulnerable populations in Lausanne, Switzerland for the first time were offered tuberculosis screening with an interferon-gamma assay. Preventive treatment was offered if indicated. Adherence to treatment was evaluated monthly over a nine month period. RESULTS: Of the 161 participants, 131 (81.4%) agreed to screening and 125 had complete examinations. Twenty-four of the 125 patients (19.2%; CI95% 12.7;27.2) had positive interferon-gamma assay results, two of which had active tuberculosis. Only five patients with LTBI completed full preventive treatments. Five others initiated the treatment but did not follow through. CONCLUSION: Screening for tuberculosis infection in this hard-to-reach population is feasible in dedicated urban clinics, and the prevalence of LTBI is high in this vulnerable population. However, the low adherence to treatment is an important public health concern, and new strategies are needed to address this problem.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Työn tarkoituksena oli suunnitella kunnonvalvontajärjestelmä kahdelle lasivillan tuotantolinjalle. Suunnitteluprosessin lisäksi työssä on esitelty erilaisia kunnonvalvontamenetelmiä. Työn alussa on kerrottu erilaisista kunnonvalvontamenetelmistä, joilla voidaan seurata erilaisten laitteiden ja koneiden toimintakuntoa.Erityisesti työssä on tarkasteltu teollisuudessa yleistyviä kunnonvalvonnan värähtelymittauksia. Työssä suunniteltu kunnonvalvontajärjestelmä perustuu viiteen eri menetelmään, jotka ovat värähtelymittaus, lämpötilanmittaus lämpökameralla, lämpötilanmittaus kannettavalla mittarilla, kuuntelu elektronisella stetoskoopilla ja pyörivien osien kunnontarkkailu stroboskoopilla. Kunnonvalvontajärjestelmän suunnittelu on tehty useassa eri vaiheessa. Ensin työssä on kartoitettu tuotannon kannalta tärkeimmät laitteet ja niiden mahdolliset vikaantumistavat. Seuraavaksi on valittu sopivat kunnonvalvontamenetelmät ja tehty mittaussuunnitelma, jossa on esitetty eri laitteille suoritettavat mittaukset ja mittausten aikavälit.Lopuksi työssä on esitelty muutama esimerkkitapaus kunnonvalvontamenetelmien käytöstä sekä kerrottu mahdollisista tulevaisuuden kehitysmahdollisuuksista.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Statistics has become an indispensable tool in biomedical research. Thanks, in particular, to computer science, the researcher has easy access to elementary "classical" procedures. These are often of a "confirmatory" nature: their aim is to test hypotheses (for example the efficacy of a treatment) prior to experimentation. However, doctors often use them in situations more complex than foreseen, to discover interesting data structures and formulate hypotheses. This inverse process may lead to misuse which increases the number of "statistically proven" results in medical publications. The help of a professional statistician thus becomes necessary. Moreover, good, simple "exploratory" techniques are now available. In addition, medical data contain quite a high percentage of outliers (data that deviate from the majority). With classical methods it is often very difficult (even for a statistician!) to detect them and the reliability of results becomes questionable. New, reliable ("robust") procedures have been the subject of research for the past two decades. Their practical introduction is one of the activities of the Statistics and Data Processing Department of the University of Social and Preventive Medicine, Lausanne.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Objective To develop a Postnatal Perceived Stress Inventory (PNPSI) and assess its psychometric properties. Design Cross-sectional quantitative study. Setting One nurse-managed labor and delivery unit in a university hospital in a major metropolitan area. Participants One hundred seventy-nine (179) primiparous French speaking women who gave birth at term. Methods The PNPSI was validated at 6 weeks postpartum. Its predictive validity for depression and anxiety was assessed at the same time. Results The exploratory analysis revealed a 19-item structure divided into six factors. This inventory has good internal consistency (Cronbach's alpha = .815). The predictive validity shows that the PNPSI significantly predicts depression and anxiety at 6 weeks postpartum, and that certain factors are particularly prominent. Conclusion The PNPSI's psychometric properties make it a useful tool for future research to evaluate interventions for perceived stress during the postnatal period. Its predictive power for depression indicates that it is also a promising tool for clinical settings.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This paper analyzes the possibilities of integrating cost information and engineering design. Special emphasis is put on finding the potential of using the activity-based costing (ABC) method. Today, the problem of cost estimation in engineering design is that there are two separate extremes of knowledge. On the one extreme, the engineers model the technical parametres behindcosts in great detail but do not get appropriate cost information to their elegant models. On the other extreme, the accounting professionals are stuck with traditional cost accounting methods driven by the procedures and cycles of financial accounting. Therefore, in many cases, the cost information needs of various decision making groups, for example design engineers, are not served satisfactorily. This paper studies if the activity-based costing (ABC) method could offer a compromise between the two extremes. Recognizing activities and activity chains as well as activity and cost drivers could be specially beneficial for design engineers. Also, recognizing the accurate and reliable product costs of existing products helps when doing variant design. However, ABC is not at its best if the cost system becomes too complicated. This is why a comprehensive ABC-cost information system with detailed cost information for the use of design engineers should be examined critically. ABC is at its best when considering such issues as which activities drive costs, the cost of product complexity, allocating indirect costs on the products, the relationships between processes and costs, and the cost of excess capacity.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This paper analyzes the possibilities of integrating cost information and engineering design. Special emphasis is on finding the potential of using the activity-based costing (ABC) method when formulating cost information for the needs of design engineers. This paper suggests that ABC is more useful than the traditional job order costing, but the negative issue is the fact that ABC models become easily too complicated, i.e. expensive to build and maintain, and difficult to use. For engineering design the most suitable elements of ABC are recognizing activities of the company, constructing acitivity chains, identifying resources, activity and cost drivers, as wellas calculating accurate product costs. ABC systems including numerous cost drivers can become complex. Therefore, a comprehensive ABC based cost information system for the use of design engineers should be considered criticaly. Combining the suitable ideas of ABC with engineering oriented thinking could give competentresults.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The solid-rotor induction motor provides a mechanically and thermally reliable solution for demanding environments where other rotor solutions are prohibited or questionable. Solid rotors, which are manufactured of single pieces of ferromagnetic material, are commonly used in motors in which the rotationspeeds exceed substantially the conventional speeds of laminated rotors with squirrel-cage. During the operation of a solid-rotor electrical machine, the rotor core forms a conductor for both the magnetic flux and the electrical current. This causes an increase in the rotor resistance and rotor leakage inductance, which essentially decreases the power factor and the efficiency of the machine. The electromagnetic problems related to the solid-rotor induction motor are mostly associated with the low performance of the rotor. Therefore, the main emphasis in this thesis is put on the solid steel rotor designs. The rotor designs studied in thisthesis are based on the fact that the rotor construction should be extremely robust and reliable to withstand the high mechanical stresses caused by the rotational velocity of the rotor. In addition, the demanding operation environment sets requirements for the applied materials because of the high temperatures and oxidizing acids, which may be present in the cooling fluid. Therefore, the solid rotors analyzed in this thesis are made of a single piece of ferromagnetic material without any additional parts, such as copper end-rings or a squirrel-cage. A pure solid rotor construction is rigid and able to keep its balance over a large speed range. It also may tolerate other environmental stresses such as corroding substances or abrasive particles. In this thesis, the main target is to improve the performance of an induction motor equipped with a solid steel rotor by traditional methods: by axial slitting of the rotor, by selecting a proper rotor core material and by coating the rotor with a high-resistive stainless ferromagnetic material. In the solid steel rotor calculation, the rotor end-effects have a significant effect on the rotor characteristics. Thus, the emphasis is also put on the comparison of different rotor endfactors. In addition, a corrective slip-dependent end-factor is proposed. The rotor designs covered in this thesis are the smooth solid rotor, the axially slitted solid rotor and the slitted rotor having a uniform ferromagnetic coating cylinder. The thesis aims at design rules for multi-megawatt machines. Typically, mega-watt-size solidrotor machines find their applications mainly in the field of electric-motor-gas-compression systems, in steam-turbine applications, and in various types of largepower pump applications, where high operational speeds are required. In this thesis, a 120 kW, 10 000 rpm solid-rotor induction motor is usedas a small-scale model for such megawatt-range solid-rotor machines. The performance of the 120 kW solid-rotor induction motors is determined by experimental measurements and finite element calculations.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Tehoelektoniikkalaitteella tarkoitetaan ohjaus- ja säätöjärjestelmää, jolla sähköä muokataan saatavilla olevasta muodosta haluttuun uuteen muotoon ja samalla hallitaan sähköisen tehon virtausta lähteestä käyttökohteeseen. Tämä siis eroaa signaalielektroniikasta, jossa sähköllä tyypillisesti siirretään tietoa hyödyntäen eri tiloja. Tehoelektroniikkalaitteita vertailtaessa katsotaan yleensä niiden luotettavuutta, kokoa, tehokkuutta, säätötarkkuutta ja tietysti hintaa. Tyypillisiä tehoelektroniikkalaitteita ovat taajuudenmuuttajat, UPS (Uninterruptible Power Supply) -laitteet, hitsauskoneet, induktiokuumentimet sekä erilaiset teholähteet. Perinteisesti näiden laitteiden ohjaus toteutetaan käyttäen mikroprosessoreja, ASIC- (Application Specific Integrated Circuit) tai IC (Intergrated Circuit) -piirejä sekä analogisia säätimiä. Tässä tutkimuksessa on analysoitu FPGA (Field Programmable Gate Array) -piirien soveltuvuutta tehoelektroniikan ohjaukseen. FPGA-piirien rakenne muodostuu erilaisista loogisista elementeistä ja niiden välisistä yhdysjohdoista.Loogiset elementit ovat porttipiirejä ja kiikkuja. Yhdysjohdot ja loogiset elementit ovat piirissä kiinteitä eikä koostumusta tai lukumäärää voi jälkikäteen muuttaa. Ohjelmoitavuus syntyy elementtien välisistä liitännöistä. Piirissä on lukuisia, jopa miljoonia kytkimiä, joiden asento voidaan asettaa. Siten piirin peruselementeistä voidaan muodostaa lukematon määrä erilaisia toiminnallisia kokonaisuuksia. FPGA-piirejä on pitkään käytetty kommunikointialan tuotteissa ja siksi niiden kehitys on viime vuosina ollut nopeaa. Samalla hinnat ovat pudonneet. Tästä johtuen FPGA-piiristä on tullut kiinnostava vaihtoehto myös tehoelektroniikkalaitteiden ohjaukseen. Väitöstyössä FPGA-piirien käytön soveltuvuutta on tutkittu käyttäen kahta vaativaa ja erilaista käytännön tehoelektroniikkalaitetta: taajuudenmuuttajaa ja hitsauskonetta. Molempiin testikohteisiin rakennettiin alan suomalaisten teollisuusyritysten kanssa soveltuvat prototyypit,joiden ohjauselektroniikka muutettiin FPGA-pohjaiseksi. Lisäksi kehitettiin tätä uutta tekniikkaa hyödyntävät uudentyyppiset ohjausmenetelmät. Prototyyppien toimivuutta verrattiin vastaaviin perinteisillä menetelmillä ohjattuihin kaupallisiin tuotteisiin ja havaittiin FPGA-piirien mahdollistaman rinnakkaisen laskennantuomat edut molempien tehoelektroniikkalaitteiden toimivuudessa. Työssä on myösesitetty uusia menetelmiä ja työkaluja FPGA-pohjaisen säätöjärjestelmän kehitykseen ja testaukseen. Esitetyillä menetelmillä tuotteiden kehitys saadaan mahdollisimman nopeaksi ja tehokkaaksi. Lisäksi työssä on kehitetty FPGA:n sisäinen ohjaus- ja kommunikointiväylärakenne, joka palvelee tehoelektroniikkalaitteiden ohjaussovelluksia. Uusi kommunikointirakenne edistää lisäksi jo tehtyjen osajärjestelmien uudelleen käytettävyyttä tulevissa sovelluksissa ja tuotesukupolvissa.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The purpose of this thesis is to analyse activity-based costing (ABC) and possible modified versions ofit in engineering design context. The design engineers need cost information attheir decision-making level and the cost information should also have a strong future orientation. These demands are high because traditional management accounting has concentrated on the direct actual costs of the products. However, cost accounting has progressed as ABC was introduced late 1980s and adopted widely bycompanies in the 1990s. The ABC has been a success, but it has gained also criticism. In some cases the ambitious ABC systems have become too complex to build,use and update. This study can be called an action-oriented case study with some normative features. In this thesis theoretical concepts are assessed and allowed to unfold gradually through interaction with data from three cases. The theoretical starting points are ABC and theory of engineering design process (chapter2). Concepts and research results from these theoretical approaches are summarized in two hypotheses (chapter 2.3). The hypotheses are analysed with two cases (chapter 3). After the two case analyses, the ABC part is extended to cover alsoother modern cost accounting methods, e.g. process costing and feature costing (chapter 4.1). The ideas from this second theoretical part are operationalized with the third case (chapter 4.2). The knowledge from the theory and three cases is summarized in the created framework (chapter 4.3). With the created frameworkit is possible to analyse ABC and its modifications in the engineering design context. The framework collects the factors that guide the choice of the costing method to be used in engineering design. It also illuminates the contents of various ABC-related costing methods. However, the framework needs to be further tested. On the basis of the three cases it can be said that ABC should be used cautiously when formulating cost information for engineering design. It is suitable when the manufacturing can be considered simple, or when the design engineers are not cost conscious, and in the beginning of the design process when doing adaptive or variant design. If the design engineers need cost information for the embodiment or detailed design, or if manufacturing can be considered complex, or when design engineers are cost conscious, the ABC has to be always evaluated critically.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Superheater corrosion causes vast annual losses for the power companies. With a reliable corrosion prediction method, the plants can be designed accordingly, and knowledge of fuel selection and determination of process conditions may be utilized to minimize superheater corrosion. Growing interest to use recycled fuels creates additional demands for the prediction of corrosion potential. Models depending on corrosion theories will fail, if relations between the inputs and the output are poorly known. A prediction model based on fuzzy logic and an artificial neural network is able to improve its performance as the amount of data increases. The corrosion rate of a superheater material can most reliably be detected with a test done in a test combustor or in a commercial boiler. The steel samples can be located in a special, temperature-controlled probe, and exposed to the corrosive environment for a desired time. These tests give information about the average corrosion potential in that environment. Samples may also be cut from superheaters during shutdowns. The analysis ofsamples taken from probes or superheaters after exposure to corrosive environment is a demanding task: if the corrosive contaminants can be reliably analyzed, the corrosion chemistry can be determined, and an estimate of the material lifetime can be given. In cases where the reason for corrosion is not clear, the determination of the corrosion chemistry and the lifetime estimation is more demanding. In order to provide a laboratory tool for the analysis and prediction, a newapproach was chosen. During this study, the following tools were generated: · Amodel for the prediction of superheater fireside corrosion, based on fuzzy logic and an artificial neural network, build upon a corrosion database developed offuel and bed material analyses, and measured corrosion data. The developed model predicts superheater corrosion with high accuracy at the early stages of a project. · An adaptive corrosion analysis tool based on image analysis, constructedas an expert system. This system utilizes implementation of user-defined algorithms, which allows the development of an artificially intelligent system for thetask. According to the results of the analyses, several new rules were developed for the determination of the degree and type of corrosion. By combining these two tools, a user-friendly expert system for the prediction and analyses of superheater fireside corrosion was developed. This tool may also be used for the minimization of corrosion risks by the design of fluidized bed boilers.