954 resultados para Process capability analysis


Relevância:

30.00% 30.00%

Publicador:

Resumo:

The aim of this work is to study the influence of several analytical parameters on the variability of Raman spectra of paint samples. In the present study, microtome thin section and direct (no preparation) analysis are considered as sample preparation. In order to evaluate their influence on the measures, an experimental design such as 'fractional full factorial' with seven factors (including the sampling process) is applied, for a total of 32 experiments representing 160 measures. Once the influence of sample preparation highlighted, a depth profile of a paint sample is carried out by changing the focusing plane in order to measure the colored layer under a clearcoat. This is undertaken in order to avoid sample preparation such a microtome sectioning. Finally, chemometric treatments such as principal component analysis are applied to the resulting spectra. The findings of this study indicate the importance of sample preparation, or more specifically, the surface roughness, on the variability of the measurements on a same sample. Moreover, the depth profile experiment highlights the influence of the refractive index of the upper layer (clearcoat) when measuring through a transparent layer.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Background: Ethical conflicts are arising as a result of the growing complexity of clinical care, coupled with technological advances. Most studies that have developed instruments for measuring ethical conflict base their measures on the variables"frequency" and"degree of conflict". In our view, however, these variables are insufficient for explaining the root of ethical conflicts. Consequently, the present study formulates a conceptual model that also includes the variable"exposure to conflict", as well as considering six"types of ethical conflict". An instrument was then designed to measure the ethical conflicts experienced by nurses who work with critical care patients. The paper describes the development process and validation of this instrument, the Ethical Conflict in Nursing Questionnaire Critical Care Version (ECNQ-CCV). Methods: The sample comprised 205 nursing professionals from the critical care units of two hospitals in Barcelona (Spain). The ECNQ-CCV presents 19 nursing scenarios with the potential to produce ethical conflict in the critical care setting. Exposure to ethical conflict was assessed by means of the Index of Exposure to Ethical Conflict (IEEC), a specific index developed to provide a reference value for each respondent by combining the intensity and frequency of occurrence of each scenario featured in the ECNQ-CCV. Following content validity, construct validity was assessed by means of Exploratory Factor Analysis (EFA), while Cronbach"s alpha was used to evaluate the instrument"s reliability. All analyses were performed using the statistical software PASW v19. Results: Cronbach"s alpha for the ECNQ-CCV as a whole was 0.882, which is higher than the values reported for certain other related instruments. The EFA suggested a unidimensional structure, with one component accounting for 33.41% of the explained variance. Conclusions: The ECNQ-CCV is shown to a valid and reliable instrument for use in critical care units. Its structure is such that the four variables on which our model of ethical conflict is based may be studied separately or in combination. The critical care nurses in this sample present moderate levels of exposure to ethical conflict. This study represents the first evaluation of the ECNQ-CCV.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Background: Ethical conflicts are arising as a result of the growing complexity of clinical care, coupled with technological advances. Most studies that have developed instruments for measuring ethical conflict base their measures on the variables"frequency" and"degree of conflict". In our view, however, these variables are insufficient for explaining the root of ethical conflicts. Consequently, the present study formulates a conceptual model that also includes the variable"exposure to conflict", as well as considering six"types of ethical conflict". An instrument was then designed to measure the ethical conflicts experienced by nurses who work with critical care patients. The paper describes the development process and validation of this instrument, the Ethical Conflict in Nursing Questionnaire Critical Care Version (ECNQ-CCV). Methods: The sample comprised 205 nursing professionals from the critical care units of two hospitals in Barcelona (Spain). The ECNQ-CCV presents 19 nursing scenarios with the potential to produce ethical conflict in the critical care setting. Exposure to ethical conflict was assessed by means of the Index of Exposure to Ethical Conflict (IEEC), a specific index developed to provide a reference value for each respondent by combining the intensity and frequency of occurrence of each scenario featured in the ECNQ-CCV. Following content validity, construct validity was assessed by means of Exploratory Factor Analysis (EFA), while Cronbach"s alpha was used to evaluate the instrument"s reliability. All analyses were performed using the statistical software PASW v19. Results: Cronbach"s alpha for the ECNQ-CCV as a whole was 0.882, which is higher than the values reported for certain other related instruments. The EFA suggested a unidimensional structure, with one component accounting for 33.41% of the explained variance. Conclusions: The ECNQ-CCV is shown to a valid and reliable instrument for use in critical care units. Its structure is such that the four variables on which our model of ethical conflict is based may be studied separately or in combination. The critical care nurses in this sample present moderate levels of exposure to ethical conflict. This study represents the first evaluation of the ECNQ-CCV.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The final year project came to us as an opportunity to get involved in a topic which has appeared to be attractive during the learning process of majoring in economics: statistics and its application to the analysis of economic data, i.e. econometrics.Moreover, the combination of econometrics and computer science is a very hot topic nowadays, given the Information Technologies boom in the last decades and the consequent exponential increase in the amount of data collected and stored day by day. Data analysts able to deal with Big Data and to find useful results from it are verydemanded in these days and, according to our understanding, the work they do, although sometimes controversial in terms of ethics, is a clear source of value added both for private corporations and the public sector. For these reasons, the essence of this project is the study of a statistical instrument valid for the analysis of large datasets which is directly related to computer science: Partial Correlation Networks.The structure of the project has been determined by our objectives through the development of it. At first, the characteristics of the studied instrument are explained, from the basic ideas up to the features of the model behind it, with the final goal of presenting SPACE model as a tool for estimating interconnections in between elements in large data sets. Afterwards, an illustrated simulation is performed in order to show the power and efficiency of the model presented. And at last, the model is put into practice by analyzing a relatively large data set of real world data, with the objective of assessing whether the proposed statistical instrument is valid and useful when applied to a real multivariate time series. In short, our main goals are to present the model and evaluate if Partial Correlation Network Analysis is an effective, useful instrument and allows finding valuable results from Big Data.As a result, the findings all along this project suggest the Partial Correlation Estimation by Joint Sparse Regression Models approach presented by Peng et al. (2009) to work well under the assumption of sparsity of data. Moreover, partial correlation networks are shown to be a very valid tool to represent cross-sectional interconnections in between elements in large data sets.The scope of this project is however limited, as there are some sections in which deeper analysis would have been appropriate. Considering intertemporal connections in between elements, the choice of the tuning parameter lambda, or a deeper analysis of the results in the real data application are examples of aspects in which this project could be completed.To sum up, the analyzed statistical tool has been proved to be a very useful instrument to find relationships that connect the elements present in a large data set. And after all, partial correlation networks allow the owner of this set to observe and analyze the existing linkages that could have been omitted otherwise.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The Mechanistic-Empirical Pavement Design Guide (MEPDG) was developed under National Cooperative Highway Research Program (NCHRP) Project 1-37A as a novel mechanistic-empirical procedure for the analysis and design of pavements. The MEPDG was subsequently supported by AASHTO’s DARWin-ME and most recently marketed as AASHTOWare Pavement ME Design software as of February 2013. Although the core design process and computational engine have remained the same over the years, some enhancements to the pavement performance prediction models have been implemented along with other documented changes as the MEPDG transitioned to AASHTOWare Pavement ME Design software. Preliminary studies were carried out to determine possible differences between AASHTOWare Pavement ME Design, MEPDG (version 1.1), and DARWin-ME (version 1.1) performance predictions for new jointed plain concrete pavement (JPCP), new hot mix asphalt (HMA), and HMA over JPCP systems. Differences were indeed observed between the pavement performance predictions produced by these different software versions. Further investigation was needed to verify these differences and to evaluate whether identified local calibration factors from the latest MEPDG (version 1.1) were acceptable for use with the latest version (version 2.1.24) of AASHTOWare Pavement ME Design at the time this research was conducted. Therefore, the primary objective of this research was to examine AASHTOWare Pavement ME Design performance predictions using previously identified MEPDG calibration factors (through InTrans Project 11-401) and, if needed, refine the local calibration coefficients of AASHTOWare Pavement ME Design pavement performance predictions for Iowa pavement systems using linear and nonlinear optimization procedures. A total of 130 representative sections across Iowa consisting of JPCP, new HMA, and HMA over JPCP sections were used. The local calibration results of AASHTOWare Pavement ME Design are presented and compared with national and locally calibrated MEPDG models.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

BACKGROUND: Recent data suggest that beta-blockers can be beneficial in subgroups of patients with chronic heart failure (CHF). For metoprolol and carvedilol, an increase in ejection fraction has been shown and favorable effects on the myocardial remodeling process have been reported in some studies. We examined the effects of bisoprolol fumarate on exercise capacity and left ventricular volume with magnetic resonance imaging (MRI) and applied a novel high-resolution MRI tagging technique to determine myocardial rotation and relaxation velocity. METHODS: Twenty-eight patients (mean age, 57 +/- 11 years; mean ejection fraction, 26 +/- 6%) were randomized to bisoprolol fumarate (n = 13) or to placebo therapy (n = 15). The dosage of the drugs was titrated to match that of the the Cardiac Insufficiency Bisoprolol Study protocol. Hemodynamic and gas exchange responses to exercise, MRI measurements of left ventricular end-systolic and end-diastolic volumes and ejection fraction, and left ventricular rotation and relaxation velocities were measured before the administration of the drug and 6 and 12 months later. RESULTS: After 1 year, heart rate was reduced in the bisoprolol fumarate group both at rest (81 +/- 12 before therapy versus 61 +/- 11 after therapy; P <.01) and peak exercise (144 +/- 20 before therapy versus 127 +/- 17 after therapy; P <.01), which indicated a reduction in sympathetic drive. No differences were observed in heart rate responses in the placebo group. No differences were observed within or between groups in peak oxygen uptake, although work rate achieved was higher (117.9 +/- 36 watts versus 146.1 +/- 33 watts; P <.05) and exercise time tended to be higher (9.1 +/- 1.7 minutes versus 11.4 +/- 2.8 minutes; P =.06) in the bisoprolol fumarate group. A trend for a reduction in left ventricular end-diastolic volume (-54 mL) and left ventricular end-systolic volume (-62 mL) in the bisoprolol fumarate group occurred after 1 year. Ejection fraction was higher in the bisoprolol fumarate group (25.0 +/- 7 versus 36.2 +/- 9%; P <.05), and the placebo group remained unchanged. Most changes in volume and ejection fraction occurred during the latter 6 months of treatment. With myocardial tagging, insignificant reductions in left ventricular rotation velocity were observed in both groups, whereas relaxation velocity was reduced only after bisoprolol fumarate therapy (by 39%; P <.05). CONCLUSION: One year of bisoprolol fumarate therapy resulted in an improvement in exercise capacity, showed trends for reductions in end-diastolic and end-systolic volumes, increased ejection fraction, and significantly reduced relaxation velocity. Although these results generally confirm the beneficial effects of beta-blockade in patients with chronic heart failure, they show differential effects on systolic and diastolic function.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

BACKGROUND: Guidelines surrounding maternal contact with the stillborn infant have been contradictory over the past thirty years. Most studies have reported that seeing and holding the stillborn baby is associated with fewer anxiety and depressive symptoms among mothers of stillborn babies than not doing so. In contrast, others studies suggest that contact with the stillborn infant can lead to poorer maternal mental health outcomes. There is a lack of research focusing on the maternal experience of this contact. The present study aimed to investigate how mothers describe their experience of spending time with their stillborn baby and how they felt retrospectively about the decision they made to see and hold their baby or not. METHOD: In depth interviews were conducted with twenty-one mothers three months after stillbirth. All mothers had decided to see and the majority to hold their baby. Qualitative analysis of the interview data was performed using Interpretive Phenomenological Analysis. RESULTS: Six superordinate themes were identified: Characteristics of Contact, Physicality; Emotional Experience; Surreal Experience; Finality; and Decision. Having contact with their stillborn infant provided mothers with time to process what had happened, to build memories, and to 'say goodbye', often sharing the experience with partners and other family members. The majority of mothers felt satisfied with their decision to spend time with their stillborn baby. Several mothers talked about their fear of seeing a damaged or dead body. Some mothers experienced strong disbelief and dissociation during the contact. CONCLUSIONS: Results indicate that preparation before contact with the baby, professional support during the contact, and professional follow-up are crucial in order to prevent the development of maternal mental health problems. Fears of seeing a damaged or dead body should be sensitively explored and ways of coping discussed. Even in cases where mothers experienced intense distress during the contact with their stillborn baby, they still described that having had this contact was important and that they had taken the right decision. This indicates a need for giving parents an informed choice by engaging in discussions about the possible benefits and risks of seeing their stillborn baby.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The integrated system of design for manufacturing and assembly (DFMA) and internet based collaborative design are presented to support product design, manufacturing process, and assembly planning for axial eccentric oil-pump design. The presented system manages and schedules group oriented collaborative activities. The design guidelines of internet based collaborative design & DFMA are expressed. The components and the manufacturing stages of axial eccentric oil-pump are expressed in detail. The file formats of the presented system include the data types of collaborative design of the product, assembly design, assembly planning and assembly system design. Product design and assembly planning can be operated synchronously and intelligently and they are integrated under the condition of internet based collaborative design and DFMA. The technologies of collaborative modelling, collaborative manufacturing, and internet based collaborative assembly for the specific pump construction are developed. A seven-security level is presented to ensure the security of the internet based collaborative design system.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Granular flow phenomena are frequently encountered in the design of process and industrial plants in the traditional fields of the chemical, nuclear and oil industries as well as in other activities such as food and materials handling. Multi-phase flow is one important branch of the granular flow. Granular materials have unusual kinds of behavior compared to normal materials, either solids or fluids. Although some of the characteristics are still not well-known yet, one thing is confirmed: the particle-particle interaction plays a key role in the dynamics of granular materials, especially for dense granular materials. At the beginning of this thesis, detailed illustration of developing two models for describing the interaction based on the results of finite-element simulation, dimension analysis and numerical simulation is presented. The first model is used to describing the normal collision of viscoelastic particles. Based on some existent models, more parameters are added to this model, which make the model predict the experimental results more accurately. The second model is used for oblique collision, which include the effects from tangential velocity, angular velocity and surface friction based on Coulomb's law. The theoretical predictions of this model are in agreement with those by finite-element simulation. I n the latter chapters of this thesis, the models are used to predict industrial granular flow and the agreement between the simulations and experiments also shows the validation of the new model. The first case presents the simulation of granular flow passing over a circular obstacle. The simulations successfully predict the existence of a parabolic steady layer and show how the characteristics of the particles, such as coefficients of restitution and surface friction affect the separation results. The second case is a spinning container filled with granular material. Employing the previous models, the simulation could also reproduce experimentally observed phenomena, such as a depression in the center of a high frequency rotation. The third application is about gas-solid mixed flow in a vertically vibrated device. Gas phase motion is added to coherence with the particle motion. The governing equations of the gas phase are solved by using the Large eddy simulation (LES) and particle motion is predicted by using the Lagrangian method. The simulation predicted some pattern formation reported by experiment.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Centrifugal compressors are widely used for example in process industry, oil and gas industry, in small gas turbines and turbochargers. In order to achieve lower consumption of energy and operation costs the efficiency of the compressor needs to be improve. In the present work different pinches and low solidity vaned diffusers were utilized in order to improve the efficiency of a medium size centrifugal compressor. In this study, pinch means the decrement of the diffuser flow passage height. First different geometries were analyzed using computational fluid dynamics. The flow solver Finflo was used to solve the flow field. Finflo is a Navier-Stokes solver. The solver is capable to solve compressible, incompressible, steady and unsteady flow fields. Chien's k-e turbulence model was used. One of the numerically investigated pinched diffuser and one low solidity vaned diffuser were studied experimentally. The overall performance of the compressor and the static pressure distribution before and after the diffuser were measured. The flow entering and leaving the diffuser was measured using a three-hole Cobra-probe and Kiel-probes. The pinch and the low solidity vaned diffuser increased the efficiency of the compressor. Highest isentropic efficiency increment obtained was 3\% of the design isentropic efficiency of the original geometry. It was noticed in the numerical results that the pinch made to the hub and the shroud wall was most beneficial to the operation of the compressor. Also the pinch made to the hub was better than the pinchmade to the shroud. The pinch did not affect the operation range of the compressor, but the low solidity vaned diffuser slightly decreased the operation range.The unsteady phenomena in the vaneless diffuser were studied experimentally andnumerically. The unsteady static pressure was measured at the diffuser inlet and outlet, and time-accurate numerical simulation was conducted. The unsteady static pressure showed that most of the pressure variations lay at the passing frequency of every second blade. The pressure variations did not vanish in the diffuser and were visible at the diffuser outlet. However, the amplitude of the pressure variations decreased in the diffuser. The time-accurate calculations showed quite a good agreement with the measured data. Agreement was very good at the design operation point, even though the computational grid was not dense enough inthe volute and in the exit cone. The time-accurate calculation over-predicted the amplitude of the pressure variations at high flow.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The mass media are assigned an important role in political campaigns on popular votes. This article asks how the press communicates political issues to citizens during referendum campaigns, and whether some minimal criteria for successful public deliberation are met. The press coverage of all 24 ballot votes on welfare state issues from 1995 to 2004 in Switzerland is examined, distinguishing seven criteria to judge how news coverage compares to idealized notions of the media's role in the democratic process: coverage intensity, time for public deliberation, balance in media coverage, source independence and inclusiveness, substantive coverage, and spatial homogeneity. The results of our quantitative analysis suggest that the press does fulfil these normative requirements to a reasonable extent and that fears about biased or deceitful media treatment of ballot issues are not well-founded. However, some potential for optimizing the coverage of referendum campaigns by the Swiss press does exist

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Tämä diplomityö käsittelee työkaluja, jotka on suunniteltu kustannusten ennakointiin ja hinnan asetantaan. Aluksi on käyty läpi perinteisen ja toimintoperusteisen kustannuslaskennan perusteita. Näiden menetelmien välisiä eroja on tarkasteltu ja toimintoperusteisen kustannuslaskennan paremmin sopivuus nykypäivän yrityksille on perusteltu. Toisena käsitellään hinnoittelu. Hinnan merkitys, hinnoittelumenetelmät ja päätös lopullisesta hinnasta on käyty läpi. Hinnoittelun jälkeen esitellään kustannusjärjestelmät ja kustannusten arviointi. Nämä asiat todistavat, että tarkat kustannusarviot ovat elintärkeitä yritykselle. Tuotteen kustannusarviointi, hinnan asetanta ja tarjoaminen ovat erittäin merkityksellisiä asioita ottaen huomioon koko projektin elinkaaren ja tulevat tuotot. Nykyään on yleistä käyttää työkaluja kustannusarvioinnissa ja joskus myös hinnoittelussa. Työkalujen luotettavuus on tiedettävä, ennenkuin työkalut otetaan käyttöön. Myös työkalujen käyttäjät täytyy perehdyttää hyvin. Muuten yritys todennäköisesti kohtaa odottamattomia ja epämiellyttäviä yllätyksiä.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Työn tavoitteena oli tutkia hyvän asiakasreferenssin ominaisuuksia suodatinvalmistaja Laroxin myynnin ja huollon sekä yrityksen asiakkaiden näkökulmasta. Larox voi käyttää saatua tietoa referenssien tehokkaampaan valintaan ja hyödyntämiseen. Kaksi internet-kyselyä toteutettiin, välineenä Webropol. Alustava kysely sunnattiin Laroxin myynnille ja huollolle. Kysely koostui viidestä kategoriasta asiakasreferenssejä, joiden tärkeyttä arvioitiin, sekä vapaista vastauksista. Tunnistettuja hyvän asiakasreferenssin ominaisuuksia ovat hyvä suhde referenssiasiakkaaseen, positiiviset jarehelliset suosittelut asiakkaalta, referenssilaitteen hyvä toimintakyky ja asiakas joka ymmärtää huollon tärkeyden. Pääkysely suunnattiin Laroxin asiakkaille. Tilastollisilla analyyseilla tutkittiin koetun riskin mallinmuuttujien välisiä yhteyksiä. Analyysit eivät paljastaneet merkittäviä riippuvuuksia asiakasreferenssin ominaisuuksien tärkeydessä eritaustaisten vastaajien tai tilannetekijöiden välillä, mutta asiakasreferenssin ominaisuuksien faktorit tukevat mallia. Referenssilaitteiden toimintakyky vaikuttaa tärkeimmältä ja huollon tärkeys on myös merkittävä.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Monimutkaisen tietokonejärjestelmän suorituskykyoptimointi edellyttää järjestelmän ajonaikaisen käyttäytymisen ymmärtämistä. Ohjelmiston koon ja monimutkaisuuden kasvun myötä suorituskykyoptimointi tulee yhä tärkeämmäksi osaksi tuotekehitysprosessia. Tehokkaampien prosessorien käytön myötä myös energiankulutus ja lämmöntuotto ovat nousseet yhä suuremmiksi ongelmiksi, erityisesti pienissä, kannettavissa laitteissa. Lämpö- ja energiaongelmien rajoittamiseksi on kehitetty suorituskyvyn skaalausmenetelmiä, jotka edelleen lisäävät järjestelmän kompleksisuutta ja suorituskykyoptimoinnin tarvetta. Tässä työssä kehitettiin visualisointi- ja analysointityökalu ajonaikaisen käyttäytymisen ymmärtämisen helpottamiseksi. Lisäksi kehitettiin suorituskyvyn mitta, joka mahdollistaa erilaisten skaalausmenetelmien vertailun ja arvioimisen suoritusympäristöstä riippumatta, perustuen joko suoritustallenteen tai teoreettiseen analyysiin. Työkalu esittää ajonaikaisesti kerätyn tallenteen helposti ymmärrettävällä tavalla. Se näyttää mm. prosessit, prosessorikuorman, skaalausmenetelmien toiminnan sekä energiankulutuksen kolmiulotteista grafiikkaa käyttäen. Työkalu tuottaa myös käyttäjän valitsemasta osasta suorituskuvaa numeerista tietoa, joka sisältää useita oleellisia suorituskykyarvoja ja tilastotietoa. Työkalun sovellettavuutta tarkasteltiin todellisesta laitteesta saatua suoritustallennetta sekä suorituskyvyn skaalauksen simulointia analysoimalla. Skaalausmekanismin parametrien vaikutus simuloidun laitteen suorituskykyyn analysoitiin.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Nowadays the used fuel variety in power boilers is widening and new boiler constructions and running models have to be developed. This research and development is done in small pilot plants where more faster analyse about the boiler mass and heat balance is needed to be able to find and do the right decisions already during the test run. The barrier on determining boiler balance during test runs is the long process of chemical analyses of collected input and outputmatter samples. The present work is concentrating on finding a way to determinethe boiler balance without chemical analyses and optimise the test rig to get the best possible accuracy for heat and mass balance of the boiler. The purpose of this work was to create an automatic boiler balance calculation method for 4 MW CFB/BFB pilot boiler of Kvaerner Pulping Oy located in Messukylä in Tampere. The calculation was created in the data management computer of pilot plants automation system. The calculation is made in Microsoft Excel environment, which gives a good base and functions for handling large databases and calculations without any delicate programming. The automation system in pilot plant was reconstructed und updated by Metso Automation Oy during year 2001 and the new system MetsoDNA has good data management properties, which is necessary for big calculations as boiler balance calculation. Two possible methods for calculating boiler balance during test run were found. Either the fuel flow is determined, which is usedto calculate the boiler's mass balance, or the unburned carbon loss is estimated and the mass balance of the boiler is calculated on the basis of boiler's heat balance. Both of the methods have their own weaknesses, so they were constructed parallel in the calculation and the decision of the used method was left to user. User also needs to define the used fuels and some solid mass flowsthat aren't measured automatically by the automation system. With sensitivity analysis was found that the most essential values for accurate boiler balance determination are flue gas oxygen content, the boiler's measured heat output and lower heating value of the fuel. The theoretical part of this work concentrates in the error management of these measurements and analyses and on measurement accuracy and boiler balance calculation in theory. The empirical part of this work concentrates on the creation of the balance calculation for the boiler in issue and on describing the work environment.