837 resultados para Entropy of a sampling design
Resumo:
BACKGROUND: Aromatase inhibitors provide superior disease control when compared with tamoxifen as adjuvant therapy for postmenopausal women with endocrine-responsive early breast cancer. PURPOSE: To present the design, history, and analytic challenges of the Breast International Group (BIG) 1-98 trial: an international, multicenter, randomized, double-blind, phase-III study comparing the aromatase inhibitor letrozole with tamoxifen in this clinical setting. METHODS: From 1998-2003, BIG 1-98 enrolled 8028 women to receive monotherapy with either tamoxifen or letrozole for 5 years, or sequential therapy of 2 years of one agent followed by 3 years of the other. Randomization to one of four treatment groups permitted two complementary analyses to be conducted several years apart. The first, reported in 2005, provided a head-to-head comparison of letrozole versus tamoxifen. Statistical power was increased by an enriched design, which included patients who were assigned sequential treatments until the time of the treatment switch. The second, reported in late 2008, used a conditional landmark approach to test the hypothesis that switching endocrine agents at approximately 2 years from randomization for patients who are disease-free is superior to continuing with the original agent. RESULTS: The 2005 analysis showed the superiority of letrozole compared with tamoxifen. The patients who were assigned tamoxifen alone were unblinded and offered the opportunity to switch to letrozole. Results from other trials increased the clinical relevance about whether or not to start treatment with letrozole or tamoxifen, and analysis plans were expanded to evaluate sequential versus single-agent strategies from randomization. LIMITATIONS: Due to the unblinding of patients assigned tamoxifen alone, analysis of updated data will require ascertainment of the influence of selective crossover from tamoxifen to letrozole. CONCLUSIONS: BIG 1-98 is an example of an enriched design, involving complementary analyses addressing different questions several years apart, and subject to evolving analytic plans influenced by new data that emerge over time.
Resumo:
OBJECTIVES: Etravirine (ETV) is metabolized by cytochrome P450 (CYP) 3A, 2C9, and 2C19. Metabolites are glucuronidated by uridine diphosphate glucuronosyltransferases (UGT). To identify the potential impact of genetic and non-genetic factors involved in ETV metabolism, we carried out a two-step pharmacogenetics-based population pharmacokinetic study in HIV-1 infected individuals. MATERIALS AND METHODS: The study population included 144 individuals contributing 289 ETV plasma concentrations and four individuals contributing 23 ETV plasma concentrations collected in a rich sampling design. Genetic variants [n=125 single-nucleotide polymorphisms (SNPs)] in 34 genes with a predicted role in ETV metabolism were selected. A first step population pharmacokinetic model included non-genetic and known genetic factors (seven SNPs in CYP2C, one SNP in CYP3A5) as covariates. Post-hoc individual ETV clearance (CL) was used in a second (discovery) step, in which the effect of the remaining 98 SNPs in CYP3A, P450 cytochrome oxidoreductase (POR), nuclear receptor genes, and UGTs was investigated. RESULTS: A one-compartment model with zero-order absorption best characterized ETV pharmacokinetics. The average ETV CL was 41 (l/h) (CV 51.1%), the volume of distribution was 1325 l, and the mean absorption time was 1.2 h. The administration of darunavir/ritonavir or tenofovir was the only non-genetic covariate influencing ETV CL significantly, resulting in a 40% [95% confidence interval (CI): 13-69%] and a 42% (95% CI: 17-68%) increase in ETV CL, respectively. Carriers of rs4244285 (CYP2C19*2) had 23% (8-38%) lower ETV CL. Co-administered antiretroviral agents and genetic factors explained 16% of the variance in ETV concentrations. None of the SNPs in the discovery step influenced ETV CL. CONCLUSION: ETV concentrations are highly variable, and co-administered antiretroviral agents and genetic factors explained only a modest part of the interindividual variability in ETV elimination. Opposing effects of interacting drugs effectively abrogate genetic influences on ETV CL, and vice-versa.
Resumo:
Iowa’s traffic safety culture is influenced by laws and policies, enforcement methods, driver education, roadway engineering, and drivers’ behaviors. The Center for Social and Behavioral Research at the University of Northern Iowa was contracted by the Iowa Department of Transportation to conduct a general population survey of adult Iowans. Telephone interviews were conducted with 1,088 adult Iowans from October to December 2011. A dual-frame (cell phone and landline) sampling design was used. The interview covered a wide range of traffic safety topics (e.g., traffic safety policies, enforcement techniques, and distracted driving). Most Iowans said driving in Iowa is about as safe now as it was 5 years ago; however, one-fourth said driving in Iowa is less safe now. There are a number of driving-related behaviors many adult Iowans consider serious threats to traffic safety and never acceptable to do while driving. Yet, many Iowans report often seeing other drivers engaging in these behaviors and admit engaging in some themselves. For example, nearly 1 in 5 adult Iowa drivers said they have sent or read a text message or email while driving in the past 30 days despite this being prohibited since July of 2011. A slight majority said they support using cameras on highways, interstates, and city streets to automatically ticket drivers for speeding, with even stronger support for red light cameras. A comprehensive approach to traffic safety in Iowa is required to encourage protective factors that enhance traffic safety and reduce the impact of detrimental factors.
Resumo:
Résumé La diminution de la biodiversité, à toutes les échelles spatiales et sur l'ensemble de la planète, compte parmi les problèmes les plus préoccupants de notre époque. En terme de conservation, il est aujourd'hui primordial de mieux comprendre les mécanismes qui créent et maintiennent la biodiversité dans les écosystèmes naturels ou anthropiques. La présente étude a pour principal objectif d'améliorer notre compréhension des patrons de biodiversité végétale et des mécanismes sous jacents, dans un écosystème complexe, riche en espèces et à forte valeur patrimoniale, les pâturages boisés jurassiens. Structure et échelle spatiales sont progressivement reconnues comme des dimensions incontournables dans l'étude des patrons de biodiversité. De plus, ces deux éléments jouent un rôle central dans plusieurs théories écologiques. Toutefois, peu d'hypothèses issues de simulations ou d'études théoriques concernant le lien entre structure spatiale du paysage et biodiversité ont été testées de façon empirique. De même, l'influence des différentes composantes de l'échelle spatiale sur les patrons de biodiversité est méconnue. Cette étude vise donc à tester quelques-unes de ces hypothèses et à explorer les patrons spatiaux de biodiversité dans un contexte multi-échelle, pour différentes mesures de biodiversité (richesse et composition en espèces) à l'aide de données de terrain. Ces données ont été collectées selon un plan d'échantillonnage hiérarchique. Dans un premier temps, nous avons testé l'hypothèse élémentaire selon laquelle la richesse spécifique (le nombre d'espèces sur une surface donnée) est liée à l'hétérogénéité environnementale quelque soit l'échelle. Nous avons décomposé l'hétérogénéité environnementale en deux parties, la variabilité des conditions environnementales et sa configuration spatiale. Nous avons montré que, en général, la richesse spécifique augmentait avec l'hétérogénéité de l'environnement : elle augmentait avec le nombre de types d'habitats et diminuait avec l'agrégation spatiale de ces habitats. Ces effets ont été observés à toutes les échelles mais leur nature variait en fonction de l'échelle, suggérant une modification des mécanismes. Dans un deuxième temps, la structure spatiale de la composition en espèces a été décomposée en relation avec 20 variables environnementales et 11 traits d'espèces. Nous avons utilisé la technique de partition de la variation et un descripteur spatial, récemment développé, donnant accès à une large gamme d'échelles spatiales. Nos résultats ont montré que la structure spatiale de la composition en espèces végétales était principalement liée à la topographie, aux échelles les plus grossières, et à la disponibilité en lumière, aux échelles les plus fines. La fraction non-environnementale de la variation spatiale de la composition spécifique avait une relation complexe avec plusieurs traits d'espèces suggérant un lien avec des processus biologiques tels que la dispersion, dépendant de l'échelle spatiale. Dans un dernier temps, nous avons testé, à plusieurs échelles spatiales, les relations entre trois composantes de la biodiversité : la richesse spécifique totale d'un échantillon (diversité gamma), la richesse spécifique moyenne (diversité alpha), mesurée sur des sous-échantillons, et les différences de composition spécifique entre les sous-échantillons (diversité beta). Les relations deux à deux entre les diversités alpha, beta et gamma ne suivaient pas les relations attendues, tout du moins à certaines échelles spatiales. Plusieurs de ces relations étaient fortement dépendantes de l'échelle. Nos résultats ont mis en évidence l'importance du rapport d'échelle (rapport entre la taille de l'échantillon et du sous-échantillon) lors de l'étude des patrons spatiaux de biodiversité. Ainsi, cette étude offre un nouvel aperçu des patrons spatiaux de biodiversité végétale et des mécanismes potentiels permettant la coexistence des espèces. Nos résultats suggèrent que les patrons de biodiversité ne peuvent être expliqués par une seule théorie, mais plutôt par une combinaison de théories. Ils ont également mis en évidence le rôle essentiel joué par la structure spatiale dans la détermination de la biodiversité, quelque soit le composant de la biodiversité considéré. Enfin, cette étude souligne l'importance de prendre en compte plusieurs échelles spatiales et différents constituants de l'échelle spatiale pour toute étude relative à la diversité spécifique. Abstract The world-wide loss of biodiversity at all scales has become a matter of urgent concern, and improving our understanding of local drivers of biodiversity in natural and anthropogenic ecosystems is now crucial for conservation. The main objective of this study was to further our comprehension of the driving forces controlling biodiversity patterns in a complex and diverse ecosystem of high conservation value, wooded pastures. Spatial pattern and scale are central to several ecological theories, and it is increasingly recognized that they must be taken -into consideration when studying biodiversity patterns. However, few hypotheses developed from simulations or theoretical studies have been tested using field data, and the evolution of biodiversity patterns with different scale components remains largely unknown. We test several such hypotheses and explore spatial patterns of biodiversity in a multi-scale context and using different measures of biodiversity (species richness and composition), with field data. Data were collected using a hierarchical sampling design. We first tested the simple hypothesis that species richness, the number of species in a given area, is related to environmental heterogeneity at all scales. We decomposed environmental heterogeneity into two parts: the variability of environmental conditions and its spatial configuration. We showed that species richness generally increased with environmental heterogeneity: species richness increased with increasing number of habitat types and with decreasing spatial aggregation of those habitats. Effects occurred at all scales but the nature of the effect changed with scale, suggesting a change in underlying mechanisms. We then decomposed the spatial structure of species composition in relation to environmental variables and species traits using variation partitioning and a recently developed spatial descriptor, allowing us to capture a wide range of spatial scales. We showed that the spatial structure of plant species composition was related to topography at the coarsest scales and insolation at finer scales. The non-environmental fraction of the spatial variation in species composition had a complex relationship with several species traits, suggesting a scale-dependent link to biological processes, particularly dispersal. Finally, we tested, at different spatial scales, the relationships between different components of biodiversity: total sample species richness (gamma diversity), mean species .richness (alpha diversity), measured in nested subsamples, and differences in species composition between subsamples (beta diversity). The pairwise relationships between alpha, beta and gamma diversity did not follow the expected patterns, at least at certain scales. Our result indicated a strong scale-dependency of several relationships, and highlighted the importance of the scale ratio when studying biodiversity patterns. Thus, our results bring new insights on the spatial patterns of biodiversity and the possible mechanisms allowing species coexistence. They suggest that biodiversity patterns cannot be explained by any single theory proposed in the literature, but a combination of theories is sufficient. Spatial structure plays a crucial role for all components of biodiversity. Results emphasize the importance of considering multiple spatial scales and multiple scale components when studying species diversity.
Resumo:
Helping behavior is any intentional behavior that benefits another living being or group (Hogg & Vaughan, 2010). People tend to underestimate the probability that others will comply with their direct requests for help (Flynn & Lake, 2008). This implies that when they need help, they will assess the probability of getting it (De Paulo, 1982, cited in Flynn & Lake, 2008) and then they will tend to estimate one that is actually lower than the real chance, so they may not even consider worth asking for it. Existing explanations for this phenomenon attribute it to a mistaken cost computation by the help seeker, who will emphasize the instrumental cost of “saying yes”, ignoring that the potential helper also needs to take into account the social cost of saying “no”. And the truth is that, especially in face-to-face interactions, the discomfort caused by refusing to help can be very high. In short, help seekers tend to fail to realize that it might be more costly to refuse to comply with a help request rather than accepting. A similar effect has been observed when estimating trustworthiness of people. Fetchenhauer and Dunning (2010) showed that people also tend to underestimate it. This bias is reduced when, instead of asymmetric feedback (getting feedback only when deciding to trust the other person), symmetric feedback (always given) was provided. This cause could as well be applicable to help seeking as people only receive feedback when they actually make their request but not otherwise. Fazio, Shook, and Eiser (2004) studied something that could be reinforcing these outcomes: Learning asymmetries. By means of a computer game called BeanFest, they showed that people learn better about negatively valenced objects (beans in this case) than about positively valenced ones. This learning asymmetry esteemed from “information gain being contingent on approach behavior” (p. 293), which could be identified with what Fetchenhauer and Dunning mention as ‘asymmetric feedback’, and hence also with help requests. Fazio et al. also found a generalization asymmetry in favor of negative attitudes versus positive ones. They attributed it to a negativity bias that “weights resemblance to a known negative more heavily than resemblance to a positive” (p. 300). Applied to help seeking scenarios, this would mean that when facing an unknown situation, people would tend to generalize and infer that is more likely that they get a negative rather than a positive outcome from it, so, along with what it was said before, people will be more inclined to think that they will get a “no” when requesting help. Denrell and Le Mens (2011) present a different perspective when trying to explain judgment biases in general. They deviate from the classical inappropriate information processing (depicted among other by Fiske & Taylor, 2007, and Tversky & Kahneman, 1974) and explain this in terms of ‘adaptive sampling’. Adaptive sampling is a sampling mechanism in which the selection of sample items is conditioned by the values of the variable of interest previously observed (Thompson, 2011). Sampling adaptively allows individuals to safeguard themselves from experiences they went through once and turned out to lay negative outcomes. However, it also prevents them from giving a second chance to those experiences to get an updated outcome that could maybe turn into a positive one, a more positive one, or just one that regresses to the mean, whatever direction that implies. That, as Denrell and Le Mens (2011) explained, makes sense: If you go to a restaurant, and you did not like the food, you do not choose that restaurant again. This is what we think could be happening when asking for help: When we get a “no”, we stop asking. And here, we want to provide a complementary explanation for the underestimation of the probability that others comply with our direct help requests based on adaptive sampling. First, we will develop and explain a model that represents the theory. Later on, we will test it empirically by means of experiments, and will elaborate on the analysis of its results.
Resumo:
A mechanical gauge was developed to monitor the movement of crack or joint openings in portland cement concrete structures, in general, and portland cement concrete pavements in particular. Designed to be inexpensive and simple to operate, this gauge is capable of recording maximum, minimum, and instantaneous crack or joint openings. Specific recommendations were made for recording minimum and maximum pavement temperature over the monitoring period. The report was written as a set of guidelines for design, fabrication, installation, and operation of the gauge as well as the temperature measuring device.
Resumo:
1. Species distribution models are increasingly used to address conservation questions, so their predictive capacity requires careful evaluation. Previous studies have shown how individual factors used in model construction can affect prediction. Although some factors probably have negligible effects compared to others, their relative effects are largely unknown. 2. We introduce a general "virtual ecologist" framework to study the relative importance of factors involved in the construction of species distribution models. 3. We illustrate the framework by examining the relative importance of five key factors-a missing covariate, spatial autocorrelation due to a dispersal process in presences/absences, sample size, sampling design and modeling technique-in a real study framework based on plants in a mountain landscape at regional scale, and show that, for the parameter values considered here, most of the variation in prediction accuracy is due to sample size and modeling technique. Contrary to repeatedly reported concerns, spatial autocorrelation has only comparatively small effects. 4. This study shows the importance of using a nested statistical framework to evaluate the relative effects of factors that may affect species distribution models.
Resumo:
The need for high performance, high precision, and energy saving in rotating machinery demands an alternative solution to traditional bearings. Because of the contactless operation principle, the rotating machines employing active magnetic bearings (AMBs) provide many advantages over the traditional ones. The advantages such as contamination-free operation, low maintenance costs, high rotational speeds, low parasitic losses, programmable stiffness and damping, and vibration insulation come at expense of high cost, and complex technical solution. All these properties make the use of AMBs appropriate primarily for specific and highly demanding applications. High performance and high precision control requires model-based control methods and accurate models of the flexible rotor. In turn, complex models lead to high-order controllers and feature considerable computational burden. Fortunately, in the last few years the advancements in signal processing devices provide new perspective on the real-time control of AMBs. The design and the real-time digital implementation of the high-order LQ controllers, which focus on fast execution times, are the subjects of this work. In particular, the control design and implementation in the field programmable gate array (FPGA) circuits are investigated. The optimal design is guided by the physical constraints of the system for selecting the optimal weighting matrices. The plant model is complemented by augmenting appropriate disturbance models. The compensation of the force-field nonlinearities is proposed for decreasing the uncertainty of the actuator. A disturbance-observer-based unbalance compensation for canceling the magnetic force vibrations or vibrations in the measured positions is presented. The theoretical studies are verified by the practical experiments utilizing a custom-built laboratory test rig. The test rig uses a prototyping control platform developed in the scope of this work. To sum up, the work makes a step in the direction of an embedded single-chip FPGA-based controller of AMBs.
Resumo:
A sign of presence in virtual environments is that people respond to situations and events as if they were real, where response may be considered at many different levels, ranging from unconscious physiological responses through to overt behavior,emotions, and thoughts. In this paper we consider two responses that gave different indications of the onset of presence in a gradually forming environment. Two aspects of the response of people to an immersive virtual environment were recorded: their eye scanpath, and their skin conductance response (SCR). The scenario was formed over a period of 2 min, by introducing an increasing number of its polygons in random order in a head-tracked head-mounted display. For one group of experimental participants (n 8) the environment formed into one in which they found themselves standing on top of a 3 m high column. For a second group of participants (n 6) the environment was otherwise the same except that the column was only 1 cm high, so that they would be standing at normal ground level. For a third group of participants (n 14) the polygons never formed into a meaningful environment. The participants who stood on top of the tall column exhibited a significant decrease in entropy of the eye scanpath and an increase in the number of SCR by 99 s into the scenario, at a time when only 65% of the polygons had been displayed. The ground level participants exhibited a similar decrease in scanpath entropy, but not the increase in SCR. The random scenario grouping did not exhibit this decrease in eye scanpath entropy. A drop in scanpath entropy indicates that the environment had cohered into a meaningful perception. An increase in the rate of SCR indicates the perception of an aversive stimulus. These results suggest that on these two dimensions (scanpath entropy and rate of SCR) participants were responding realistically to the scenario shown in the virtual environment. In addition, the response occurred well before the entire scenario had been displayed, suggesting that once a set of minimal cues exists within a scenario,it is enough to form a meaningful perception. Moreover, at the level of the sympathetic nervous system, the participants who were standing on top of the column exhibited arousal as if their experience might be real. This is an important practical aspect of the concept of presence.
Resumo:
Tässä työssä tutkitaan ohjelmistoarkkitehtuurisuunnitteluominaisuuksien vaikutusta erään client-server –arkkitehtuuriin perustuvan mobiilipalvelusovelluksen suunnittelu- ja toteutusaikaan. Kyseinen tutkimus perustuu reaalielämän projektiin, jonka kvalitatiivinen analyysi paljasti arkkitehtuurikompponenttien välisten kytkentöjen merkittävästi vaikuttavan projektin työmäärään. Työn päätavoite oli kvantitatiivisesti tutkia yllä mainitun havainnon oikeellisuus. Tavoitteen saavuttamiseksi suunniteltiin ohjelmistoarkkitehtuurisuunnittelun mittaristo kuvaamaan kyseisen järjestelmän alijärjestelmien arkkitehtuuria ja luotiin kaksi suunniteltua mittaristoa käyttävää, työmäärää (komponentin suunnittelu-, toteutus- ja testausaikojen summa) arvioivaa mallia, joista toinen on lineaarinen ja toinen epälineaarinen. Näiden mallien kertoimet sovitettiin optimoimalla niiden arvot epälineaarista gloobaalioptimointimenetelmää, differentiaalievoluutioalgoritmia, käyttäen, niin että mallien antamat arvot vastasivat parhaiten mitattua työmäärää sekä kaikilla ominaisuuksilla eli attribuuteilla että vain osalla niistä (yksi jätettiin vuorotellen pois). Kun arkkitehtuurikompenttien väliset kytkennät jätettiin malleista pois, mitattujen ja arvoitujen työmäärien välinen ero (ilmaistuna virheenä) kasvoi eräässä tapauksessa 367 % entisestä tarkoittaen sitä, että näin muodostettu malli vastasi toteutusaikoja huonosti annetulla ainestolla. Tämä oli suurin havaitu virhe kaikkien poisjätettyjen ominaisuuksien kesken. Saadun tuloksen perusteella päätettiin, että kyseisen järjestelmän toteutusajat ovat vahvasti riippuvaisia kytkentöjen määrästä, ja näin ollen kytkentöjen määrä oli mitä todennäköisemmin kaikista tärkein työmäärään vaikuttava tekijä tutkitun järjestelmän arkkitehtuurisuunnittelussa.
Resumo:
BACKGROUND & AIMS: Trace elements (TE) are involved in the immune and antioxidant defences which are of particular importance during critical illness. Determining plasma TE levels is costly. The present quality control study aimed at assessing the economic impact of a computer reminded blood sampling versus a risk guided on-demand monitoring of plasma concentrations of selenium, copper, and zinc. METHODS: Retrospective analysis of 2 cohorts of patients admitted during 6 months periods in 2006 and 2009 to the ICU of a University hospital. INCLUSION CRITERIA: to receive intravenous micronutrient supplements and/or to have a TE sampling during ICU stay. The TE samplings were triggered by computerized reminder in 2006 versus guided by nutritionists in 2009. RESULTS: During the 2 periods 636 patients met the inclusion criteria out of 2406 consecutive admissions, representing 29.7% and 24.9% respectively of the periods' admissions. The 2009 patients had higher SAPS2 scores (p = 0.02) and lower BMI compared to 2006 (p = 0.007). The number of laboratory determinations was drastically reduced in 2009, particularly during the first week, despite the higher severity of the cohort, resulting in à 55% cost reduction. CONCLUSIONS: The monitoring of TE concentrations guided by a nutritionist resulted in a reduction of the sampling frequency, and targeting on the sickest high risk patients, requiring a nutritional prescription adaptation. This control leads to cost reduction compared to an automated sampling prescription.
Resumo:
This study aimed at comparing the efficiency of various sampling materials for the collection and subsequent analysis of organic gunshot residues (OGSR). To the best of our knowledge, it is the first time that sampling devices were investigated in detail for further quantitation of OGSR by LC-MS. Seven sampling materials, namely two "swab"-type and five "stub"-type collection materials, were tested. The investigation started with the development of a simple and robust LC-MS method able to separate and quantify molecules typically found in gunpowders, such as diphenylamine or ethylcentralite. The evaluation of sampling materials was then systematically carried out by first analysing blank extracts of the materials to check for potential interferences and determining matrix effects. Based on these results, the best four materials, namely cotton buds, polyester swabs, a tape from 3M and PTFE were compared in terms of collection efficiency during shooting experiments using a set of 9 mm Luger ammunition. It was found that the tape was capable of recovering the highest amounts of OGSR. As tape-lifting is the technique currently used in routine for inorganic GSR, OGSR analysis might be implemented without modifying IGSR sampling and analysis procedure.
Resumo:
In this paper, manufacturability analysis and collection of design aspects is made for a microwave test-fixture. Aspects of applying systematic design for a microwave test-fixture design and manufacturing are also analysed. Special questionnaires for the component and machining are made in order to enable necessary information to ensure DFM(A) – aspects of the component. The aspects of easy manufacturing for machining the microwave test-fixture are collected. Material selection is discussed and manufacturing stages of prototype manufacturing are presented.
Resumo:
The uncertainty of any analytical determination depends on analysis and sampling. Uncertainty arising from sampling is usually not controlled and methods for its evaluation are still little known. Pierre Gy’s sampling theory is currently the most complete theory about samplingwhich also takes the design of the sampling equipment into account. Guides dealing with the practical issues of sampling also exist, published by international organizations such as EURACHEM, IUPAC (International Union of Pure and Applied Chemistry) and ISO (International Organization for Standardization). In this work Gy’s sampling theory was applied to several cases, including the analysis of chromite concentration estimated on SEM (Scanning Electron Microscope) images and estimation of the total uncertainty of a drug dissolution procedure. The results clearly show that Gy’s sampling theory can be utilized in both of the above-mentioned cases and that the uncertainties achieved are reliable. Variographic experiments introduced in Gy’s sampling theory are beneficially applied in analyzing the uncertainty of auto-correlated data sets such as industrial process data and environmental discharges. The periodic behaviour of these kinds of processes can be observed by variographic analysis as well as with fast Fourier transformation and auto-correlation functions. With variographic analysis, the uncertainties are estimated as a function of the sampling interval. This is advantageous when environmental data or process data are analyzed as it can be easily estimated how the sampling interval is affecting the overall uncertainty. If the sampling frequency is too high, unnecessary resources will be used. On the other hand, if a frequency is too low, the uncertainty of the determination may be unacceptably high. Variographic methods can also be utilized to estimate the uncertainty of spectral data produced by modern instruments. Since spectral data are multivariate, methods such as Principal Component Analysis (PCA) are needed when the data are analyzed. Optimization of a sampling plan increases the reliability of the analytical process which might at the end have beneficial effects on the economics of chemical analysis,
Resumo:
Aim: To investigate and understand patient's satisfaction with nursing care in the intensive care unit to identify the dimensions of the concept of"satisfaction" from the patient's point of view. To design and validate a questionnaire that measures satisfaction levels in critical patients. Background: There are many instruments capable of measuring satisfaction with nursing care; however, they do not address the reality for critical patients nor are they applicable in our context. Design: A dual approach study comprising: a qualitative phase employing Grounded Theory and a quantitative and descriptive phase to prepare and validate the questionnaire. Methods: Data collection in the qualitative phase will consist of: in-depth interview after theoretical sampling, on-site diary and expert discussion group. The sample size will depend on the expected theoretical saturation n = 27-36. Analysis will be based on Grounded Theory. For the quantitative phase, the sampling will be based on convenience (n = 200). A questionnaire will be designed on the basis of qualitative data. Descriptive and inferential statistics will be used. The validation will be developed on the basis of the validity of the content, the criteria of the construct and reliability of the instrument by the Cronbach's alpha and test-retest approach. Approval date for this protocol was November 2010. Discussion: Self-perceptions, beliefs, experiences, demographic, socio-cultural epistemological and political factors are determinants for satisfaction, and these should be taken into account when compiling a questionnaire on satisfaction with nursing care among critical patients.