983 resultados para estimation risk


Relevância:

30.00% 30.00%

Publicador:

Resumo:

Agricultural workers are exposed to various risks, including chemical agents, noise, and many other factors. One of the most characteristic and least known risk factors is constituted by the microclimatic conditions in the different phases of work (in field, in greenhouse, etc). A typical condition is thermal stress due to high temperatures during harvesting operations in open fields or in greenhouses. In Italy, harvesting is carried out for many hours during the day, mainly in the summer, with temperatures often higher than 30 degrees C. According to ISO 7243, these conditions can be considered dangerous for workers' health. The aim of this study is to assess the risks of exposure to microclimatic conditions (heat) for fruit and vegetable harvesters in central Italy by applying methods established by international standards. In order to estimate the risk for workers, the air temperature, radiative temperature, and air speed were measured using instruments in conformity with ISO 7726. Thermodynamic parameters and two more subjective parameters, clothing and the metabolic heat production rate related to the worker's physical activity, were used to calculate the predicted heat strain (PHS) for the exposed workers in conformity with ISO 7933. Environmental and subjective parameters were also measured for greenhouse workers, according to ISO 7243, in order to calculate the wet-bulb globe temperature (WBGT). The results show a slight risk for workers during manual harvesting in the field. On the other hand, the data collected in the greenhouses show that the risk for workers must not be underestimated. The results of the study show that, for manual harvesting work in climates similar to central Italy, it is essential to provide plenty of drinking water and acclimatization for the workers in order to reduce health risks. Moreover, the study emphasizes that the possible health risks for greenhouse workers increase from the month of April through July.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Background Guidelines for the prevention of coronary heart disease (CHD) recommend use of Framingham-based risk scores that were developed in white middle-aged populations. It remains unclear whether and how CHD risk prediction might be improved among older adults. We aimed to compare the prognostic performance of the Framingham risk score (FRS), directly and after recalibration, with refit functions derived from the present cohort, as well as to assess the utility of adding other routinely available risk parameters to FRS. Methods Among 2193 black and white older adults (mean age, 73.5 years) without pre-existing cardiovascular disease from the Health ABC cohort, we examined adjudicated CHD events, defined as incident myocardial infarction, CHD death, and hospitalization for angina or coronary revascularization. Results During 8-year follow-up, 351 participants experienced CHD events. The FRS poorly discriminated between persons who experienced CHD events vs. not (C-index: 0.577 in women; 0.583 in men) and underestimated absolute risk prediction by 51% in women and 8% in men. Recalibration of the FRS improved absolute risk prediction, particulary for women. For both genders, refitting these functions substantially improved absolute risk prediction, with similar discrimination to the FRS. Results did not differ between whites and blacks. The addition of lifestyle variables, waist circumference and creatinine did not improve risk prediction beyond risk factors of the FRS. Conclusions The FRS underestimates CHD risk in older adults, particularly in women, although traditional risk factors remain the best predictors of CHD. Re-estimated risk functions using these factors improve accurate estimation of absolute risk.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Mendelian models can predict who carries an inherited deleterious mutation of known disease genes based on family history. For example, the BRCAPRO model is commonly used to identify families who carry mutations of BRCA1 and BRCA2, based on familial breast and ovarian cancers. These models incorporate the age of diagnosis of diseases in relatives and current age or age of death. We develop a rigorous foundation for handling multiple diseases with censoring. We prove that any disease unrelated to mutations can be excluded from the model, unless it is sufficiently common and dependent on a mutation-related disease time. Furthermore, if a family member has a disease with higher probability density among mutation carriers, but the model does not account for it, then the carrier probability is deflated. However, even if a family only has diseases the model accounts for, if the model excludes a mutation-related disease, then the carrier probability will be inflated. In light of these results, we extend BRCAPRO to account for surviving all non-breast/ovary cancers as a single outcome. The extension also enables BRCAPRO to extract more useful information from male relatives. Using 1500 familes from the Cancer Genetics Network, accounting for surviving other cancers improves BRCAPRO’s concordance index from 0.758 to 0.762 (p = 0.046), improves its positive predictive value from 35% to 39% (p < 10−6) without impacting its negative predictive value, and improves its overall calibration, although calibration slightly worsens for those with carrier probability < 10%. Copyright c 2000 John Wiley & Sons, Ltd.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In biostatistical applications, interest often focuses on the estimation of the distribution of time T between two consecutive events. If the initial event time is observed and the subsequent event time is only known to be larger or smaller than an observed monitoring time, then the data is described by the well known singly-censored current status model, also known as interval censored data, case I. We extend this current status model by allowing the presence of a time-dependent process, which is partly observed and allowing C to depend on T through the observed part of this time-dependent process. Because of the high dimension of the covariate process, no globally efficient estimators exist with a good practical performance at moderate sample sizes. We follow the approach of Robins and Rotnitzky (1992) by modeling the censoring variable, given the time-variable and the covariate-process, i.e., the missingness process, under the restriction that it satisfied coarsening at random. We propose a generalization of the simple current status estimator of the distribution of T and of smooth functionals of the distribution of T, which is based on an estimate of the missingness. In this estimator the covariates enter only through the estimate of the missingness process. Due to the coarsening at random assumption, the estimator has the interesting property that if we estimate the missingness process more nonparametrically, then we improve its efficiency. We show that by local estimation of an optimal model or optimal function of the covariates for the missingness process, the generalized current status estimator for smooth functionals become locally efficient; meaning it is efficient if the right model or covariate is consistently estimated and it is consistent and asymptotically normal in general. Estimation of the optimal model requires estimation of the conditional distribution of T, given the covariates. Any (prior) knowledge of this conditional distribution can be used at this stage without any risk of losing root-n consistency. We also propose locally efficient one step estimators. Finally, we show some simulation results.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The development of a clinical decision tree based on knowledge about risks and reported outcomes of therapy is a necessity for successful planning and outcome of periodontal therapy. This requires a well-founded knowledge of the disease entity and a broad knowledge of how different risk conditions attribute to periodontitis. The infectious etiology, a complex immune response, and influence from a large number of co-factors are challenging conditions in clinical periodontal risk assessment. The difficult relationship between independent and dependent risk conditions paired with limited information on periodontitis prevalence adds to difficulties in periodontal risk assessment. The current information on periodontitis risk attributed to smoking habits, socio-economic conditions, general health and subjects' self-perception of health, is not comprehensive, and this contributes to limited success in periodontal risk assessment. New models for risk analysis have been advocated. Their utility for the estimation of periodontal risk assessment and prognosis should be tested. The present review addresses several of these issues associated with periodontal risk assessment.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The amount and type of ground cover is an important characteristic to measure when collecting soil disturbance monitoring data after a timber harvest. Estimates of ground cover and bare soil can be used for tracking changes in invasive species, plant growth and regeneration, woody debris loadings, and the risk of surface water runoff and soil erosion. A new method of assessing ground cover and soil disturbance was recently published by the U.S. Forest Service, the Forest Soil Disturbance Monitoring Protocol (FSDMP). This protocol uses the frequency of cover types in small circular (15cm) plots to compare ground surface in pre- and post-harvest condition. While both frequency and percent cover are common methods of describing vegetation, frequency has rarely been used to measure ground surface cover. In this study, three methods for assessing ground cover percent (step-point, 15cm dia. circular and 1x5m visual plot estimates) were compared to the FSDMP frequency method. Results show that the FSDMP method provides significantly higher estimates of ground surface condition for most soil cover types, except coarse wood. The three cover methods had similar estimates for most cover values. The FSDMP method also produced the highest value when bare soil estimates were used to model erosion risk. In a person-hour analysis, estimating ground cover percent in 15cm dia. plots required the least sampling time, and provided standard errors similar to the other cover estimates even at low sampling intensities (n=18). If ground cover estimates are desired in soil monitoring, then a small plot size (15cm dia. circle), or a step-point method can provide a more accurate estimate in less time than the current FSDMP method.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Pulse wave velocity (PWV) is a surrogate of arterial stiffness and represents a non-invasive marker of cardiovascular risk. The non-invasive measurement of PWV requires tracking the arrival time of pressure pulses recorded in vivo, commonly referred to as pulse arrival time (PAT). In the state of the art, PAT is estimated by identifying a characteristic point of the pressure pulse waveform. This paper demonstrates that for ambulatory scenarios, where signal-to-noise ratios are below 10 dB, the performance in terms of repeatability of PAT measurements through characteristic points identification degrades drastically. Hence, we introduce a novel family of PAT estimators based on the parametric modeling of the anacrotic phase of a pressure pulse. In particular, we propose a parametric PAT estimator (TANH) that depicts high correlation with the Complior(R) characteristic point D1 (CC = 0.99), increases noise robustness and reduces by a five-fold factor the number of heartbeats required to obtain reliable PAT measurements.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

BACKGROUND: The estimation of physiologic ability and surgical stress (E-PASS) has been used to produce a numerical estimate of expected mortality and morbidity after elective gastrointestinal surgery. The aim of this study was to validate E-PASS in a selected cohort of patients requiring liver resections (LR). METHODS: In this retrospective study, E-PASS predictor equations for morbidity and mortality were applied to the prospective data from 243 patients requiring LR. The observed rates were compared with predicted rates using Fisher's exact test. The discriminative capability of E-PASS was evaluated using receiver-operating characteristic (ROC) curve analysis. RESULTS: The observed and predicted overall mortality rates were both 3.3% and the morbidity rates were 31.3 and 26.9%, respectively. There was a significant difference in the comprehensive risk scores for deceased and surviving patients (p = 0.043). However, the scores for patients with or without complications were not significantly different (p = 0.120). Subsequent ROC curve analysis revealed a poor predictive accuracy for morbidity. CONCLUSIONS: The E-PASS score seems to effectively predict mortality in this specific group of patients but is a poor predictor of complications. A new modified logistic regression might be required for LR in order to better predict the postoperative outcome.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

BACKGROUND: Reduced bone mineral density (BMD) is common in adults infected with human immunodeficiency virus (HIV). The role of proximal renal tubular dysfunction (PRTD) and alterations in bone metabolism in HIV-related low BMD are incompletely understood. METHODS: We quantified BMD (dual-energy x-ray absorptiometry), blood and urinary markers of bone metabolism and renal function, and risk factors for low BMD (hip or spine T score, -1 or less) in an ambulatory care setting. We determined factors associated with low BMD and calculated 10-year fracture risks using the World Health Organization FRAX equation. RESULTS: We studied 153 adults (98% men; median age, 48 years; median body mass index, 24.5; 67 [44%] were receiving tenofovir, 81 [53%] were receiving a boosted protease inhibitor [PI]). Sixty-five participants (42%) had low BMD, and 11 (7%) had PRTD. PI therapy was associated with low BMD in multivariable analysis (odds ratio, 2.69; 95% confidence interval, 1.09-6.63). Tenofovir use was associated with increased osteoblast and osteoclast activity (P< or = .002). The mean estimated 10-year risks were 1.2% for hip fracture and 5.4% for any major osteoporotic fracture. CONCLUSIONS: In this mostly male population, low BMD was significantly associated with PI therapy. Tenofovir recipients showed evidence of increased bone turnover. Measurement of BMD and estimation of fracture risk may be warranted in treated HIV-infected adults.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

BACKGROUND In recent years, the occurrence and the relevance of Mycoplasma hyopneumoniae infections in suckling pigs has been examined in several studies. Whereas most of these studies were focused on sole prevalence estimation within different age groups, follow-up of infected piglets or assessment of pathological findings, none of the studies included a detailed analysis of individual and environmental risk factors. Therefore, the aim of the present study was to investigate the frequency of M. hyopneumoniae infections in suckling pigs of endemically infected herds and to identify individual risk factors potentially influencing the infection status of suckling pigs at the age of weaning. RESULTS The animal level prevalence of M. hyopneumoniae infections in suckling pigs examined in three conventional pig breeding herds was 3.6% (41/1127) at the time of weaning. A prevalence of 1.2% was found in the same pigs at the end of their nursery period. In a multivariable Poisson regression model it was found that incidence rate ratios (IRR) for suckling pigs are significantly lower than 1 when teeth grinding was conducted (IRR: 0.10). Moreover, high temperatures in the piglet nest during the first two weeks of life (occasionally >40°C) were associated with a decrease of the probability of an infection (IRR: 0.23-0.40). Contrary, the application of PCV2 vaccines to piglets was associated with an increased infection risk (IRR: 9.72). CONCLUSIONS Since single infected piglets are supposed to act as initiators for the transmission of this pathogen in nursery and fattening pigs, the elimination of the risk factors described in this study should help to reduce the incidence rate of M. hyopneumoniae infections and thereby might contribute to a reduced probability of high prevalences in older pigs.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

BACKGROUND The role of surgery for patients with metastatic esophagogastric adenocarcinoma (EGC) is not defined. The purpose of this study was to define selection criteria for patients who may benefit from resection following systemic chemotherapy. METHODS From 1987 to 2007, 160 patients presenting with synchronous metastatic EGC (cT3/4 cNany cM0/1 finally pM1) were treated with chemotherapy followed by resection of the primary tumor and metastases. Clinical and histopathological data, site and number of metastases were analyzed. A prognostic score was established and validated in a second cohort from another academic center (n = 32). RESULTS The median survival (MS) in cohort 1 was 13.6 months. Significant prognostic factors were grading (p = 0.046), ypT- (p = 0.001), ypN- (p = 0.011) and R-category (p = 0.015), lymphangiosis (p = 0.021), clinical (p = 0.004) and histopathological response (p = 0.006), but not localization or number of metastases. The addition of grading (G1/2:0 points; G3/4:1 points), clinical response (responder: 0; nonresponder: 1) and R-category (complete:0; R1:1; R2:2) defines two groups of patients with significantly different survival (p = 0.001) [low risk group (Score 0/1), n = 22: MS 35.3 months, 3-year-survival 47.6%); high risk group (Score 2/3/4) n = 126: MS 12.0 months, 3-year-survival 14.2%]. The score showed a strong trend in the validation cohort (p = 0.063) [low risk group (MS not reached, 3-year-survival 57.1%); high risk group (MS 19.9 months, 3-year-survival 6.7%)]. CONCLUSION We observed long-term survival after resection of metastatic EGC. A simple clinical score may help to identify a subgroup of patients with a high chance of benefit from resection. However, the accurate estimation of achieving a complete resection, which is an integral element of the score, remains challenging.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

OBJECTIVE Cognitive impairments are regarded as a core component of schizophrenia. However, the cognitive dimension of psychosis is hardly considered by ultra-high risk (UHR) criteria. Therefore, we studied whether the combination of symptomatic UHR criteria and the basic symptom criterion "cognitive disturbances" (COGDIS) is superior in predicting first-episode psychosis. METHOD In a naturalistic 48-month follow-up study, the conversion rate to first-episode psychosis was studied in 246 outpatients of an early detection of psychosis service (FETZ); thereby, the association between conversion, and the combined and singular use of UHR criteria and COGDIS was compared. RESULTS Patients that met UHR criteria and COGDIS (n=127) at baseline had a significantly higher risk of conversion (hr=0.66 at month 48) and a shorter time to conversion than patients that met only UHR criteria (n=37; hr=0.28) or only COGDIS (n=30; hr=0.23). Furthermore, the risk of conversion was higher for the combined criteria than for UHR criteria (n=164; hr=0.56 at month 48) and COGDIS (n=158; hr=0.56 at month 48) when considered irrespective of each other. CONCLUSIONS Our findings support the merits of considering both COGDIS and UHR criteria in the early detection of persons who are at high risk of developing a first psychotic episode within 48months. Applying both sets of criteria improves sensitivity and individual risk estimation, and may thereby support the development of stage-targeted interventions. Moreover, since the combined approach enables the identification of considerably more homogeneous at-risk samples, it should support both preventive and basic research.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

BACKGROUND Prediction studies in subjects at Clinical High Risk (CHR) for psychosis are hampered by a high proportion of uncertain outcomes. We therefore investigated whether quantitative EEG (QEEG) parameters can contribute to an improved identification of CHR subjects with a later conversion to psychosis. METHODS This investigation was a project within the European Prediction of Psychosis Study (EPOS), a prospective multicenter, naturalistic field study with an 18-month follow-up period. QEEG spectral power and alpha peak frequencies (APF) were determined in 113 CHR subjects. The primary outcome measure was conversion to psychosis. RESULTS Cox regression yielded a model including frontal theta (HR=1.82; [95% CI 1.00-3.32]) and delta (HR=2.60; [95% CI 1.30-5.20]) power, and occipital-parietal APF (HR=.52; [95% CI .35-.80]) as predictors of conversion to psychosis. The resulting equation enabled the development of a prognostic index with three risk classes (hazard rate 0.057 to 0.81). CONCLUSIONS Power in theta and delta ranges and APF contribute to the short-term prediction of psychosis and enable a further stratification of risk in CHR samples. Combined with (other) clinical ratings, EEG parameters may therefore be a useful tool for individualized risk estimation and, consequently, targeted prevention.