868 resultados para Negative Binomial Regression Model (NBRM)
Resumo:
The health impacts of exposure to ambient temperature have been drawing increasing attention from the environmental health research community, government, society, industries, and the public. Case-crossover and time series models are most commonly used to examine the effects of ambient temperature on mortality. However, some key methodological issues remain to be addressed. For example, few studies have used spatiotemporal models to assess the effects of spatial temperatures on mortality. Few studies have used a case-crossover design to examine the delayed (distributed lag) and non-linear relationship between temperature and mortality. Also, little evidence is available on the effects of temperature changes on mortality, and on differences in heat-related mortality over time. This thesis aimed to address the following research questions: 1. How to combine case-crossover design and distributed lag non-linear models? 2. Is there any significant difference in effect estimates between time series and spatiotemporal models? 3. How to assess the effects of temperature changes between neighbouring days on mortality? 4. Is there any change in temperature effects on mortality over time? To combine the case-crossover design and distributed lag non-linear model, datasets including deaths, and weather conditions (minimum temperature, mean temperature, maximum temperature, and relative humidity), and air pollution were acquired from Tianjin China, for the years 2005 to 2007. I demonstrated how to combine the case-crossover design with a distributed lag non-linear model. This allows the case-crossover design to estimate the non-linear and delayed effects of temperature whilst controlling for seasonality. There was consistent U-shaped relationship between temperature and mortality. Cold effects were delayed by 3 days, and persisted for 10 days. Hot effects were acute and lasted for three days, and were followed by mortality displacement for non-accidental, cardiopulmonary, and cardiovascular deaths. Mean temperature was a better predictor of mortality (based on model fit) than maximum or minimum temperature. It is still unclear whether spatiotemporal models using spatial temperature exposure produce better estimates of mortality risk compared with time series models that use a single site’s temperature or averaged temperature from a network of sites. Daily mortality data were obtained from 163 locations across Brisbane city, Australia from 2000 to 2004. Ordinary kriging was used to interpolate spatial temperatures across the city based on 19 monitoring sites. A spatiotemporal model was used to examine the impact of spatial temperature on mortality. A time series model was used to assess the effects of single site’s temperature, and averaged temperature from 3 monitoring sites on mortality. Squared Pearson scaled residuals were used to check the model fit. The results of this study show that even though spatiotemporal models gave a better model fit than time series models, spatiotemporal and time series models gave similar effect estimates. Time series analyses using temperature recorded from a single monitoring site or average temperature of multiple sites were equally good at estimating the association between temperature and mortality as compared with a spatiotemporal model. A time series Poisson regression model was used to estimate the association between temperature change and mortality in summer in Brisbane, Australia during 1996–2004 and Los Angeles, United States during 1987–2000. Temperature change was calculated by the current day's mean temperature minus the previous day's mean. In Brisbane, a drop of more than 3 �C in temperature between days was associated with relative risks (RRs) of 1.16 (95% confidence interval (CI): 1.02, 1.31) for non-external mortality (NEM), 1.19 (95% CI: 1.00, 1.41) for NEM in females, and 1.44 (95% CI: 1.10, 1.89) for NEM aged 65.74 years. An increase of more than 3 �C was associated with RRs of 1.35 (95% CI: 1.03, 1.77) for cardiovascular mortality and 1.67 (95% CI: 1.15, 2.43) for people aged < 65 years. In Los Angeles, only a drop of more than 3 �C was significantly associated with RRs of 1.13 (95% CI: 1.05, 1.22) for total NEM, 1.25 (95% CI: 1.13, 1.39) for cardiovascular mortality, and 1.25 (95% CI: 1.14, 1.39) for people aged . 75 years. In both cities, there were joint effects of temperature change and mean temperature on NEM. A change in temperature of more than 3 �C, whether positive or negative, has an adverse impact on mortality even after controlling for mean temperature. I examined the variation in the effects of high temperatures on elderly mortality (age . 75 years) by year, city and region for 83 large US cities between 1987 and 2000. High temperature days were defined as two or more consecutive days with temperatures above the 90th percentile for each city during each warm season (May 1 to September 30). The mortality risk for high temperatures was decomposed into: a "main effect" due to high temperatures using a distributed lag non-linear function, and an "added effect" due to consecutive high temperature days. I pooled yearly effects across regions and overall effects at both regional and national levels. The effects of high temperature (both main and added effects) on elderly mortality varied greatly by year, city and region. The years with higher heat-related mortality were often followed by those with relatively lower mortality. Understanding this variability in the effects of high temperatures is important for the development of heat-warning systems. In conclusion, this thesis makes contribution in several aspects. Case-crossover design was combined with distribute lag non-linear model to assess the effects of temperature on mortality in Tianjin. This makes the case-crossover design flexibly estimate the non-linear and delayed effects of temperature. Both extreme cold and high temperatures increased the risk of mortality in Tianjin. Time series model using single site’s temperature or averaged temperature from some sites can be used to examine the effects of temperature on mortality. Temperature change (no matter significant temperature drop or great temperature increase) increases the risk of mortality. The high temperature effect on mortality is highly variable from year to year.
Resumo:
Extending recent research on the importance of specific resources and skills for the internationalization of start-ups, this article tests a negative binomial model on a sample of 520 recently created high technology firms from the UK and Germany. The results show that previous international experience of entrepreneurs facilitates the rapid penetration of foreign markets, especially when the company features a clear and deliberate strategic intent of internationalization from the outset. This research provides one of the first empirical studies linking the influence of entrepreneurial teams to a high probability of success in the internationalization of high-technology ventures.
Resumo:
Background: Developing sampling strategies to target biological pests such as insects in stored grain is inherently difficult owing to species biology and behavioural characteristics. The design of robust sampling programmes should be based on an underlying statistical distribution that is sufficiently flexible to capture variations in the spatial distribution of the target species. Results: Comparisons are made of the accuracy of four probability-of-detection sampling models - the negative binomial model,1 the Poisson model,1 the double logarithmic model2 and the compound model3 - for detection of insects over a broad range of insect densities. Although the double log and negative binomial models performed well under specific conditions, it is shown that, of the four models examined, the compound model performed the best over a broad range of insect spatial distributions and densities. In particular, this model predicted well the number of samples required when insect density was high and clumped within experimental storages. Conclusions: This paper reinforces the need for effective sampling programs designed to detect insects over a broad range of spatial distributions. The compound model is robust over a broad range of insect densities and leads to substantial improvement in detection probabilities within highly variable systems such as grain storage.
Resumo:
Nitrous oxide (N2O) is one of the greenhouse gases that can contribute to global warming. Spatial variability of N2O can lead to large uncertainties in prediction. However, previous studies have often ignored the spatial dependency to quantify the N2O - environmental factors relationships. Few researches have examined the impacts of various spatial correlation structures (e.g. independence, distance-based and neighbourhood based) on spatial prediction of N2O emissions. This study aimed to assess the impact of three spatial correlation structures on spatial predictions and calibrate the spatial prediction using Bayesian model averaging (BMA) based on replicated, irregular point-referenced data. The data were measured in 17 chambers randomly placed across a 271 m(2) field between October 2007 and September 2008 in the southeast of Australia. We used a Bayesian geostatistical model and a Bayesian spatial conditional autoregressive (CAR) model to investigate and accommodate spatial dependency, and to estimate the effects of environmental variables on N2O emissions across the study site. We compared these with a Bayesian regression model with independent errors. The three approaches resulted in different derived maps of spatial prediction of N2O emissions. We found that incorporating spatial dependency in the model not only substantially improved predictions of N2O emission from soil, but also better quantified uncertainties of soil parameters in the study. The hybrid model structure obtained by BMA improved the accuracy of spatial prediction of N2O emissions across this study region.
Resumo:
The use of graphical processing unit (GPU) parallel processing is becoming a part of mainstream statistical practice. The reliance of Bayesian statistics on Markov Chain Monte Carlo (MCMC) methods makes the applicability of parallel processing not immediately obvious. It is illustrated that there are substantial gains in improved computational time for MCMC and other methods of evaluation by computing the likelihood using GPU parallel processing. Examples use data from the Global Terrorism Database to model terrorist activity in Colombia from 2000 through 2010 and a likelihood based on the explicit convolution of two negative-binomial processes. Results show decreases in computational time by a factor of over 200. Factors influencing these improvements and guidelines for programming parallel implementations of the likelihood are discussed.
Resumo:
We present a novel approach for developing summary statistics for use in approximate Bayesian computation (ABC) algorithms using indirect infer- ence. We embed this approach within a sequential Monte Carlo algorithm that is completely adaptive. This methodological development was motivated by an application involving data on macroparasite population evolution modelled with a trivariate Markov process. The main objective of the analysis is to compare inferences on the Markov process when considering two di®erent indirect mod- els. The two indirect models are based on a Beta-Binomial model and a three component mixture of Binomials, with the former providing a better ¯t to the observed data.
Resumo:
Real world business process models may consist of hundreds of elements and have sophisticated structure. Although there are tasks where such models are valuable and appreciated, in general complexity has a negative influence on model comprehension and analysis. Thus, means for managing the complexity of process models are needed. One approach is abstraction of business process models-creation of a process model which preserves the main features of the initial elaborate process model, but leaves out insignificant details. In this paper we study the structural aspects of process model abstraction and introduce an abstraction approach based on process structure trees (PST). The developed approach assures that the abstracted process model preserves the ordering constraints of the initial model. It surpasses pattern-based process model abstraction approaches, allowing to handle graph-structured process models of arbitrary structure. We also provide an evaluation of the proposed approach.
Resumo:
A cross-sectional survey of female office workers (n=333) was undertaken to determine the level of neck pain and disability (Neck Disability Index—NDI) and to explore the relationship between individual and workplace risk factors with the NDI score and the presence of pain. Workers reported nil (32%), mild (53%), moderate (14%) and severe (1%) neck pain. There were more risk factors associated with the NDI score than the presence of neck pain. The presence of neck pain was associated with a history of neck trauma (OR: 4.8), using a graduated lens (OR: 4.6), and negative affectivity (OR: 2.7) in the multiple regression model. Factors associated with higher NDI score were using the computer mouse for more than 6 h per day, higher negative affectivity, older age and an uncomfortable workstation. These results suggest that measuring the level of neck pain and disability rather than just the presence of neck pain provides more specific directives for the prevention and management of this disorder.
Resumo:
This study investigated the relative contribution of individual, workplace, psychosocial and physiological features associated with neck pain in female office workers towards developing appropriate intervention programs. Workers without disability (Neck Disability Index (NDI) score≤8, n=33); workers with neck pain and disability (NDI≥9/100, n=52) and 22 controls (women who did not work and without neck pain) participated in this study. Two logistic regression models were constructed to test the association between various measures in (1) workers with and without disability, and (2) workers without disability and controls. Measures included those found to be significantly associated with higher NDI in our previous studies: psychosocial domains; individual factors; task demands; quantitative sensory measures and measures of motor function. In the final model, higher score on negative affectivity scale (OR=4.47), greater activity in the neck flexors during cranio-cervical flexion (OR=1.44), cold hyperalgesia (OR=1.27) and longer duration of symptoms (OR=1.19) remained significantly associated with neck pain in workers. Workers without disability and controls could only be differentiated by greater muscle activity in the cervical flexors and extensors during a typing task. No psychosocial domains remained in either regression model. These results suggest that impairments in the sensory and motor system should be considered in any assessment of the office worker with neck pain and may have stronger influences on the presenting symptoms than workplace and psychosocial features.
Resumo:
This paper uses a correlated multinomial logit model and a Poisson regression model to measure the factors affecting demand for different types of transportation by elderly and disabled people in rural Virginia. The major results are: (a) A paratransit system providing door-to-door service is highly valued by transportation-handicapped people; (b) Taxis are probably a potential but inferior alternative even when subsidized; (c) Buses are a poor alternative, especially in rural areas where distances to bus stops may be long; (d) Making buses handicap-accessible would have a statistically significant but small effect on mode choice; (e) Demand is price inelastic; and (f) The total number of trips taken is insensitive to mode availability and characteristics. These results suggest that transportation-handicapped people take a limited number of trips. Those they do take are in some sense necessary (given the low elasticity with respect to mode price or availability). People will substitute away from relying upon others when appropriate transportation is available, at least to some degree. But such transportation needs to be flexible enough to meet the needs of the people involved.
Impact of child labor on academic performance : evidence from the program "Edúcame Primero Colombia"
Resumo:
In this study, the effects of different variables of child labor on academic performance are investigated. To this end, 3302 children participating in the child labor eradication program “Edúcame Primero Colombia” were interviewed. The interview format used for the children's enrollment into the program was a template from which socioeconomic conditions, academic performance, and child labor variables were evaluated. The academic performance factor was determined using the Analytic Hierarchy Process (AHP). The data were analyzed through a logistic regression model that took into account children who engaged in a type of labor (n = 921). The results showed that labor conditions, the number of weekly hours dedicated to work, and the presence of work scheduled in the morning negatively affected the academic performance of child laborers. These results show that the relationship between child labor and academic performance is based on the conflict between these two activities. These results do not indicate a linear and simple relationship associated with the recognition of the presence or absence of child labor. This study has implications for the formulation of policies, programs, and interventions for preventing, eradicating, and attenuating the negative effects of child labor on the social and educational development of children.
Resumo:
Narrative text is a useful way of identifying injury circumstances from the routine emergency department data collections. Automatically classifying narratives based on machine learning techniques is a promising technique, which can consequently reduce the tedious manual classification process. Existing works focus on using Naive Bayes which does not always offer the best performance. This paper proposes the Matrix Factorization approaches along with a learning enhancement process for this task. The results are compared with the performance of various other classification approaches. The impact on the classification results from the parameters setting during the classification of a medical text dataset is discussed. With the selection of right dimension k, Non Negative Matrix Factorization-model method achieves 10 CV accuracy of 0.93.
Resumo:
Understanding pedestrian crash causes and contributing factors in developing countries is critically important as they account for about 55% of all traffic crashes. Not surprisingly, considerable attention in the literature has been paid to road traffic crash prediction models and methodologies in developing countries of late. Despite this interest, there are significant challenges confronting safety managers in developing countries. For example, in spite of the prominence of pedestrian crashes occurring on two-way two-lane rural roads, it has proven difficult to develop pedestrian crash prediction models due to a lack of both traffic and pedestrian exposure data. This general lack of available data has further hampered identification of pedestrian crash causes and subsequent estimation of pedestrian safety performance functions. The challenges are similar across developing nations, where little is known about the relationship between pedestrian crashes, traffic flow, and road environment variables on rural two-way roads, and where unique predictor variables may be needed to capture the unique crash risk circumstances. This paper describes pedestrian crash safety performance functions for two-way two-lane rural roads in Ethiopia as a function of traffic flow, pedestrian flows, and road geometry characteristics. In particular, random parameter negative binomial model was used to investigate pedestrian crashes. The models and their interpretations make important contributions to road crash analysis and prevention in developing countries. They also assist in the identification of the contributing factors to pedestrian crashes, with the intent to identify potential design and operational improvements.
Resumo:
Ordinal qualitative data are often collected for phenotypical measurements in plant pathology and other biological sciences. Statistical methods, such as t tests or analysis of variance, are usually used to analyze ordinal data when comparing two groups or multiple groups. However, the underlying assumptions such as normality and homogeneous variances are often violated for qualitative data. To this end, we investigated an alternative methodology, rank regression, for analyzing the ordinal data. The rank-based methods are essentially based on pairwise comparisons and, therefore, can deal with qualitative data naturally. They require neither normality assumption nor data transformation. Apart from robustness against outliers and high efficiency, the rank regression can also incorporate covariate effects in the same way as the ordinary regression. By reanalyzing a data set from a wheat Fusarium crown rot study, we illustrated the use of the rank regression methodology and demonstrated that the rank regression models appear to be more appropriate and sensible for analyzing nonnormal data and data with outliers.
Resumo:
This article is motivated by a lung cancer study where a regression model is involved and the response variable is too expensive to measure but the predictor variable can be measured easily with relatively negligible cost. This situation occurs quite often in medical studies, quantitative genetics, and ecological and environmental studies. In this article, by using the idea of ranked-set sampling (RSS), we develop sampling strategies that can reduce cost and increase efficiency of the regression analysis for the above-mentioned situation. The developed method is applied retrospectively to a lung cancer study. In the lung cancer study, the interest is to investigate the association between smoking status and three biomarkers: polyphenol DNA adducts, micronuclei, and sister chromatic exchanges. Optimal sampling schemes with different optimality criteria such as A-, D-, and integrated mean square error (IMSE)-optimality are considered in the application. With set size 10 in RSS, the improvement of the optimal schemes over simple random sampling (SRS) is great. For instance, by using the optimal scheme with IMSE-optimality, the IMSEs of the estimated regression functions for the three biomarkers are reduced to about half of those incurred by using SRS.