922 resultados para Generalized Linear Model


Relevância:

90.00% 90.00%

Publicador:

Resumo:

We compare Bayesian methodology utilizing free-ware BUGS (Bayesian Inference Using Gibbs Sampling) with the traditional structural equation modelling approach based on another free-ware package, Mx. Dichotomous and ordinal (three category) twin data were simulated according to different additive genetic and common environment models for phenotypic variation. Practical issues are discussed in using Gibbs sampling as implemented by BUGS to fit subject-specific Bayesian generalized linear models, where the components of variation may be estimated directly. The simulation study (based on 2000 twin pairs) indicated that there is a consistent advantage in using the Bayesian method to detect a correct model under certain specifications of additive genetics and common environmental effects. For binary data, both methods had difficulty in detecting the correct model when the additive genetic effect was low (between 10 and 20%) or of moderate range (between 20 and 40%). Furthermore, neither method could adequately detect a correct model that included a modest common environmental effect (20%) even when the additive genetic effect was large (50%). Power was significantly improved with ordinal data for most scenarios, except for the case of low heritability under a true ACE model. We illustrate and compare both methods using data from 1239 twin pairs over the age of 50 years, who were registered with the Australian National Health and Medical Research Council Twin Registry (ATR) and presented symptoms associated with osteoarthritis occurring in joints of the hand.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

This paper proposes a template for modelling complex datasets that integrates traditional statistical modelling approaches with more recent advances in statistics and modelling through an exploratory framework. Our approach builds on the well-known and long standing traditional idea of 'good practice in statistics' by establishing a comprehensive framework for modelling that focuses on exploration, prediction, interpretation and reliability assessment, a relatively new idea that allows individual assessment of predictions. The integrated framework we present comprises two stages. The first involves the use of exploratory methods to help visually understand the data and identify a parsimonious set of explanatory variables. The second encompasses a two step modelling process, where the use of non-parametric methods such as decision trees and generalized additive models are promoted to identify important variables and their modelling relationship with the response before a final predictive model is considered. We focus on fitting the predictive model using parametric, non-parametric and Bayesian approaches. This paper is motivated by a medical problem where interest focuses on developing a risk stratification system for morbidity of 1,710 cardiac patients given a suite of demographic, clinical and preoperative variables. Although the methods we use are applied specifically to this case study, these methods can be applied across any field, irrespective of the type of response.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Em geral, produtos agrícolas são produzidos em larga escala e essa produtividade cresce proporcionalmente ao seu consumo. Entretanto, outro fator também cresce de forma proporcional, as perdas pós-colheita, o que sugere a utilização de tecnologias para aumentar a utilização desses produtos mitigando o desperdício e aumentando sua a vida de prateleira. Além disso, oferecer o produto durante o período de entressafra. No presente trabalho, foi utilizado à tecnologia de secagem em leito de espuma aplicada a cenoura, beterraba, tomate e morango, produtos amplamente produzidos e consumidos no Brasil. Neste trabalho, os quatros produtos foram submetidos à secagem em leito de espuma em secador com ar circulado em temperaturas controladas de 40, 50, 60, 70 e 80 °C. A descrição da cinética de secagem foi realizada pelo ajuste de modelos matemáticos para cada temperatura do ar de secagem. Além disso, foi proposto um modelo matemático generalizado ajustado por regressão não linear. O modelo de Page obteve o melhor ajuste sobre os dados de secagem em todos os produtos testados, com um coeficiente de determinação (R²) superior a 98% em todas as temperaturas avaliadas. Além disso, foi possível modelar a influência da temperatura do ar sobre o parâmetro k do modelo de Page através da utilização de um modelo exponencial. O coeficiente de difusão efetiva aumentou com a elevação da temperatura, apresentando valores entre 10-8e 10-7 m².s-¹ para as temperaturas de processo. A relação entre o coeficiente de difusão efetiva e a temperatura de secagem pôde ser descrita pela equação de Arrhenius.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Dissertação de Mestrado, Estudos Integrados dos Oceanos, 25 de Março de 2013, Universidade dos Açores.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Consider a multihop network comprising Ethernet switches. The traffic is described with flows and each flow is characterized by its source node, its destination node, its route and parameters in the generalized multiframe model. Output queues on Ethernet switches are scheduled by static-priority scheduling and tasks executing on the processor in an Ethernet switch are scheduled by stride scheduling. We present schedulability analysis for this setting.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Submitted in partial fulfillment for the Requirements for the Degree of PhD in Mathematics, in the Speciality of Statistics in the Faculdade de Ciências e Tecnologia

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Dissertação de mestrado integrado em Engenharia Civil

Relevância:

90.00% 90.00%

Publicador:

Resumo:

There is recent interest in the generalization of classical factor models in which the idiosyncratic factors are assumed to be orthogonal and there are identification restrictions on cross-sectional and time dimensions. In this study, we describe and implement a Bayesian approach to generalized factor models. A flexible framework is developed to determine the variations attributed to common and idiosyncratic factors. We also propose a unique methodology to select the (generalized) factor model that best fits a given set of data. Applying the proposed methodology to the simulated data and the foreign exchange rate data, we provide a comparative analysis between the classical and generalized factor models. We find that when there is a shift from classical to generalized, there are significant changes in the estimates of the structures of the covariance and correlation matrices while there are less dramatic changes in the estimates of the factor loadings and the variation attributed to common factors.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Species distribution models (SDMs) are widely used to explain and predict species ranges and environmental niches. They are most commonly constructed by inferring species' occurrence-environment relationships using statistical and machine-learning methods. The variety of methods that can be used to construct SDMs (e.g. generalized linear/additive models, tree-based models, maximum entropy, etc.), and the variety of ways that such models can be implemented, permits substantial flexibility in SDM complexity. Building models with an appropriate amount of complexity for the study objectives is critical for robust inference. We characterize complexity as the shape of the inferred occurrence-environment relationships and the number of parameters used to describe them, and search for insights into whether additional complexity is informative or superfluous. By building 'under fit' models, having insufficient flexibility to describe observed occurrence-environment relationships, we risk misunderstanding the factors shaping species distributions. By building 'over fit' models, with excessive flexibility, we risk inadvertently ascribing pattern to noise or building opaque models. However, model selection can be challenging, especially when comparing models constructed under different modeling approaches. Here we argue for a more pragmatic approach: researchers should constrain the complexity of their models based on study objective, attributes of the data, and an understanding of how these interact with the underlying biological processes. We discuss guidelines for balancing under fitting with over fitting and consequently how complexity affects decisions made during model building. Although some generalities are possible, our discussion reflects differences in opinions that favor simpler versus more complex models. We conclude that combining insights from both simple and complex SDM building approaches best advances our knowledge of current and future species ranges.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Question Does a land-use variable improve spatial predictions of plant species presence-absence and abundance models at the regional scale in a mountain landscape? Location Western Swiss Alps. Methods Presence-absence generalized linear models (GLM) and abundance ordinal logistic regression models (LRM) were fitted to data on 78 mountain plant species, with topo-climatic and/or land-use variables available at a 25-m resolution. The additional contribution of land use when added to topo-climatic models was evaluated by: (1) assessing the changes in model fit and (2) predictive power, (3) partitioning the deviance respectively explained by the topo-climatic variables and the land-use variable through variation partitioning, and (5) comparing spatial projections. Results Land use significantly improved the fit of presence-absence models but not their predictive power. In contrast, land use significantly improved both the fit and predictive power of abundance models. Variation partitioning also showed that the individual contribution of land use to the deviance explained by presence-absence models was, on average, weak for both GLM and LRM (3.7% and 4.5%, respectively), but changes in spatial projections could nevertheless be important for some species. Conclusions In this mountain area and at our regional scale, land use is important for predicting abundance, but not presence-absence. The importance of adding land-use information depends on the species considered. Even without a marked effect on model fit and predictive performance, adding land use can affect spatial projections of both presence-absence and abundance models.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

We present models predicting the potential distribution of a threatened ant species, Formica exsecta Nyl., in the Swiss National Park ( SNP). Data to fit the models have been collected according to a random-stratified design with an equal number of replicates per stratum. The basic aim of such a sampling strategy is to allow the formal testing of biological hypotheses about those factors most likely to account for the distribution of the modeled species. The stratifying factors used in this study were: vegetation, slope angle and slope aspect, the latter two being used as surrogates of solar radiation, considered one of the basic requirements of F. exsecta. Results show that, although the basic stratifying predictors account for more than 50% of the deviance, the incorporation of additional non-spatially explicit predictors into the model, as measured in the field, allows for an increased model performance (up to nearly 75%). However, this was not corroborated by permutation tests. Implementation on a national scale was made for one model only, due to the difficulty of obtaining similar predictors on this scale. The resulting map on the national scale suggests that the species might once have had a broader distribution in Switzerland. Reasons for its particular abundance within the SNP might possibly be related to habitat fragmentation and vegetation transformation outside the SNP boundaries.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

During adolescence, cognitive abilities increase robustly. To search for possible related structural alterations of the cerebral cortex, we measured neuronal soma dimension (NSD = width times height), cortical thickness and neuronal densities in different types of neocortex in post-mortem brains of five 12-16 and five 17-24 year-olds (each 2F, 3M). Using a generalized mixed model analysis, mean normalized NSD comparing the age groups shows layer-specific change for layer 2 (p < .0001) and age-related differences between categorized type of cortex: primary/primary association cortex (BA 1, 3, 4, and 44) shows a generalized increase; higher-order regions (BA 9, 21, 39, and 45) also show increase in layers 2 and 5 but decrease in layers 3, 4, and 6 while limbic/orbital cortex (BA 23, 24, and 47) undergoes minor decrease (BA 1, 3, 4, and 44 vs. BA 9, 21, 39, and 45: p = .036 and BA 1, 3, 4, and 44 vs. BA 23, 24, and 47: p = .004). These data imply the operation of cortical layer- and type-specific processes of growth and regression adding new evidence that the human brain matures during adolescence not only functionally but also structurally.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Abstract OBJECTIVE To identify the factors associated with involuntary hospital admissions of technology-dependent children, in the municipality of Ribeirão Preto, São Paulo State, Brazil. METHOD A cross-sectional study, with a quantitative approach. After an active search, 124 children who qualified under the inclusion criteria, that is to say, children from birth to age 12, were identified. Data was collected in home visits to mothers or the people responsible for the children, through the application of a questionnaire. Analysis of the data followed the assumptions of the Generalized Linear Models technique. RESULTS 102 technology-dependent children aged between 6 months and 12 years participated in the study, of whom 57% were male. The average number of involuntary hospital admissions in the previous year among the children studied was 0.71 (±1.29). In the final model the following variables were significantly associated with the outcome: age (OR=0.991; CI95%=0.985-0.997), and the number of devices (OR=0.387; CI95%=0.219-0.684), which were characterized as factors of protection and quantity of medications (OR=1.532; CI95%=1.297-1.810), representing a risk factor for involuntary hospital admissions in technology-dependent children. CONCLUSION The results constitute input data for consideration of the process of care for technology-dependent children by supplying an explanatory model for involuntary hospital admissions for this client group.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

La regressió basada en distàncies és un mètode de predicció que consisteix en dos passos: a partir de les distàncies entre observacions obtenim les variables latents, les quals passen a ser els regressors en un model lineal de mínims quadrats ordinaris. Les distàncies les calculem a partir dels predictors originals fent us d'una funció de dissimilaritats adequada. Donat que, en general, els regressors estan relacionats de manera no lineal amb la resposta, la seva selecció amb el test F usual no és possible. En aquest treball proposem una solució a aquest problema de selecció de predictors definint tests estadístics generalitzats i adaptant un mètode de bootstrap no paramètric per a l'estimació dels p-valors. Incluim un exemple numèric amb dades de l'assegurança d'automòbils.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

En este documento se ilustra de un modo práctico, el empleo de tres instrumentos que permiten al actuario definir grupos arancelarios y estimar premios de riesgo en el proceso que tasa la clase para el seguro de no vida. El primero es el análisis de segmentación (CHAID y XAID) usado en primer lugar en 1997 por UNESPA en su cartera común de coches. El segundo es un proceso de selección gradual con el modelo de regresión a base de distancia. Y el tercero es un proceso con el modelo conocido y generalizado de regresión linear, que representa la técnica más moderna en la bibliografía actuarial. De estos últimos, si combinamos funciones de eslabón diferentes y distribuciones de error, podemos obtener el aditivo clásico y modelos multiplicativos