965 resultados para nested multinomial logit
Resumo:
Improving the performance of private sector small and medium sized enterprises (SMEs) in a cost effective manner is a major concern for government. Governments have saved costs by moving information online rather than through more expensive face-to-face exchanges between advisers and clients. Building on previous work that distinguished between types of advice, this article evaluates whether these changes to delivery mechanisms affect the type of advice received. Using a multinomial logit model of 1334 cases of business advice to small firms collected in England, the study found that advice to improve capabilities was taken by smaller firms who were less likely to have limited liability or undertake business planning. SMEs sought word-of-mouth referrals before taking internal, capability-enhancing advice. This is also the case when that advice was part of a wider package of assistance involving both internal and external aspects. Only when firms took advice that used extant capabilities did they rely on the Internet. Therefore, when the Internet is privileged over face-to-face advice the changes made by each recipient of advice are likely to diminish causing less impact from advice within the economy. It implies that fewer firms will adopt the sorts of management practices that would improve their productivity. © 2014 Taylor & Francis.
Resumo:
Objective: The objective of the study is to explore preferences of gastroenterologists for biosimilar drugs in Crohn’s Disease and reveal trade-offs between the perceived risks and benefits related to biosimilar drugs. Method: Discrete choice experiment was carried out involving 51 Hungarian gastroenterologists in May, 2014. The following attributes were used to describe hypothetical choice sets: 1) type of the treatment (biosimilar/originator) 2) severity of disease 3) availability of continuous medicine supply 4) frequency of the efficacy check-ups. Multinomial logit model was used to differentiate between three attitude types: 1) always opting for the originator 2) willing to consider biosimilar for biological-naïve patients only 3) willing to consider biosimilar treatment for both types of patients. Conditional logit model was used to estimate the probabilities of choosing a given profile. Results: Men, senior consultants, working in IBD center and treating more patients are more likely to willing to consider biosimilar for biological-naïve patients only. Treatment type (originator/biosimilar) was the most important determinant of choice for patients already treated with biologicals, and the availability of continuous medicine supply in the case biological-naïve patients. The probabilities of choosing the biosimilar with all the benefits offered over the originator under current reimbursement conditions are 89% vs 11% for new patients, and 44% vs 56% for patients already treated with biological. Conclusions: Gastroenterologists were willing to trade between perceived risks and benefits of biosimilars. The continuous medical supply would be one of the major benefits of biosimilars. However, benefits offered in the scenarios do not compensate for the change from the originator to the biosimilar treatment of patients already treated with biologicals.
Resumo:
The goal of this study was to develop Multinomial Logit models for the mode choice behavior of immigrants, with key focuses on neighborhood effects and behavioral assimilation. The first aspect shows the relationship between social network ties and immigrants’ chosen mode of transportation, while the second aspect explores the gradual changes toward alternative mode usage with regard to immigrants’ migrating period in the United States (US). Mode choice models were developed for work, shopping, social, recreational, and other trip purposes to evaluate the impacts of various land use patterns, neighborhood typology, socioeconomic-demographic and immigrant related attributes on individuals’ travel behavior. Estimated coefficients of mode choice determinants were compared between each alternative mode (i.e., high-occupancy vehicle, public transit, and non-motorized transport) with single-occupant vehicles. The model results revealed the significant influence of neighborhood and land use variables on the usage of alternative modes among immigrants. Incorporating these indicators into the demand forecasting process will provide a better understanding of the diverse travel patterns for the unique composition of population groups in Florida.
Resumo:
Many modern applications fall into the category of "large-scale" statistical problems, in which both the number of observations n and the number of features or parameters p may be large. Many existing methods focus on point estimation, despite the continued relevance of uncertainty quantification in the sciences, where the number of parameters to estimate often exceeds the sample size, despite huge increases in the value of n typically seen in many fields. Thus, the tendency in some areas of industry to dispense with traditional statistical analysis on the basis that "n=all" is of little relevance outside of certain narrow applications. The main result of the Big Data revolution in most fields has instead been to make computation much harder without reducing the importance of uncertainty quantification. Bayesian methods excel at uncertainty quantification, but often scale poorly relative to alternatives. This conflict between the statistical advantages of Bayesian procedures and their substantial computational disadvantages is perhaps the greatest challenge facing modern Bayesian statistics, and is the primary motivation for the work presented here.
Two general strategies for scaling Bayesian inference are considered. The first is the development of methods that lend themselves to faster computation, and the second is design and characterization of computational algorithms that scale better in n or p. In the first instance, the focus is on joint inference outside of the standard problem of multivariate continuous data that has been a major focus of previous theoretical work in this area. In the second area, we pursue strategies for improving the speed of Markov chain Monte Carlo algorithms, and characterizing their performance in large-scale settings. Throughout, the focus is on rigorous theoretical evaluation combined with empirical demonstrations of performance and concordance with the theory.
One topic we consider is modeling the joint distribution of multivariate categorical data, often summarized in a contingency table. Contingency table analysis routinely relies on log-linear models, with latent structure analysis providing a common alternative. Latent structure models lead to a reduced rank tensor factorization of the probability mass function for multivariate categorical data, while log-linear models achieve dimensionality reduction through sparsity. Little is known about the relationship between these notions of dimensionality reduction in the two paradigms. In Chapter 2, we derive several results relating the support of a log-linear model to nonnegative ranks of the associated probability tensor. Motivated by these findings, we propose a new collapsed Tucker class of tensor decompositions, which bridge existing PARAFAC and Tucker decompositions, providing a more flexible framework for parsimoniously characterizing multivariate categorical data. Taking a Bayesian approach to inference, we illustrate empirical advantages of the new decompositions.
Latent class models for the joint distribution of multivariate categorical, such as the PARAFAC decomposition, data play an important role in the analysis of population structure. In this context, the number of latent classes is interpreted as the number of genetically distinct subpopulations of an organism, an important factor in the analysis of evolutionary processes and conservation status. Existing methods focus on point estimates of the number of subpopulations, and lack robust uncertainty quantification. Moreover, whether the number of latent classes in these models is even an identified parameter is an open question. In Chapter 3, we show that when the model is properly specified, the correct number of subpopulations can be recovered almost surely. We then propose an alternative method for estimating the number of latent subpopulations that provides good quantification of uncertainty, and provide a simple procedure for verifying that the proposed method is consistent for the number of subpopulations. The performance of the model in estimating the number of subpopulations and other common population structure inference problems is assessed in simulations and a real data application.
In contingency table analysis, sparse data is frequently encountered for even modest numbers of variables, resulting in non-existence of maximum likelihood estimates. A common solution is to obtain regularized estimates of the parameters of a log-linear model. Bayesian methods provide a coherent approach to regularization, but are often computationally intensive. Conjugate priors ease computational demands, but the conjugate Diaconis--Ylvisaker priors for the parameters of log-linear models do not give rise to closed form credible regions, complicating posterior inference. In Chapter 4 we derive the optimal Gaussian approximation to the posterior for log-linear models with Diaconis--Ylvisaker priors, and provide convergence rate and finite-sample bounds for the Kullback-Leibler divergence between the exact posterior and the optimal Gaussian approximation. We demonstrate empirically in simulations and a real data application that the approximation is highly accurate, even in relatively small samples. The proposed approximation provides a computationally scalable and principled approach to regularized estimation and approximate Bayesian inference for log-linear models.
Another challenging and somewhat non-standard joint modeling problem is inference on tail dependence in stochastic processes. In applications where extreme dependence is of interest, data are almost always time-indexed. Existing methods for inference and modeling in this setting often cluster extreme events or choose window sizes with the goal of preserving temporal information. In Chapter 5, we propose an alternative paradigm for inference on tail dependence in stochastic processes with arbitrary temporal dependence structure in the extremes, based on the idea that the information on strength of tail dependence and the temporal structure in this dependence are both encoded in waiting times between exceedances of high thresholds. We construct a class of time-indexed stochastic processes with tail dependence obtained by endowing the support points in de Haan's spectral representation of max-stable processes with velocities and lifetimes. We extend Smith's model to these max-stable velocity processes and obtain the distribution of waiting times between extreme events at multiple locations. Motivated by this result, a new definition of tail dependence is proposed that is a function of the distribution of waiting times between threshold exceedances, and an inferential framework is constructed for estimating the strength of extremal dependence and quantifying uncertainty in this paradigm. The method is applied to climatological, financial, and electrophysiology data.
The remainder of this thesis focuses on posterior computation by Markov chain Monte Carlo. The Markov Chain Monte Carlo method is the dominant paradigm for posterior computation in Bayesian analysis. It has long been common to control computation time by making approximations to the Markov transition kernel. Comparatively little attention has been paid to convergence and estimation error in these approximating Markov Chains. In Chapter 6, we propose a framework for assessing when to use approximations in MCMC algorithms, and how much error in the transition kernel should be tolerated to obtain optimal estimation performance with respect to a specified loss function and computational budget. The results require only ergodicity of the exact kernel and control of the kernel approximation accuracy. The theoretical framework is applied to approximations based on random subsets of data, low-rank approximations of Gaussian processes, and a novel approximating Markov chain for discrete mixture models.
Data augmentation Gibbs samplers are arguably the most popular class of algorithm for approximately sampling from the posterior distribution for the parameters of generalized linear models. The truncated Normal and Polya-Gamma data augmentation samplers are standard examples for probit and logit links, respectively. Motivated by an important problem in quantitative advertising, in Chapter 7 we consider the application of these algorithms to modeling rare events. We show that when the sample size is large but the observed number of successes is small, these data augmentation samplers mix very slowly, with a spectral gap that converges to zero at a rate at least proportional to the reciprocal of the square root of the sample size up to a log factor. In simulation studies, moderate sample sizes result in high autocorrelations and small effective sample sizes. Similar empirical results are observed for related data augmentation samplers for multinomial logit and probit models. When applied to a real quantitative advertising dataset, the data augmentation samplers mix very poorly. Conversely, Hamiltonian Monte Carlo and a type of independence chain Metropolis algorithm show good mixing on the same dataset.
Resumo:
This paper presents a study of the effects of alcohol consumption on household income in Ireland using the Slán National Health and Lifestyle Survey 2007 dataset, accounting for endogeneity and selection bias. Drinkers are categorised into one of four categories based on the recommended weekly drinking levels by the Irish Health Promotion Unit; those who never drank, non-drinkers, moderate and heavy drinkers. A multinomial logit OLS Two Step Estimate is used to explain individual's choice of drinking status and to correct for selection bias which would result in the selection into a particular category of drinking being endogenous. Endogeneity which may arise through the simultaneity of drinking status and income either due to the reverse causation between the two variables, income affecting alcohol consumption or alcohol consumption affecting income, or due to unobserved heterogeneity, is addressed. This paper finds that the household income of drinkers is higher than that of non-drinkers and of those who never drank. There is very little difference between the household income of moderate and heavy drinkers, with heavy drinkers earning slightly more. Weekly household income for those who never drank is €454.20, non-drinkers is €506.26, compared with €683.36 per week for moderate drinkers and €694.18 for heavy drinkers.
Resumo:
In this paper, the start-up process is split conceptually into four stages: considering entrepreneurship, intending to start a new business in the next 3 years, nascent entrepreneurship and owning-managing a newly established business. We investigate the determinants of all of these jointly, using a multinomial logit model; it allows for the effects of resources and capabilities to vary across these stages. We employ the Global Entrepreneurship Monitor database for the years 2006–2009, containing 8269 usable observations from respondents drawn from the Lower Layer Super Output Areas in the East Midlands (UK) so that individual observations are linked to space. Our results show that the role of education, experience, and availability of ‘entrepreneurial capital’ in the local neighbourhood varies along the different stages of the entrepreneurial process. In the early stages, the negative (opportunity cost) effect of resources endowment dominates, yet it tends to reverse in the advanced stages, where the positive effect of resources becomes stronger.
Resumo:
Solving multi-stage oligopoly models by backward induction can easily become a com- plex task when rms are multi-product and demands are derived from a nested logit frame- work. This paper shows that under the assumption that within-segment rm shares are equal across segments, the analytical expression for equilibrium pro ts can be substantially simpli ed. The size of the error arising when this condition does not hold perfectly is also computed. Through numerical examples, it is shown that the error is rather small in general. Therefore, using this assumption allows to gain analytical tractability in a class of models that has been used to approach relevant policy questions, such as for example rm entry in an industry or the relation between competition and location. The simplifying approach proposed in this paper is aimed at helping improving these type of models for reaching more accurate recommendations.
Resumo:
Infections of the central nervous systems (CNS) present a diagnostic problem for which an accurate laboratory diagnosis is essential. Invasive practices, such as cerebral biopsy, have been replaced by obtaining a polymerase chain reaction (PCR) diagnosis using cerebral spinal fluid (CSF) as a reference method. Tests on DNA extracted from plasma are noninvasive, thus avoiding all of the collateral effects and patient risks associated with CSF collection. This study aimed to determine whether plasma can replace CSF in nested PCR analysis for the detection of CNS human herpesvirus (HHV) diseases by analysing the proportion of patients whose CSF nested PCR results were positive for CNS HHV who also had the same organism identified by plasma nested PCR. In this study, CSF DNA was used as the gold standard, and nested PCR was performed on both types of samples. Fifty-two patients with symptoms of nervous system infection were submitted to CSF and blood collection. For the eight HHV, one positive DNA result-in plasma and/or CSF nested PCR-was considered an active HHV infection, whereas the occurrence of two or more HHVs in the same sample was considered a coinfection. HHV infections were positively detected in 27/52 (51.9%) of the CSF and in 32/52 (61.5%) of the plasma, difference not significant, thus nested PCR can be performed on plasma instead of CSF. In conclusion, this findings suggest that plasma as a useful material for the diagnosis of cases where there is any difficulty to perform a CSF puncture.
Resumo:
Bovine coronavirus (BCoV) is a member of the group 2 of the Coronavirus (Nidovirales: Coronaviridae) and the causative agent of enteritis in both calves and adult bovine, as well as respiratory disease in calves. The present study aimed to develop a semi-nested RT-PCR for the detection of BCoV based on representative up-to-date sequences of the nucleocapsid gene, a conserved region of coronavirus genome. Three primers were designed, the first round with a 463bp and the second (semi-nested) with a 306bp predicted fragment. The analytical sensitivity was determined by 10-fold serial dilutions of the BCoV Kakegawa strain (HA titre: 256) in DEPC treated ultra-pure water, in fetal bovine serum (FBS) and in a BCoV-free fecal suspension, when positive results were found up to the 10-2, 10-3 and 10-7 dilutions, respectively, which suggests that the total amount of RNA in the sample influence the precipitation of pellets by the method of extraction used. When fecal samples was used, a large quantity of total RNA serves as carrier of BCoV RNA, demonstrating a high analytical sensitivity and lack of possible substances inhibiting the PCR. The final semi-nested RT-PCR protocol was applied to 25 fecal samples from adult cows, previously tested by a nested RT-PCR RdRp used as a reference test, resulting in 20 and 17 positives for the first and second tests, respectively, and a substantial agreement was found by kappa statistics (0.694). The high sensitivity and specificity of the new proposed method and the fact that primers were designed based on current BCoV sequences give basis to a more accurate diagnosis of BCoV-caused diseases, as well as to further insights on protocols for the detection of other Coronavirus representatives of both Animal and Public Health importance.
Resumo:
Background: Areas that are endemic for malaria are also highly endemic for hepatitis B virus (HBV) infection. Nevertheless, it is unknown whether HBV infection modifies the clinical presentation of malaria. This study aimed to address this question. Methodology and Findings: An observational study of 636 individuals was performed in Rondonia, western Amazon, Brazil between 2006 and 2007. Active and passive case detections identified Plasmodium infection by field microscopy and nested Polymerase Chain Reaction (PCR). HBV infections were identified by serology and confirmed by real-time PCR. Epidemiological information and plasma cytokine profiles were studied. The data were analyzed using adjusted multinomial logistic regression. Plasmodium-infected individuals with active HBV infection were more likely to be asymptomatic (OR: 120.13, P < 0.0001), present with lower levels of parasitemia and demonstrate a decreased inflammatory cytokine profile. Nevertheless, co-infected individuals presented higher HBV viremia. Plasmodium parasitemia inversely correlated with plasma HBV DNA levels (r=-0.6; P=0.0003). Conclusion: HBV infection diminishes the intensity of malaria infection in individuals from this endemic area. This effect seems related to cytokine balance and control of inflammatory responses. These findings add important insights to the understanding of the factors affecting the clinical outcomes of malaria in endemic regions.
Resumo:
The use of improved microbiological procedures associated with molecular techniques has increased the identification of Candida bloodstream infections, even if the isolation of more than one species by culture methods remains uncommon. We report the cases of two children presenting with severe gastrointestinal disorders and other risk factors that contribute to Candida infections. In the first patient, C. albicans DNA was initially detected by a nested-amplification and C. tropicalis was found later during hospitalization, while blood cultures were persistently negative. In the second child, there was amplification of C. albicans and C. glabrata DNA in the same samples, but blood cultures yielded only C. albicans. Both patients received antifungal therapy but had unfavorable outcomes. These two cases illustrate that PCR was more successful than culture methods in detecting Candida in the bloodstream of high risk children, and was also able to detect the presence of more than one species in the same patient that might impact therapy when the fungi are resistant to azole compounds.
Resumo:
Neonatal calf diarrhea is a multi-etiology syndrome of cattle and direct detection of the two major agents of the syndrome, group A rotavirus and Bovine coronavirus (BCoV) is hampered by their fastidious growth in cell culture. This study aimed at developing a multiplex semi-nested RT-PCR for simultaneous detection of BCoV (N gene) and group A rotavirus (VP1 gene) with the addition of an internal control (mRNA ND5). The assay was tested in 75 bovine feces samples tested previously for rotavirus using PAGE and for BCoV using nested RT-PCR targeted to RdRp gene. Agreement with reference tests was optimal for BCoV (kappa = 0.833) and substantial for rotavirus detection (kappa = 0.648). the internal control, ND5 mRNA, was detected successfully in all reactions. Results demonstrated that this multiplex semi-nested RT-PCR was effective in the detection of BCoV and rotavirus, with high sensitivity and specificity for simultaneous detection of both viruses at a lower cost, providing an important tool for studies on the etiology of diarrhea in cattle. (C) 2010 Elsevier B.V. All rights reserved.
Resumo:
A standardised nested-PCR method that amplifies a region of the glycoprotein E gene of avian infectious laryngotracheitis virus (ILTV) has been developed for the diagnosis of infection by Gallid herpesvirus 1. The two sets of primers employed produced the expected ampIification products of 524bp(externa I primers) and 219bp (internal primers) in the presence of ILTV DNA, whereas no Such amplicons were obtained with other avian respiratory pathogens or with DNA extracted from the cells of uninfected chickens. The identity of the 219bp amplified product was con firmed by DNA sequencing. The standardised nested-PCR method detected ILTV DNA from trachea, lung, conjunctiva and trigeminal ganglia samples from flocks of birds with and without clinical signs. and showed hi.-h sensitivity (95.4%) and specificity (93.1%) when compared with the reference test involving virus isolation in specific-pathogen-free chicken embryos. The standardised nested-PCR method described may be used to detect clinical and latent ILTV infections, and will be of significant value for both diagnostic and epidemiological Studies. (c) 2008 Elsevier B.V. All rights reserved.
Resumo:
Despite its widespread use, the Coale-Demeny model life table system does not capture the extensive variation in age-specific mortality patterns observed in contemporary populations, particularly those of the countries of Eastern Europe and populations affected by HIV/AIDS. Although relational mortality models, such as the Brass logit system, can identify these variations, these models show systematic bias in their predictive ability as mortality levels depart from the standard. We propose a modification of the two-parameter Brass relational model. The modified model incorporates two additional age-specific correction factors (gamma(x), and theta(x)) based on mortality levels among children and adults, relative to the standard. Tests of predictive validity show deviations in age-specific mortality rates predicted by the proposed system to be 30-50 per cent lower than those predicted by the Coale-Demeny system and 15-40 per cent lower than those predicted using the original Brass system. The modified logit system is a two-parameter system, parameterized using values of l(5) and l(60).