959 resultados para Statistics in sensory analysis
Resumo:
In this paper, we investigate the impact of linear mode coupling on the efficiency of intermodal four-wave mixing and on the group delay statistics in few-mode fibres. The investigation will include not only the weak or strong linear coupling regimes, but also the transition region between them, the intermediate coupling regime. This analysis will allow to assess the level of coupling strength require to suppress the nonlinear distortion in a few-mode fibre below the level of distortion for single-mode propagation without mode coupling.
Resumo:
Many modern applications fall into the category of "large-scale" statistical problems, in which both the number of observations n and the number of features or parameters p may be large. Many existing methods focus on point estimation, despite the continued relevance of uncertainty quantification in the sciences, where the number of parameters to estimate often exceeds the sample size, despite huge increases in the value of n typically seen in many fields. Thus, the tendency in some areas of industry to dispense with traditional statistical analysis on the basis that "n=all" is of little relevance outside of certain narrow applications. The main result of the Big Data revolution in most fields has instead been to make computation much harder without reducing the importance of uncertainty quantification. Bayesian methods excel at uncertainty quantification, but often scale poorly relative to alternatives. This conflict between the statistical advantages of Bayesian procedures and their substantial computational disadvantages is perhaps the greatest challenge facing modern Bayesian statistics, and is the primary motivation for the work presented here.
Two general strategies for scaling Bayesian inference are considered. The first is the development of methods that lend themselves to faster computation, and the second is design and characterization of computational algorithms that scale better in n or p. In the first instance, the focus is on joint inference outside of the standard problem of multivariate continuous data that has been a major focus of previous theoretical work in this area. In the second area, we pursue strategies for improving the speed of Markov chain Monte Carlo algorithms, and characterizing their performance in large-scale settings. Throughout, the focus is on rigorous theoretical evaluation combined with empirical demonstrations of performance and concordance with the theory.
One topic we consider is modeling the joint distribution of multivariate categorical data, often summarized in a contingency table. Contingency table analysis routinely relies on log-linear models, with latent structure analysis providing a common alternative. Latent structure models lead to a reduced rank tensor factorization of the probability mass function for multivariate categorical data, while log-linear models achieve dimensionality reduction through sparsity. Little is known about the relationship between these notions of dimensionality reduction in the two paradigms. In Chapter 2, we derive several results relating the support of a log-linear model to nonnegative ranks of the associated probability tensor. Motivated by these findings, we propose a new collapsed Tucker class of tensor decompositions, which bridge existing PARAFAC and Tucker decompositions, providing a more flexible framework for parsimoniously characterizing multivariate categorical data. Taking a Bayesian approach to inference, we illustrate empirical advantages of the new decompositions.
Latent class models for the joint distribution of multivariate categorical, such as the PARAFAC decomposition, data play an important role in the analysis of population structure. In this context, the number of latent classes is interpreted as the number of genetically distinct subpopulations of an organism, an important factor in the analysis of evolutionary processes and conservation status. Existing methods focus on point estimates of the number of subpopulations, and lack robust uncertainty quantification. Moreover, whether the number of latent classes in these models is even an identified parameter is an open question. In Chapter 3, we show that when the model is properly specified, the correct number of subpopulations can be recovered almost surely. We then propose an alternative method for estimating the number of latent subpopulations that provides good quantification of uncertainty, and provide a simple procedure for verifying that the proposed method is consistent for the number of subpopulations. The performance of the model in estimating the number of subpopulations and other common population structure inference problems is assessed in simulations and a real data application.
In contingency table analysis, sparse data is frequently encountered for even modest numbers of variables, resulting in non-existence of maximum likelihood estimates. A common solution is to obtain regularized estimates of the parameters of a log-linear model. Bayesian methods provide a coherent approach to regularization, but are often computationally intensive. Conjugate priors ease computational demands, but the conjugate Diaconis--Ylvisaker priors for the parameters of log-linear models do not give rise to closed form credible regions, complicating posterior inference. In Chapter 4 we derive the optimal Gaussian approximation to the posterior for log-linear models with Diaconis--Ylvisaker priors, and provide convergence rate and finite-sample bounds for the Kullback-Leibler divergence between the exact posterior and the optimal Gaussian approximation. We demonstrate empirically in simulations and a real data application that the approximation is highly accurate, even in relatively small samples. The proposed approximation provides a computationally scalable and principled approach to regularized estimation and approximate Bayesian inference for log-linear models.
Another challenging and somewhat non-standard joint modeling problem is inference on tail dependence in stochastic processes. In applications where extreme dependence is of interest, data are almost always time-indexed. Existing methods for inference and modeling in this setting often cluster extreme events or choose window sizes with the goal of preserving temporal information. In Chapter 5, we propose an alternative paradigm for inference on tail dependence in stochastic processes with arbitrary temporal dependence structure in the extremes, based on the idea that the information on strength of tail dependence and the temporal structure in this dependence are both encoded in waiting times between exceedances of high thresholds. We construct a class of time-indexed stochastic processes with tail dependence obtained by endowing the support points in de Haan's spectral representation of max-stable processes with velocities and lifetimes. We extend Smith's model to these max-stable velocity processes and obtain the distribution of waiting times between extreme events at multiple locations. Motivated by this result, a new definition of tail dependence is proposed that is a function of the distribution of waiting times between threshold exceedances, and an inferential framework is constructed for estimating the strength of extremal dependence and quantifying uncertainty in this paradigm. The method is applied to climatological, financial, and electrophysiology data.
The remainder of this thesis focuses on posterior computation by Markov chain Monte Carlo. The Markov Chain Monte Carlo method is the dominant paradigm for posterior computation in Bayesian analysis. It has long been common to control computation time by making approximations to the Markov transition kernel. Comparatively little attention has been paid to convergence and estimation error in these approximating Markov Chains. In Chapter 6, we propose a framework for assessing when to use approximations in MCMC algorithms, and how much error in the transition kernel should be tolerated to obtain optimal estimation performance with respect to a specified loss function and computational budget. The results require only ergodicity of the exact kernel and control of the kernel approximation accuracy. The theoretical framework is applied to approximations based on random subsets of data, low-rank approximations of Gaussian processes, and a novel approximating Markov chain for discrete mixture models.
Data augmentation Gibbs samplers are arguably the most popular class of algorithm for approximately sampling from the posterior distribution for the parameters of generalized linear models. The truncated Normal and Polya-Gamma data augmentation samplers are standard examples for probit and logit links, respectively. Motivated by an important problem in quantitative advertising, in Chapter 7 we consider the application of these algorithms to modeling rare events. We show that when the sample size is large but the observed number of successes is small, these data augmentation samplers mix very slowly, with a spectral gap that converges to zero at a rate at least proportional to the reciprocal of the square root of the sample size up to a log factor. In simulation studies, moderate sample sizes result in high autocorrelations and small effective sample sizes. Similar empirical results are observed for related data augmentation samplers for multinomial logit and probit models. When applied to a real quantitative advertising dataset, the data augmentation samplers mix very poorly. Conversely, Hamiltonian Monte Carlo and a type of independence chain Metropolis algorithm show good mixing on the same dataset.
Resumo:
The advances in three related areas of state-space modeling, sequential Bayesian learning, and decision analysis are addressed, with the statistical challenges of scalability and associated dynamic sparsity. The key theme that ties the three areas is Bayesian model emulation: solving challenging analysis/computational problems using creative model emulators. This idea defines theoretical and applied advances in non-linear, non-Gaussian state-space modeling, dynamic sparsity, decision analysis and statistical computation, across linked contexts of multivariate time series and dynamic networks studies. Examples and applications in financial time series and portfolio analysis, macroeconomics and internet studies from computational advertising demonstrate the utility of the core methodological innovations.
Chapter 1 summarizes the three areas/problems and the key idea of emulating in those areas. Chapter 2 discusses the sequential analysis of latent threshold models with use of emulating models that allows for analytical filtering to enhance the efficiency of posterior sampling. Chapter 3 examines the emulator model in decision analysis, or the synthetic model, that is equivalent to the loss function in the original minimization problem, and shows its performance in the context of sequential portfolio optimization. Chapter 4 describes the method for modeling the steaming data of counts observed on a large network that relies on emulating the whole, dependent network model by independent, conjugate sub-models customized to each set of flow. Chapter 5 reviews those advances and makes the concluding remarks.
Outcomes and Predictors of Mortality in Neurosurgical Patients at Mbarara Regional Referral Hospital
Resumo:
Background:
Knowing the scope of neurosurgical disease at Mbarara Hospital is critical for infrastructure planning, education and training. In this study, we aim to evaluate the neurosurgical outcomes and identify predictors of mortality in order to potentiate platforms for more effective interventions and inform future research efforts at Mbarara Hospital.
Methods:
This is retrospective chart review including patients of all ages with a neurosurgical disease or injury presenting to Mbarara Regional Referral Hospital (MRRH) between January 2012 to September 2015. Descriptive statistics were presented. A univariate analysis was used to obtain the odds ratios of mortality and 95% confidence intervals. Predictors of mortality were determined using multivariate logistic regression model.
Results:
A total of 1876 charts were reviewed. Of these, 1854 (had complete data and were?) were included in the analysis. The overall mortality rate was 12.75%; the mortality rates among all persons who underwent a neurosurgical procedure was 9.72%, and was 13.68% among those who did not undergo a neurosurgical procedure. Over 50% of patients were between 19 and 40 years old and the majority of were males (76.10%). The overall median length of stay was 5 days. Of all neurosurgical admissions, 87% were trauma patients. In comparison to mild head injury, closed head injury and intracranial hematoma patients were 5 (95% CI: 3.77, 8.26) and 2.5 times (95% CI: 1.64,3.98) more likely to die respectively. Procedure and diagnostic imaging were independent negative predictors of mortality (P <0.05). While age, ICU admission, admission GCS were positive predictors of mortality (P <0.05).
Conclusions:
The majority of hospital admissions were TBI patients, with RTIs being the most common mechanism of injury. Age, ICU admission, admission GCS, diagnostic imaging and undergoing surgery were independent predictors of mortality. Going forward, further exploration of patient characteristics is necessary to fully describe mortality outcomes and implement resource appropriate interventions that ultimately improve morbidity and mortality.
Resumo:
Research on women’s employment has proliferated over recent decades, often under a perspective that conceptualizes female labour market activity as independent of male presences and absences in the productive and reproductive spheres. In the face of these approaches, the article argues the need to focus on the couple as the unit of analysis of work-life articulation. After referring to the main theoretical arguments that, from a gender perspective within labour studies, have pointed out the relevance of placing the household as the central space for the analysis of the sexual division of labour, the article reviews different empirical contributions that have incorporated such perspective in the international literature. Next, the state of the art in the Spanish literature is presented, before arguing the desirability of applying such framework of analysis to the study of employment and care work in Spanish households, which are at present undergoing major transformations.
Resumo:
The article presents a study of a CEFR B2-level reading subtest that is part of the Slovenian national secondary school leaving examination in English as a foreign language, and compares the test-taker actual performance (objective difficulty) with the test-taker and expert perceptions of item difficulty (subjective difficulty). The study also analyses the test-takers’ comments on item difficulty obtained from a while-reading questionnaire. The results are discussed in the framework of the existing research in the fields of (the assessment of) reading comprehension, and are addressed with regard to their implications for item-writing, FL teaching and curriculum development.
Resumo:
While violence against children is a common occurrence only a minority of incidents come to the attention of the authorities. Low reporting rates notwithstanding, official data such as child protection referrals and recorded crime statistics provide valuable information on the numbers of children experiencing harm which come to the attention of professionals in any given year. In the UK, there has been a strong tendency to focus on child protection statistics while children as victims of crime remain largely invisible in annual crime reports and associated compendia. This is despite the implementation of a raft of policies aimed at improving the system response to victims and witnesses of crime across the UK. This paper demonstrates the utility of a more detailed analysis of crime statistics in providing information on the patterns of crime against children and examining case outcomes. Based on data made available by the Police Service for Northern Ireland, it highlights how violent crime differentially impacts on older children and how detection rates vary depending on case characteristics. It makes an argument for developing recorded crime practice to make child victims of crime more visible and to facilitate assessment of the effectiveness of current initiatives and policy developments. Copyright © 2013 John Wiley & Sons, Ltd.
Resumo:
Energy efficiency improvement has been a key objective of China’s long-term energy policy. In this paper, we derive single-factor technical energy efficiency (abbreviated as energy efficiency) in China from multi-factor efficiency estimated by means of a translog production function and a stochastic frontier model on the basis of panel data on 29 Chinese provinces over the period 2003–2011. We find that average energy efficiency has been increasing over the research period and that the provinces with the highest energy efficiency are at the east coast and the ones with the lowest in the west, with an intermediate corridor in between. In the analysis of the determinants of energy efficiency by means of a spatial Durbin error model both factors in the own province and in first-order neighboring provinces are considered. Per capita income in the own province has a positive effect. Furthermore, foreign direct investment and population density in the own province and in neighboring provinces have positive effects, whereas the share of state-owned enterprises in Gross Provincial Product in the own province and in neighboring provinces has negative effects. From the analysis it follows that inflow of foreign direct investment and reform of state-owned enterprises are important policy handles.
Resumo:
We study a minimal integrate-and-fire based model of a "ghostbursting" neuron under periodic stimulation. These neurons are involved in sensory processing in weakly electric fish. There exist regions in parameter space in which the model neuron is mode-locked to the stimulation. We analyse this locked behavior and examine the bifurcations that define the boundaries of these regions. Due to the discontinuous nature of the flow, some of these bifurcations are nonsmooth. This exact analysis is in excellent agreement with numerical simulations, and can be used to understand the response of such a model neuron to biologically realistic input.
Resumo:
The ability to predict the properties of magnetic materials in a device is essential to ensuring the correct operation and optimization of the design as well as the device behavior over a wide range of input frequencies. Typically, development and simulation of wide-bandwidth models requires detailed, physics-based simulations that utilize significant computational resources. Balancing the trade-offs between model computational overhead and accuracy can be cumbersome, especially when the nonlinear effects of saturation and hysteresis are included in the model. This study focuses on the development of a system for analyzing magnetic devices in cases where model accuracy and computational intensity must be carefully and easily balanced by the engineer. A method for adjusting model complexity and corresponding level of detail while incorporating the nonlinear effects of hysteresis is presented that builds upon recent work in loss analysis and magnetic equivalent circuit (MEC) modeling. The approach utilizes MEC models in conjunction with linearization and model-order reduction techniques to process magnetic devices based on geometry and core type. The validity of steady-state permeability approximations is also discussed.
Resumo:
A lean muscle line (L) and a fat muscle line (F) of rainbow trout were established (Quillet et al., 2005) by a two-way selection for muscle lipid content performed on pan-size rainbow trout using a non-destructive measurement of muscle lipid content (Distell Fish Fat Meter®). The aim of the present study was to evaluate the consequences of this selective breeding on flesh quality of pan size (290 g) diploid and triploid trout after three generations of selection. Instrumental evaluations of fillet color and pH measurement were performed at slaughter. Flesh color, pH, dry matter content and mechanical resistance were measured at 48 h and 96 h postmortem on raw and cooked flesh, respectively. A sensorial profile analysis was performed on cooked fillets. Fillets from the selected fatty muscle line (F) had a higher dry matter content and were more colorful for both raw and cooked fillets. Mechanical evaluation indicated a tendency of raw flesh from F fish to be less firm, but this was not confirmed after cooking, neither instrumentally or by sensory analysis. The sensory analysis revealed higher fat loss, higher intensity of flavor of cooked potato, higher exudation, higher moisture content and a more fatty film left on the tongue for flesh from F fish. Triploid fish had mechanically softer raw and cooked fillets, but the difference was not perceived by the sensorial panel. The sensorial evaluation also revealed a lower global intensity of odor, more exudation and a higher moisture content in the fillets from triploid fish. These differences in quality parameters among groups of fish were associated with larger white muscle fibers in F fish and in triploid fish. The data provide additional information about the relationship between muscle fat content, muscle cellularity and flesh quality.
Resumo:
Capacity analysis using simulation is not a new thing in literature. Most of the development process of UMTS standardization have used simulation tools; however, we thing that the use of GIS planning tools and matrix manipulation capacity of MATLAB can show us different scenarios and make a more realistic analysis. Some work is been doing in COST 273 in order to have more realistic scenarios for UMTS planning. Our work initially was centered in uplink analysis, but we are now working in downlink analysis, specifically in two areas: capacity in number of users for RT and NRT services, and Node B power. In this work we will show results for up-link capacity and some results for downlink capacity and BS power consumption.
Resumo:
The graph Laplacian operator is widely studied in spectral graph theory largely due to its importance in modern data analysis. Recently, the Fourier transform and other time-frequency operators have been defined on graphs using Laplacian eigenvalues and eigenvectors. We extend these results and prove that the translation operator to the i’th node is invertible if and only if all eigenvectors are nonzero on the i’th node. Because of this dependency on the support of eigenvectors we study the characteristic set of Laplacian eigenvectors. We prove that the Fiedler vector of a planar graph cannot vanish on large neighborhoods and then explicitly construct a family of non-planar graphs that do exhibit this property. We then prove original results in modern analysis on graphs. We extend results on spectral graph wavelets to create vertex-dyanamic spectral graph wavelets whose support depends on both scale and translation parameters. We prove that Spielman’s Twice-Ramanujan graph sparsifying algorithm cannot outperform his conjectured optimal sparsification constant. Finally, we present numerical results on graph conditioning, in which edges of a graph are rescaled to best approximate the complete graph and reduce average commute time.
Resumo:
Gating of sensory (e.g. auditory) information has been demonstrated as a reduction in the auditory-evoked potential responses recorded in the brain of both normal animals and human subjects. Auditory gating is perturbed in schizophrenic patients and pharmacologically by drugs such as amphetamine, phencyclidine or ketamine, which precipitate schizophrenic-like symptoms in normal subjects. The neurobiological basis underlying this sensory gating can be investigated using local field potential recordings from single electrodes. In this paper we use such technology to investigate the role of cannabinoids in sensory gating. Cannabinoids represent a fundamentally new class of retrograde messengers which are released postsynaptically and bind to presynaptic receptors. In this way they allow fine-tuning of neuronal response, and in particular can lead to so-called depolarization-induced suppression of inhibition (DSI). Our experimental results show that application of the exogenous cannabinoid WIN55, 212-2 can abolish sensory gating as measured by the amplitude of local field responses in rat hippocampal region CA3. Importantly we develop a simple firing rate population model of CA3 and show that gating is heavily dependent upon the presence of a slow inhibitory (GABAB) pathway. Moreover, a simple phenomenological model of cannabinoid dynamics underlying DSI is shown to abolish gating in a manner consistent with our experimental findings.