929 resultados para likelihood ratio test


Relevância:

80.00% 80.00%

Publicador:

Resumo:

The quick detection of an abrupt unknown change in the conditional distribution of a dependent stochastic process has numerous applications. In this paper, we pose a minimax robust quickest change detection problem for cases where there is uncertainty about the post-change conditional distribution. Our minimax robust formulation is based on the popular Lorden criteria of optimal quickest change detection. Under a condition on the set of possible post-change distributions, we show that the widely known cumulative sum (CUSUM) rule is asymptotically minimax robust under our Lorden minimax robust formulation as a false alarm constraint becomes more strict. We also establish general asymptotic bounds on the detection delay of misspecified CUSUM rules (i.e. CUSUM rules that are designed with post- change distributions that differ from those of the observed sequence). We exploit these bounds to compare the delay performance of asymptotically minimax robust, asymptotically optimal, and other misspecified CUSUM rules. In simulation examples, we illustrate that asymptotically minimax robust CUSUM rules can provide better detection delay performance at greatly reduced computation effort compared to competing generalised likelihood ratio procedures.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

In this paper we propose a novel scheme for carrying out speaker diarization in an iterative manner. We aim to show that the information obtained through the first pass of speaker diarization can be reused to refine and improve the original diarization results. We call this technique speaker rediarization and demonstrate the practical application of our rediarization algorithm using a large archive of two-speaker telephone conversation recordings. We use the NIST 2008 SRE summed telephone corpora for evaluating our speaker rediarization system. This corpus contains recurring speaker identities across independent recording sessions that need to be linked across the entire corpus. We show that our speaker rediarization scheme can take advantage of inter-session speaker information, linked in the initial diarization pass, to achieve a 30% relative improvement over the original diarization error rate (DER) after only two iterations of rediarization.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

In the analysis of tagging data, it has been found that the least-squares method, based on the increment function known as the Fabens method, produces biased estimates because individual variability in growth is not allowed for. This paper modifies the Fabens method to account for individual variability in the length asymptote. Significance tests using t-statistics or log-likelihood ratio statistics may be applied to show the level of individual variability. Simulation results indicate that the modified method reduces the biases in the estimates to negligible proportions. Tagging data from tiger prawns (Penaeus esculentus and Penaeus semisulcatus) and rock lobster (Panulirus ornatus) are analysed as an illustration.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Sepsis is associated with a systemic inflammatory response. It is characterised by an early proinflammatory response and followed by a state of immunosuppression. In order to improve the outcome of patients with infection and sepsis, novel therapies that influence the systemic inflammatory response are being developed and utilised. Thus, an accurate and early diagnosis of infection and evaluation of immune state are crucial. In this thesis, various markers of systemic inflammation were studied with respect to enhancing the diagnostics of infection and of predicting outcome in patients with suspected community-acquired infection. A total of 1092 acutely ill patients admitted to a university hospital medical emergency department were evaluated, and 531 patients with a suspicion of community-acquired infection were included for the analysis. Markers of systemic inflammation were determined from a blood sample obtained simultaneously with a blood culture sample on admission to hospital. Levels of phagocyte CD11b/CD18 and CD14 expression were measured by whole blood flow cytometry. Concentrations of soluble CD14, interleukin (IL)-8, and soluble IL-2 receptor α (sIL-2Rα) were determined by ELISA, those of sIL-2R, IL-6, and IL-8 by a chemiluminescent immunoassay, that of procalcitonin by immunoluminometric assay, and that of C-reactive protein by immunoturbidimetric assay. Clinical data were collected retrospectively from the medical records. No marker of systemic inflammation, neither CRP, PCT, IL-6, IL-8, nor sIL-2R predicted bacteraemia better than did the clinical signs of infection, i.e., the presence of infectious focus or fever or both. IL-6 and PCT had the highest positive likelihood ratios to identify patients with hidden community-acquired infection. However, the use of a single marker failed to detect all patients with infection. A combination of markers including a fast-responding reactant (CD11b expression), a later-peaking reactant (CRP), and a reactant originating from inflamed tissues (IL-8) detected all patients with infection. The majority of patients (86.5%) with possible but not verified infection showed levels exceeding at least one cut-off limit of combination, supporting the view that infection was the cause of their acute illness. The 28-day mortality of patients with community-acquired infection was low (3.4%). On admission to hospital, the low expression of cell-associated lipopolysaccharide receptor CD14 (mCD14) was predictive for 28-day mortality. In the patients with severe forms of community-acquired infection, namely pneumonia and sepsis, high levels of soluble CD14 alone did not predict mortality, but a high sCD14 level measured simultaneously with a low mCD14 raised the possibility of poor prognosis. In conclusion, to further enhance the diagnostics of hidden community-acquired infection, a combination of inflammatory markers is useful; 28-day mortality is associated with low levels of mCD14 expression at an early phase of the disease.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

This thesis studies the informational efficiency of the European Union emission allowance (EUA) market. In an efficient market, the market price is unpredictable and profits above average are impossible in the long run. The main research problem is does the EUA price follow a random walk. The method is an econometric analysis of the price series, which includes an autocorrelation coefficient test and a variance ratio test. The results reveal that the price series is autocorrelated and therefore a nonrandom walk. In order to find out the extent of predictability, the price series is modelled with an autoregressive model. The conclusion is that the EUA price is autocorrelated only to a small degree and that the predictability cannot be used to make extra profits. The EUA market is therefore considered informationally efficient, although the price series does not fulfill the requirements of a random walk. A market review supports the conclusion, but it is clear that the maturing of the market is still in process.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Merton's model views equity as a call option on the asset of the firm. Thus the asset is partially observed through the equity. Then using nonlinear filtering an explicit expression for likelihood ratio for underlying parameters in terms of the nonlinear filter is obtained. As the evolution of the filter itself depends on the parameters in question, this does not permit direct maximum likelihood estimation, but does pave the way for the `Expectation-Maximization' method for estimating parameters. (C) 2010 Elsevier B.V. All rights reserved.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Low density parity-check (LDPC) codes are a class of linear block codes that are decoded by running belief propagation (BP) algorithm or log-likelihood ratio belief propagation (LLR-BP) over the factor graph of the code. One of the disadvantages of LDPC codes is the onset of an error floor at high values of signal to noise ratio caused by trapping sets. In this paper, we propose a two stage decoder to deal with different types of trapping sets. Oscillating trapping sets are taken care by the first stage of the decoder and the elementary trapping sets are handled by the second stage of the decoder. Simulation results on the regular PEG (504,252,3,6) code and the irregular PEG (1024,518,15,8) code shows that the proposed two stage decoder performs significantly better than the standard decoder.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

We consider cooperative spectrum sensing for cognitive radios. We develop an energy efficient detector with low detection delay using sequential hypothesis testing. Sequential Probability Ratio Test (SPRT) is used at both the local nodes and the fusion center. We also analyse the performance of this algorithm and compare with the simulations. Modelling uncertainties in the distribution parameters are considered. Slow fading with and without perfect channel state information at the cognitive radios is taken into account.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

This paper considers cooperative spectrum sensing in Cognitive Radios. In our previous work we have developed DualSPRT, a distributed algorithm for cooperative spectrum sensing using Sequential Probability Ratio Test (SPRT) at the Cognitive Radios as well as at the fusion center. This algorithm works well, but is not optimal. In this paper we propose an improved algorithm- SPRT-CSPRT, which is motivated from Cumulative Sum Procedures (CUSUM). We analyse it theoretically. We also modify this algorithm to handle uncertainties in SNR's and fading.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

We consider cooperative spectrum sensing for cognitive radios. We develop an energy efficient detector with low detection delay using sequential hypothesis testing. Sequential Probability Ratio Test (SPRT) is used at both the local nodes and the fusion center. We also analyse the performance of this algorithm and compare with the simulations. Modelling uncertainties in the distribution parameters are considered. Slow fading with and without perfect channel state information at the cognitive radios is taken into account.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

This paper considers cooperative spectrum sensing algorithms for Cognitive Radios which focus on reducing the number of samples to make a reliable detection. We propose algorithms based on decentralized sequential hypothesis testing in which the Cognitive Radios sequentially collect the observations, make local decisions and send them to the fusion center for further processing to make a final decision on spectrum usage. The reporting channel between the Cognitive Radios and the fusion center is assumed more realistically as a Multiple Access Channel (MAC) with receiver noise. Furthermore the communication for reporting is limited, thereby reducing the communication cost. We start with an algorithm where the fusion center uses an SPRT-like (Sequential Probability Ratio Test) procedure and theoretically analyze its performance. Asymptotically, its performance is close to the optimal centralized test without fusion center noise. We further modify this algorithm to improve its performance at practical operating points. Later we generalize these algorithms to handle uncertainties in SNR and fading. (C) 2014 Elsevier B.V. All rights reserved.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

The Restricted Boltzmann Machines (RBM) can be used either as classifiers or as generative models. The quality of the generative RBM is measured through the average log-likelihood on test data. Due to the high computational complexity of evaluating the partition function, exact calculation of test log-likelihood is very difficult. In recent years some estimation methods are suggested for approximate computation of test log-likelihood. In this paper we present an empirical comparison of the main estimation methods, namely, the AIS algorithm for estimating the partition function, the CSL method for directly estimating the log-likelihood, and the RAISE algorithm that combines these two ideas.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

This thesis studies decision making under uncertainty and how economic agents respond to information. The classic model of subjective expected utility and Bayesian updating is often at odds with empirical and experimental results; people exhibit systematic biases in information processing and often exhibit aversion to ambiguity. The aim of this work is to develop simple models that capture observed biases and study their economic implications.

In the first chapter I present an axiomatic model of cognitive dissonance, in which an agent's response to information explicitly depends upon past actions. I introduce novel behavioral axioms and derive a representation in which beliefs are directionally updated. The agent twists the information and overweights states in which his past actions provide a higher payoff. I then characterize two special cases of the representation. In the first case, the agent distorts the likelihood ratio of two states by a function of the utility values of the previous action in those states. In the second case, the agent's posterior beliefs are a convex combination of the Bayesian belief and the one which maximizes the conditional value of the previous action. Within the second case a unique parameter captures the agent's sensitivity to dissonance, and I characterize a way to compare sensitivity to dissonance between individuals. Lastly, I develop several simple applications and show that cognitive dissonance contributes to the equity premium and price volatility, asymmetric reaction to news, and belief polarization.

The second chapter characterizes a decision maker with sticky beliefs. That is, a decision maker who does not update enough in response to information, where enough means as a Bayesian decision maker would. This chapter provides axiomatic foundations for sticky beliefs by weakening the standard axioms of dynamic consistency and consequentialism. I derive a representation in which updated beliefs are a convex combination of the prior and the Bayesian posterior. A unique parameter captures the weight on the prior and is interpreted as the agent's measure of belief stickiness or conservatism bias. This parameter is endogenously identified from preferences and is easily elicited from experimental data.

The third chapter deals with updating in the face of ambiguity, using the framework of Gilboa and Schmeidler. There is no consensus on the correct way way to update a set of priors. Current methods either do not allow a decision maker to make an inference about her priors or require an extreme level of inference. In this chapter I propose and axiomatize a general model of updating a set of priors. A decision maker who updates her beliefs in accordance with the model can be thought of as one that chooses a threshold that is used to determine whether a prior is plausible, given some observation. She retains the plausible priors and applies Bayes' rule. This model includes generalized Bayesian updating and maximum likelihood updating as special cases.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Fish growth is commonly estimated from length-at-age data obtained from otoliths. There are several techniques for estimating length-at-age from otoliths including 1) direct observed counts of annual increments; 2) age adjustment based on a categorization of otolith margins; 3) age adjustment based on known periods of spawning and annuli formation; 4) back-calculation to all annuli, and 5) back-calculation to the last annulus only. In this study we compared growth estimates (von Bertalanffy growth functions) obtained from the above five methods for estimating length-at-age from otoliths for two large scombrids: narrow-barred Spanish mackerel (Scomberomorus commerson) and broad-barred king mackerel (Scomberomorus semifasciatus). Likelihood ratio tests revealed that the largest differences in growth occurred between the back-calculation methods and the observed and adjusted methods for both species of mackerel. The pattern, however, was more pronounced for S. commerson than for S. semifasciatus, because of the pronounced effect of gear selectivity demonstrated for S. commerson. We propose a method of substituting length-at-age data from observed or adjusted methods with back-calculated length-at-age data to provide more appropriate estimates of population growth than those obtained with the individual methods alone, particularly when faster growing young fish are disproportionately selected for. Substitution of observed or adjusted length-at-age data with back-calculated length-at-age data provided more realistic estimates of length for younger ages than observed or adjusted methods as well as more realistic estimates of mean maximum length than those derived from backcalculation methods alone.