951 resultados para One-shot information theory
Resumo:
This thesis presents a thorough and principled investigation into the application of artificial neural networks to the biological monitoring of freshwater. It contains original ideas on the classification and interpretation of benthic macroinvertebrates, and aims to demonstrate their superiority over the biotic systems currently used in the UK to report river water quality. The conceptual basis of a new biological classification system is described, and a full review and analysis of a number of river data sets is presented. The biological classification is compared to the common biotic systems using data from the Upper Trent catchment. This data contained 292 expertly classified invertebrate samples identified to mixed taxonomic levels. The neural network experimental work concentrates on the classification of the invertebrate samples into biological class, where only a subset of the sample is used to form the classification. Other experimentation is conducted into the identification of novel input samples, the classification of samples from different biotopes and the use of prior information in the neural network models. The biological classification is shown to provide an intuitive interpretation of a graphical representation, generated without reference to the class labels, of the Upper Trent data. The selection of key indicator taxa is considered using three different approaches; one novel, one from information theory and one from classical statistical methods. Good indicators of quality class based on these analyses are found to be in good agreement with those chosen by a domain expert. The change in information associated with different levels of identification and enumeration of taxa is quantified. The feasibility of using neural network classifiers and predictors to develop numeric criteria for the biological assessment of sediment contamination in the Great Lakes is also investigated.
Resumo:
The effect of unitary noise on the discrete one-dimensional quantum walk is studied using computer simulations. For the noiseless quantum walk, starting at the origin (n=0) at time t=0, the position distribution P-t(n) at time t is very different from the Gaussian distribution obtained for the classical random walk. Furthermore, its standard deviation, sigma(t) scales as sigma(t)similar tot, unlike the classical random walk for which sigma(t)similar toroott. It is shown that when the quantum walk is exposed to unitary noise, it exhibits a crossover from quantum behavior for short times to classical-like behavior for long times. The crossover time is found to be Tsimilar toalpha(-2), where alpha is the standard deviation of the noise.
Resumo:
This paper studies the fundamental operational limits of a class of Gaussian multicast channels with an interference setting. In particular, the paper considers two base stations multicasting separate messages to distinct sets of users. In the presence of channel state information at the transmitters and at the respective receivers, the capacity region of the Gaussian multicast channel with interference is characterized to within one bit. At the crux of this result is an extension to the multicast channel with interference of the Han-Kobayashi or the Chong-Motani-Garg achievable region for the interference channel.
Resumo:
The effect of the heat flux on the rate of chemical reaction in dilute gases is shown to be important for reactions characterized by high activation energies and in the presence of very large temperature gradients. This effect, obtained from the second-order terms in the distribution function (similar to those obtained in the Burnett approximation to the solution of the Boltzmann equation), is derived on the basis of information theory. It is shown that the analytical results describing the effect are simpler if the kinetic definition for the nonequilibrium temperature is introduced than if the thermodynamic definition is introduced. The numerical results are nearly the same for both definitions
Resumo:
We consider bipartitions of one-dimensional extended systems whose probability distribution functions describe stationary states of stochastic models. We define estimators of the information shared between the two subsystems. If the correlation length is finite, the estimators stay finite for large system sizes. If the correlation length diverges, so do the estimators. The definition of the estimators is inspired by information theory. We look at several models and compare the behaviors of the estimators in the finite-size scaling limit. Analytical and numerical methods as well as Monte Carlo simulations are used. We show how the finite-size scaling functions change for various phase transitions, including the case where one has conformal invariance.
Resumo:
This paper studies a model of a sequential auction where bidders are allowed to acquire further information about their valuations of the object in the middle of the auction. It is shown that, in any equilibrium where the distribution of the final price is atornless, a bidder's best response has a simple characterization. In particular, the optimal information acquisition point is the same, regardless of the other bidders' actions. This makes it natural to focus on symmetric, undominated equilibria, as in the Vickrey auction. An existence theorem for such a class of equilibria is presented. The paper also presents some results and numerical simulations that compare this sequential auction with the one-shot auction. 8equential auctions typically yield more expected revenue for the seller than their one-shot counterparts. 80 the possibility of mid-auction information acquisition can provide an explanation for why sequential procedures are more often adopted.
Resumo:
Similarity measure is one of the main factors that affect the accuracy of intensity-based 2D/3D registration of X-ray fluoroscopy to CT images. Information theory has been used to derive similarity measure for image registration leading to the introduction of mutual information, an accurate similarity measure for multi-modal and mono-modal image registration tasks. However, it is known that the standard mutual information measure only takes intensity values into account without considering spatial information and its robustness is questionable. Previous attempt to incorporate spatial information into mutual information either requires computing the entropy of higher dimensional probability distributions, or is not robust to outliers. In this paper, we show how to incorporate spatial information into mutual information without suffering from these problems. Using a variational approximation derived from the Kullback-Leibler bound, spatial information can be effectively incorporated into mutual information via energy minimization. The resulting similarity measure has a least-squares form and can be effectively minimized by a multi-resolution Levenberg-Marquardt optimizer. Experimental results are presented on datasets of two applications: (a) intra-operative patient pose estimation from a few (e.g. 2) calibrated fluoroscopic images, and (b) post-operative cup alignment estimation from single X-ray radiograph with gonadal shielding.
Resumo:
Since the UsedSoft ruling of the CJEU in 2012, there has been the distinct feeling that – like the big bang - UsedSoft signals the start of a new beginning. As we enter this brave new world, the Copyright Directive will be read anew: misalignments in the treatment of physical and digital content will be resolved; accessibility and affordability for consumers will be heightened; and lock-in will be reduced as e-exhaustion takes hold. With UsedSoft as a precedent, the Court can do nothing but keep expanding its own ruling. For big bang theorists, it is only a matter of time until the digital first sale meteor strikes non-software downloads also. This paper looks at whether the UsedSoft ruling could indeed be the beginning of a wider doctrine of e-exhaustion, or if it is simply a one-shot comet restrained by provisions of the Computer Program Directive on which it was based. Fighting the latter corner, we have the strict word of the law; in the UsedSoft ruling, the Court appears to willingly bypass the international legal framework of the WCT. As far as expansion goes, the Copyright Directive was conceived specifically to implement the WCT, thus the legislative intent is clear. The Court would not, surely, invoke its modicum of creativity there also... With perhaps undue haste in a digital market of many unknowns, it seems this might well be the case. Provoking the big bang theory of e-exhaustion, the UsedSoft ruling can be read as distinctly purposive, but rather than having copyright norms in mind, the standard for the Court is the same free movement rules that underpin the exhaustion doctrine in the physical world. With an endowed sense of principled equivalence, the Court clearly wishes the tangible and intangible rules to be aligned. Against the backdrop of the European internal market, perhaps few legislative instruments would staunchly stand in its way. With firm objectives in mind, the UsedSoft ruling could be a rather disruptive meteor indeed.
Resumo:
The security of quantum key distribution protocols is guaranteed by the laws of quantum mechanics. However, a precise analysis of the security properties requires tools from both classical cryptography and information theory. Here, we employ recent results in non-asymptotic classical information theory to show that information reconciliation imposes fundamental limitations on the amount of secret key that can be extracted in the finite key regime. In particular, we find that an often used approximation for the information leakage during one-way information reconciliation is flawed and we propose an improved estimate.
Resumo:
In the first part of this work, we show how certain techniques from quantum information theory can be used in order to obtain very sharp embeddings between noncommutative Lp-spaces. Then, we use these estimates to study the classical capacity with restricted assisted entanglement of the quantum erasure channel and the quantum depolarizing channel. In particular, we exactly compute the capacity of the first one and we show that certain nonmultiplicative results hold for the second one.
Resumo:
Thesis (Ph.D.)--University of Washington, 2016-04
Resumo:
The problem of distributed compression for correlated quantum sources is considered. The classical version of this problem was solved by Slepian and Wolf, who showed that distributed compression could take full advantage of redundancy in the local sources created by the presence of correlations. Here it is shown that, in general, this is not the case for quantum sources, by proving a lower bound on the rate sum for irreducible sources of product states which is stronger than the one given by a naive application of Slepian-Wolf. Nonetheless, strategies taking advantage of correlation do exist for some special classes of quantum sources. For example, Devetak and Winter demonstrated the existence of such a strategy when one of the sources is classical. Optimal nontrivial strategies for a different extreme, sources of Bell states, are presented here. In addition, it is explained how distributed compression is connected to other problems in quantum information theory, including information-disturbance questions, entanglement distillation and quantum error correction.
Resumo:
The thesis deals with the problem of Model Selection (MS) motivated by information and prediction theory, focusing on parametric time series (TS) models. The main contribution of the thesis is the extension to the multivariate case of the Misspecification-Resistant Information Criterion (MRIC), a criterion introduced recently that solves Akaike’s original research problem posed 50 years ago, which led to the definition of the AIC. The importance of MS is witnessed by the huge amount of literature devoted to it and published in scientific journals of many different disciplines. Despite such a widespread treatment, the contributions that adopt a mathematically rigorous approach are not so numerous and one of the aims of this project is to review and assess them. Chapter 2 discusses methodological aspects of MS from information theory. Information criteria (IC) for the i.i.d. setting are surveyed along with their asymptotic properties; and the cases of small samples, misspecification, further estimators. Chapter 3 surveys criteria for TS. IC and prediction criteria are considered for: univariate models (AR, ARMA) in the time and frequency domain, parametric multivariate (VARMA, VAR); nonparametric nonlinear (NAR); and high-dimensional models. The MRIC answers Akaike’s original question on efficient criteria, for possibly-misspecified (PM) univariate TS models in multi-step prediction with high-dimensional data and nonlinear models. Chapter 4 extends the MRIC to PM multivariate TS models for multi-step prediction introducing the Vectorial MRIC (VMRIC). We show that the VMRIC is asymptotically efficient by proving the decomposition of the MSPE matrix and the consistency of its Method-of-Moments Estimator (MoME), for Least Squares multi-step prediction with univariate regressor. Chapter 5 extends the VMRIC to the general multiple regressor case, by showing that the MSPE matrix decomposition holds, obtaining consistency for its MoME, and proving its efficiency. The chapter concludes with a digression on the conditions for PM VARX models.
Resumo:
Oggigiorno il concetto di informazione è diventato cruciale in fisica, pertanto, siccome la migliore teoria che abbiamo per compiere predizioni riguardo l'universo è la meccanica quantistica, assume una particolare importanza lo sviluppo di una versione quantistica della teoria dell'informazione. Questa centralità è confermata dal fatto che i buchi neri hanno entropia. Per questo motivo, in questo lavoro sono presentati elementi di teoria dell'informazione quantistica e della comunicazione quantistica e alcuni sono illustrati riferendosi a modelli quantistici altamente idealizzati della meccanica di buco nero. In particolare, nel primo capitolo sono forniti tutti gli strumenti quanto-meccanici per la teoria dell'informazione e della comunicazione quantistica. Successivamente, viene affrontata la teoria dell'informazione quantistica e viene trovato il limite di Bekenstein alla quantità di informazione chiudibile entro una qualunque regione spaziale. Tale questione viene trattata utilizzando un modello quantistico idealizzato della meccanica di buco nero supportato dalla termodinamica. Nell'ultimo capitolo, viene esaminato il problema di trovare un tasso raggiungibile per la comunicazione quantistica facendo nuovamente uso di un modello quantistico idealizzato di un buco nero, al fine di illustrare elementi della teoria. Infine, un breve sommario della fisica dei buchi neri è fornito in appendice.
Resumo:
In a quantum critical chain, the scaling regime of the energy and momentum of the ground state and low-lying excitations are described by conformal field theory (CFT). The same holds true for the von Neumann and Renyi entropies of the ground state, which display a universal logarithmic behavior depending on the central charge. In this Letter we generalize this result to those excited states of the chain that correspond to primary fields in CFT. It is shown that the nth Renyi entropy is related to a 2n-point correlator of primary fields. We verify this statement for the critical XX and XXZ chains. This result uncovers a new link between quantum information theory and CFT.