156 resultados para Bayesian Latent Class
Resumo:
It is important to identify the ``correct'' number of topics in mechanisms like Latent Dirichlet Allocation(LDA) as they determine the quality of features that are presented as features for classifiers like SVM. In this work we propose a measure to identify the correct number of topics and offer empirical evidence in its favor in terms of classification accuracy and the number of topics that are naturally present in the corpus. We show the merit of the measure by applying it on real-world as well as synthetic data sets(both text and images). In proposing this measure, we view LDA as a matrix factorization mechanism, wherein a given corpus C is split into two matrix factors M-1 and M-2 as given by C-d*w = M1(d*t) x Q(t*w).Where d is the number of documents present in the corpus anti w is the size of the vocabulary. The quality of the split depends on ``t'', the right number of topics chosen. The measure is computed in terms of symmetric KL-Divergence of salient distributions that are derived from these matrix factors. We observe that the divergence values are higher for non-optimal number of topics - this is shown by a `dip' at the right value for `t'.
Resumo:
Stochastic behavior of an aero-engine failure/repair process has been analyzed from a Bayesian perspective. Number of failures/repairs in the component-sockets of this multi-component system are assumed to follow independent renewal processes with Weibull inter-arrival times. Based on the field failure/repair data of a large number of such engines and independent Gamma priors on the scale parameters and log-concave priors on the shape parameters, an exact method of sampling from the resulting posterior distributions of the parameters has been proposed. These generated parameter values are next utilised in obtaining the posteriors of the expected number of system repairs, system failure rate, and the conditional intensity function, which are computed using a recursive formula.
Resumo:
Recently we have reported the effect of (S)-6-aryl urea/thiourea substituted-2-amino-4,5,6,7-tetrahydrobenzod]thiazole derivatives as potent anti-leukemic agents. To elucidate further the Structure Activity Relationship (SAR) studies on the anti-leukemic activity of (S)-2,6-diamino-4,5,6,7 tetrahydrobenzod]thiazole moiety, a series of 2-arlycarboxamide substituted-(S)-6-amino-4,5,6,7-tetrahydrobenzod]thiazole were designed, synthesized and evaluated for their anti-leukemic activity by trypan blue exclusion, 3-(4,5-dimethylthiazol-2-yl)-2,5-diphenyltetrazolium bromide (MTT), lactate dehydrogenase (LDH) assays and cell cycle analysis. Results suggest that the position, number and bulkiness of the substituent on the phenyl ring of aryl carboxamide moiety at 2nd position of 6-amino-4,5,6,7-tetrhydrobenzod]thiazole play a key role in inhibiting the proliferation of leukemia cells. Compounds with ortho substitution showed poor activity and with meta and para substitution showed good activity. (C) 2010 Elsevier Masson SAS. All rights reserved.
Resumo:
Representation and quantification of uncertainty in climate change impact studies are a difficult task. Several sources of uncertainty arise in studies of hydrologic impacts of climate change, such as those due to choice of general circulation models (GCMs), scenarios and downscaling methods. Recently, much work has focused on uncertainty quantification and modeling in regional climate change impacts. In this paper, an uncertainty modeling framework is evaluated, which uses a generalized uncertainty measure to combine GCM, scenario and downscaling uncertainties. The Dempster-Shafer (D-S) evidence theory is used for representing and combining uncertainty from various sources. A significant advantage of the D-S framework over the traditional probabilistic approach is that it allows for the allocation of a probability mass to sets or intervals, and can hence handle both aleatory or stochastic uncertainty, and epistemic or subjective uncertainty. This paper shows how the D-S theory can be used to represent beliefs in some hypotheses such as hydrologic drought or wet conditions, describe uncertainty and ignorance in the system, and give a quantitative measurement of belief and plausibility in results. The D-S approach has been used in this work for information synthesis using various evidence combination rules having different conflict modeling approaches. A case study is presented for hydrologic drought prediction using downscaled streamflow in the Mahanadi River at Hirakud in Orissa, India. Projections of n most likely monsoon streamflow sequences are obtained from a conditional random field (CRF) downscaling model, using an ensemble of three GCMs for three scenarios, which are converted to monsoon standardized streamflow index (SSFI-4) series. This range is used to specify the basic probability assignment (bpa) for a Dempster-Shafer structure, which represents uncertainty associated with each of the SSFI-4 classifications. These uncertainties are then combined across GCMs and scenarios using various evidence combination rules given by the D-S theory. A Bayesian approach is also presented for this case study, which models the uncertainty in projected frequencies of SSFI-4 classifications by deriving a posterior distribution for the frequency of each classification, using an ensemble of GCMs and scenarios. Results from the D-S and Bayesian approaches are compared, and relative merits of each approach are discussed. Both approaches show an increasing probability of extreme, severe and moderate droughts and decreasing probability of normal and wet conditions in Orissa as a result of climate change. (C) 2010 Elsevier Ltd. All rights reserved.
Resumo:
A feature common to many adaptive systems for identification and control is the adjustment.of gain parameters in a manner ensuring the stability of the overall system. This paper puts forward a principle which assures such a result for arbitrary systems which are linear and time invariant except for the adjustable parameters. The principle only demands that a transfer function be positive real. This transfer function dependent on the structure of the system with respect to the parameters. Several examples from adaptive identification, control and observer schemes are given as illustrations of the conceptual simplification provided by the structural principle.
Resumo:
We show how, for large classes of systems with purely second-class constraints, further information can be obtained about the constraint algebra. In particular, a subset consisting of half the full set of constraints is shown to have vanishing mutual brackets. Some other constraint brackets are also shown to be zero. The class of systems for which our results hold includes examples from non-relativistic particle mechanics as well as relativistic field theory. The results are derived at the classical level for Poisson brackets, but in the absence of commutator anomalies the same results will hold for the commutators of the constraint operators in the corresponding quantised theories.
Resumo:
Given a classical dynamical theory with second-class constraints, it is sometimes possible to construct another theory with first-class constraints, i.e., a gauge-invariant one, which is physically equivalent to the first theory. We identify some conditions under which this may be done, explaining the general principles and working out several examples. Field theoretic applications include the chiral Schwinger model and the non-linear sigma model. An interesting connection with the work of Faddeev and Shatashvili is pointed out.
Resumo:
We address risk minimizing option pricing in a regime switching market where the floating interest rate depends on a finite state Markov process. The growth rate and the volatility of the stock also depend on the Markov process. Using the minimal martingale measure, we show that the locally risk minimizing prices for certain exotic options satisfy a system of Black-Scholes partial differential equations with appropriate boundary conditions. We find the corresponding hedging strategies and the residual risk. We develop suitable numerical methods to compute option prices.
Resumo:
The cell envelope of Mycobacterium tuberculosis (M. tuberculosis) is composed of a variety of lipids including mycolic acids, sulpholipids, lipoarabinomannans, etc., which impart rigidity crucial for its survival and pathogenesis. Acyl CoA carboxylase (ACC) provides malonyl-CoA and methylmalonyl-CoA, committed precursors for fatty acid and essential for mycolic acid synthesis respectively. Biotin Protein Ligase (BPL/BirA) activates apo-biotin carboxyl carrier protein (BCCP) by biotinylating it to an active holo-BCCP. A minimal peptide (Schatz), an efficient substrate for Escherichia coli BirA, failed to serve as substrate for M. tuberculosis Biotin Protein Ligase (MtBPL). MtBPL specifically biotinylates homologous BCCP domain, MtBCCP87, but not EcBCCP87. This is a unique feature of MtBPL as EcBirA lacks such a stringent substrate specificity. This feature is also reflected in the lack of self/promiscuous biotinylation by MtBPL. The N-terminus/HTH domain of EcBirA has the selfbiotinable lysine residue that is inhibited in the presence of Schatz peptide, a peptide designed to act as a universal acceptor for EcBirA. This suggests that when biotin is limiting, EcBirA preferentially catalyzes, biotinylation of BCCP over selfbiotinylation. R118G mutant of EcBirA showed enhanced self and promiscuous biotinylation but its homologue, R69A MtBPL did not exhibit these properties. The catalytic domain of MtBPL was characterized further by limited proteolysis. Holo-MtBPL is protected from proteolysis by biotinyl-59 AMP, an intermediate of MtBPL catalyzed reaction. In contrast, apo-MtBPL is completely digested by trypsin within 20 min of co-incubation. Substrate selectivity and inability to promote self biotinylation are exquisite features of MtBPL and are a consequence of the unique molecular mechanism of an enzyme adapted for the high turnover of fatty acid biosynthesis.
Resumo:
The decision to patent a technology is a difficult one to make for the top management of any organization. The expected value that the patent might deliver in the market is an important factor that impacts this judgement. Earlier researchers have suggested that patent prices are better indicators of value of a patent and that auction prices are the best way of determining value. However, the lack of public data on pricing has prevented research on understanding the dynamics of patent pricing. Our paper uses singleton patent auction price data of Ocean Tomo LLC to study the prices of patents. We describe price characteristics of these patents. The price of these patents was correlated with their age, and a significant correlation was found. A price - age matrix was developed and we describe the price characteristics of patents using four quadrants of the matrix, namely young and old patents with low and high prices. We also found that patents owned by small firms get transacted more often and inventor owned patents attracted a better price than assignee owned patents.
Resumo:
This paper considers the problem of spectrum sensing, i.e., the detection of whether or not a primary user is transmitting data by a cognitive radio. The Bayesian framework is adopted, with the performance measure being the probability of detection error. A decentralized setup, where N sensors use M observations each to arrive at individual decisions that are combined at a fusion center to form the overall decision is considered. The unknown fading channel between the primary sensor and the cognitive radios makes the individual decision rule computationally complex, hence, a generalized likelihood ratio test (GLRT)-based approach is adopted. Analysis of the probabilities of false alarm and miss detection of the proposed method reveals that the error exponent with respect to M is zero. Also, the fusion of N individual decisions offers a diversity advantage, similar to diversity reception in communication systems, and a tight bound on the error exponent is presented. Through an analysis in the low power regime, the number of observations needed as a function of received power, to achieve a given probability of error is determined. Monte-Carlo simulations confirm the accuracy of the analysis.
Resumo:
In the distributed storage setting introduced by Dimakis et al., B units of data are stored across n nodes in the network in such a way that the data can be recovered by connecting to any k nodes. Additionally one can repair a failed node by connecting to any d nodes while downloading at most beta units of data from each node. In this paper, we introduce a flexible framework in which the data can be recovered by connecting to any number of nodes as long as the total amount of data downloaded is at least B. Similarly, regeneration of a failed node is possible if the new node connects to the network using links whose individual capacity is bounded above by beta(max) and whose sum capacity equals or exceeds a predetermined parameter gamma. In this flexible setting, we obtain the cut-set lower bound on the repair bandwidth along with a constructive proof for the existence of codes meeting this bound for all values of the parameters. An explicit code construction is provided which is optimal in certain parameter regimes.
Resumo:
An explicit construction of all the homogeneous holomorphic Hermitian vector bundles over the unit disc D is given. It is shown that every such vector bundle is a direct sum of irreducible ones. Among these irreducible homogeneous holomorphic Hermitian vector bundles over D, the ones corresponding to operators in the Cowen-Douglas class B-n(D) are identified. The classification of homogeneous operators in B-n(D) is completed using an explicit realization of these operators. We also show how the homogeneous operators in B-n(D) split into similarity classes. (C) 2011 Elsevier Inc. All rights reserved.
Resumo:
Nuclear import of proteins is mediated by the nuclear pore complexes in the nuclear envelope and requires the presence of a nuclear localization signal (NLS) on the karyophilic protein. In this paper, we describe studies with a monoclonal antibody, Mab E2, which recognizes a class of nuclear pore proteins of 60-76 kDa with a common phosphorylated epitope on rat nuclear envelopes. The Mab Ea-reactive proteins fractionated with the relatively insoluble pore complex-containing component of the envelope and gave a finely punctate pattern of nuclear staining in immunofluorescence assays. The antibody did not bind to any cytosolic proteins. Mab E2 inhibited the interaction of a simian virus 40 large T antigen NLS peptide with a specific 60-kDa NLS-binding protein from rat nuclear envelopes in photoaffinity labeling experiments. The antibody blocked the nuclear import of NLS-albumin conjugates in an in vitro nuclear transport assay with digitonin-permeabilized cells, but did not affect passive diffusion of a small nonnuclear protein, lysozyme, across the pore. Mab E2 may inhibit protein transport by directly interacting with the 60-kDa NLS-binding protein, thereby blocking signal-mediated nuclear import across the nuclear pore complex. (C) 1994 Academic Press, Inc.
Resumo:
The impulse response of a typical wireless multipath channel can be modeled as a tapped delay line filter whose non-zero components are sparse relative to the channel delay spread. In this paper, a novel method of estimating such sparse multipath fading channels for OFDM systems is explored. In particular, Sparse Bayesian Learning (SBL) techniques are applied to jointly estimate the sparse channel and its second order statistics, and a new Bayesian Cramer-Rao bound is derived for the SBL algorithm. Further, in the context of OFDM channel estimation, an enhancement to the SBL algorithm is proposed, which uses an Expectation Maximization (EM) framework to jointly estimate the sparse channel, unknown data symbols and the second order statistics of the channel. The EM-SBL algorithm is able to recover the support as well as the channel taps more efficiently, and/or using fewer pilot symbols, than the SBL algorithm. To further improve the performance of the EM-SBL, a threshold-based pruning of the estimated second order statistics that are input to the algorithm is proposed, and its mean square error and symbol error rate performance is illustrated through Monte-Carlo simulations. Thus, the algorithms proposed in this paper are capable of obtaining efficient sparse channel estimates even in the presence of a small number of pilots.