973 resultados para Number of samples


Relevância:

100.00% 100.00%

Publicador:

Resumo:

It is important to identify the ``correct'' number of topics in mechanisms like Latent Dirichlet Allocation(LDA) as they determine the quality of features that are presented as features for classifiers like SVM. In this work we propose a measure to identify the correct number of topics and offer empirical evidence in its favor in terms of classification accuracy and the number of topics that are naturally present in the corpus. We show the merit of the measure by applying it on real-world as well as synthetic data sets(both text and images). In proposing this measure, we view LDA as a matrix factorization mechanism, wherein a given corpus C is split into two matrix factors M-1 and M-2 as given by C-d*w = M1(d*t) x Q(t*w).Where d is the number of documents present in the corpus anti w is the size of the vocabulary. The quality of the split depends on ``t'', the right number of topics chosen. The measure is computed in terms of symmetric KL-Divergence of salient distributions that are derived from these matrix factors. We observe that the divergence values are higher for non-optimal number of topics - this is shown by a `dip' at the right value for `t'.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Design of speaker identification schemes for a small number of speakers (around 10) with a high degree of accuracy in controlled environment is a practical proposition today. When the number of speakers is large (say 50–100), many of these schemes cannot be directly extended, as both recognition error and computation time increase monotonically with population size. The feature selection problem is also complex for such schemes. Though there were earlier attempts to rank order features based on statistical distance measures, it has been observed only recently that the best two independent measurements are not the same as the combination in two's for pattern classification. We propose here a systematic approach to the problem using the decision tree or hierarchical classifier with the following objectives: (1) Design of optimal policy at each node of the tree given the tree structure i.e., the tree skeleton and the features to be used at each node. (2) Determination of the optimal feature measurement and decision policy given only the tree skeleton. Applicability of optimization procedures such as dynamic programming in the design of such trees is studied. The experimental results deal with the design of a 50 speaker identification scheme based on this approach.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Tiivistelmä ReferatAbstract Metabolomics is a rapidly growing research field that studies the response of biological systems to environmental factors, disease states and genetic modifications. It aims at measuring the complete set of endogenous metabolites, i.e. the metabolome, in a biological sample such as plasma or cells. Because metabolites are the intermediates and end products of biochemical reactions, metabolite compositions and metabolite levels in biological samples can provide a wealth of information on on-going processes in a living system. Due to the complexity of the metabolome, metabolomic analysis poses a challenge to analytical chemistry. Adequate sample preparation is critical to accurate and reproducible analysis, and the analytical techniques must have high resolution and sensitivity to allow detection of as many metabolites as possible. Furthermore, as the information contained in the metabolome is immense, the data set collected from metabolomic studies is very large. In order to extract the relevant information from such large data sets, efficient data processing and multivariate data analysis methods are needed. In the research presented in this thesis, metabolomics was used to study mechanisms of polymeric gene delivery to retinal pigment epithelial (RPE) cells. The aim of the study was to detect differences in metabolomic fingerprints between transfected cells and non-transfected controls, and thereafter to identify metabolites responsible for the discrimination. The plasmid pCMV-β was introduced into RPE cells using the vector polyethyleneimine (PEI). The samples were analyzed using high performance liquid chromatography (HPLC) and ultra performance liquid chromatography (UPLC) coupled to a triple quadrupole (QqQ) mass spectrometer (MS). The software MZmine was used for raw data processing and principal component analysis (PCA) was used in statistical data analysis. The results revealed differences in metabolomic fingerprints between transfected cells and non-transfected controls. However, reliable fingerprinting data could not be obtained because of low analysis repeatability. Therefore, no attempts were made to identify metabolites responsible for discrimination between sample groups. Repeatability and accuracy of analyses can be influenced by protocol optimization. However, in this study, optimization of analytical methods was hindered by the very small number of samples available for analysis. In conclusion, this study demonstrates that obtaining reliable fingerprinting data is technically demanding, and the protocols need to be thoroughly optimized in order to approach the goals of gaining information on mechanisms of gene delivery.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The integral diaphragm pressure transducers machined out of precipitation hardened martensite stainless steel (APX4) are widely used for propellant pressure measurements in space applications. These transducers are expected to exhibit dimensional stability and linearity for their entire useful life. These vital factors are very critical for the reliable performance and dependability of the pressure transducers. However, these transducers invariably develop internal stresses during various stages of machining. These stresses have an adverse effect on the performance of the transducers causing deviation from linearity. In order to eliminate these possibilities, it was planned to cryotreat the machined transducers to improve both the long-term linearity and dimensional stability. To study these effects, an experimental cryotreatment unit was designed and developed based on the concept of indirect cooling using the concept of cold nitrogen gas forced closed loop convection currents. The system has the capability of cryotreating large number of samples for varied rates of cooling, soaking and warm-up. After obtaining the initial levels of residual stress and retained austenite using X-ray diffraction techniques, the pressure transducers were cryotreated at 98 K for 36 h. Immediately after cryotreatment, the transducers were tempered at 510 degrees C for 3 h in vacuum furnace. Results after cryo treatment clearly indicated significant reduction in residual stress levels and conversion of retained austenite to martensite. These changes have brought in improvements in long term zero drift and dimensional stability. The cryotreated pressure transducers have been incorporated for actual space applications. (c) 2010 Published by Elsevier Ltd.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The study of soil microbiota and their activities is central to the understanding of many ecosystem processes such as decomposition and nutrient cycling. The collection of microbiological data from soils generally involves several sequential steps of sampling, pretreatment and laboratory measurements. The reliability of results is dependent on reliable methods in every step. The aim of this thesis was to critically evaluate some central methods and procedures used in soil microbiological studies in order to increase our understanding of the factors that affect the measurement results and to provide guidance and new approaches for the design of experiments. The thesis focuses on four major themes: 1) soil microbiological heterogeneity and sampling, 2) storage of soil samples, 3) DNA extraction from soil, and 4) quantification of specific microbial groups by the most-probable-number (MPN) procedure. Soil heterogeneity and sampling are discussed as a single theme because knowledge on spatial (horizontal and vertical) and temporal variation is crucial when designing sampling procedures. Comparison of adjacent forest, meadow and cropped field plots showed that land use has a strong impact on the degree of horizontal variation of soil enzyme activities and bacterial community structure. However, regardless of the land use, the variation of microbiological characteristics appeared not to have predictable spatial structure at 0.5-10 m. Temporal and soil depth-related patterns were studied in relation to plant growth in cropped soil. The results showed that most enzyme activities and microbial biomass have a clear decreasing trend in the top 40 cm soil profile and a temporal pattern during the growing season. A new procedure for sampling of soil microbiological characteristics based on stratified sampling and pre-characterisation of samples was developed. A practical example demonstrated the potential of the new procedure to reduce the analysis efforts involved in laborious microbiological measurements without loss of precision. The investigation of storage of soil samples revealed that freezing (-20 °C) of small sample aliquots retains the activity of hydrolytic enzymes and the structure of the bacterial community in different soil matrices relatively well whereas air-drying cannot be recommended as a storage method for soil microbiological properties due to large reductions in activity. Freezing below -70 °C was the preferred method of storage for samples with high organic matter content. Comparison of different direct DNA extraction methods showed that the cell lysis treatment has a strong impact on the molecular size of DNA obtained and on the bacterial community structure detected. An improved MPN method for the enumeration of soil naphthalene degraders was introduced as an alternative to more complex MPN protocols or the DNA-based quantification approach. The main advantage of the new method is the simple protocol and the possibility to analyse a large number of samples and replicates simultaneously.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The mechanical properties of composites of polymethylmethacrylate (PMMA) with two-dimensional graphene-like boron nitride (BN) have been investigated to explore the dependence of the properties on the number of BN layers. This study demonstrates that significantly improved mechanical properties are exhibited by the composite with the fewest number of BN layers. Thus, with incorporation of three BN layers, the hardness and elastic modulus of the composite showed an increase of 125% and 130%, respectively, relative to pure PMMA. (C) 2010 Acta Materialia Inc. Published by Elsevier Ltd. All rights reserved.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Higher level of inversion is achieved with a less number of switches in the proposed scheme. The scheme proposes a five-level inverter for an open-end winding induction motor which uses only two DC-link rectifiers of voltage rating of Vdc/4, a neutral-point clamped (NPC) three-level inverter and a two-level inverter. Even though the two-level inverter is connected to the high-voltage side, it is always in square-wave operation. Since the two-level inverter is not switching in a pulse width modulated fashion and the magnitude of switching transient is only half compared to the convention three-level NPC inverter, the switching losses and electromagnetic interference is not so high. The scheme is experimentally verified on a 2.5 kW induction machine.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Thiobacillus ferrooxidans MAL4-1, an isolate from Malanjkhand copper mines, India, was adapted to grow in the presence of high concentration (30 gL(-1)) of Cu2+, resulting in a 15-fold increase in its tolerance to Cu2+. While wild-type T. ferrooxidans MAL4-1 contained multiple plasmids, cultures adapted to Cu2+ concentrations of 20 gL(-1) or more showed a drastic reduction in the copy number of the plasmids. The reduction for three of the plasmids was estimated to be over 50-fold. Examination of the plasmid profiles of the strains adapted to high concentration of SO42- anion (as Na2SO4 or ZnSO4) indicated that the reduction in plasmid copy number is not owing to SO42- anion, but is specific for Cu2+. The effect of mercury on the plasmids was similar to that of copper. Deadaptation of the Cu2+- Or Hg2+-adapted T. ferrooxidans resulted in restoration of the plasmids to the original level within the first passage. The fact that the plasmid copy number, in general, is drastically reduced in Cu2+-adapted T. ferrooxidans suggests that resistance to copper is chromosome mediated. This is the first report of a selective negative influence of copper ions on the copy number of plasmids in T. ferrooxidans.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Suspensions of testicular germ cells from six species of mammals were prepared and stained for the DNA content with a fluorochrome (ethidium bromide) adopting a common technique and subjected to DNA flow cytometry. While uniform staining of the germ cells of the mouse, hamster, rat and monkey could be obtained by treating with 0.5% pepsin for 60 min followed by staining with ethidium bromide for 30 min, that of the guinea pig and rabbit required for optimal staining pepsinization for 90 min and treatment with ethidium bromide for 60 min. The procedure adopted here provided a uniform recovery of over 80% of germ cells with each one of the species tested and the cell population distributed itself according to the DNA content (expressed as C values) into 5 major classes-spermatogonia (2C), cells in S-phase, primary spermatocytes (4C), round spermatids (1C), and elongating/elongated spermatids (HC). Comparison of the DNA distribution pattern of the germ cell populations between species revealed little variation in the relative quantities of cells with 2C (8-11%), S-phase (6-9%), and 4C (6-9%) amount of DNA. Though the spermatid cell populations exhibited variations (1C:31-46%, HCI:7-20% and and HC2:11-25%) they represented the bulk of germ cells (70-80%). The overall conversion of 2C to 1C (1C:2C ratio) and meiotic transformation of 4C cells to IC (1C:4C ratio) kinetics were relatively constant between the species studied. The present study clearly demonstrates that DNA flow cytometry can be adopted with ease and assurance to quantify germ cell transformation and as such spermatogenesis by analysing a large number of samples with consistency both within and across the species barrier. Any variation from the norms in germ cell proportions observed following treatment, for e.g. hormonal stimulation or deprivation can then be ascribed due to a specific effect of the hormone/drug on single/multiple steps in germ cell transformation

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Let G be an undirected graph with a positive real weight on each edge. It is shown that the number of minimum-weight cycles of G is bounded above by a polynomial in the number of edges of G. A similar bound holds if we wish to count the number of cycles with weight at most a constant multiple of the minimum weight of a cycle of G.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

We consider a system comprising a finite number of nodes, with infinite packet buffers, that use unslotted ALOHA with Code Division Multiple Access (CDMA) to share a channel for transmitting packetised data. We propose a simple model for packet transmission and retransmission at each node, and show that saturation throughput in this model yields a sufficient condition for the stability of the packet buffers; we interpret this as the capacity of the access method. We calculate and compare the capacities of CDMA-ALOHA (with and without code sharing) and TDMA-ALOHA; we also consider carrier sensing and collision detection versions of these protocols. In each case, saturation throughput can be obtained via analysis pf a continuous time Markov chain. Our results show how saturation throughput degrades with code-sharing. Finally, we also present some simulation results for mean packet delay. Our work is motivated by optical CDMA in which "chips" can be optically generated, and hence the achievable chip rate can exceed the achievable TDMA bit rate which is limited by electronics. Code sharing may be useful in the optical CDMA context as it reduces the number of optical correlators at the receivers. Our throughput results help to quantify by how much the CDMA chip rate should exceed the TDMA bit rate so that CDMA-ALOHA yields better capacity than TDMA-ALOHA.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

We address the problem of allocating a single divisible good to a number of agents. The agents have concave valuation functions parameterized by a scalar type. The agents report only the type. The goal is to find allocatively efficient, strategy proof, nearly budget balanced mechanisms within the Groves class. Near budget balance is attained by returning as much of the received payments as rebates to agents. Two performance criteria are of interest: the maximum ratio of budget surplus to efficient surplus, and the expected budget surplus, within the class of linear rebate functions. The goal is to minimize them. Assuming that the valuation functions are known, we show that both problems reduce to convex optimization problems, where the convex constraint sets are characterized by a continuum of half-plane constraints parameterized by the vector of reported types. We then propose a randomized relaxation of these problems by sampling constraints. The relaxed problem is a linear programming problem (LP). We then identify the number of samples needed for ``near-feasibility'' of the relaxed constraint set. Under some conditions on the valuation function, we show that value of the approximate LP is close to the optimal value. Simulation results show significant improvements of our proposed method over the Vickrey-Clarke-Groves (VCG) mechanism without rebates. In the special case of indivisible goods, the mechanisms in this paper fall back to those proposed by Moulin, by Guo and Conitzer, and by Gujar and Narahari, without any need for randomization. Extension of the proposed mechanisms to situations when the valuation functions are not known to the central planner are also discussed. Note to Practitioners-Our results will be useful in all resource allocation problems that involve gathering of information privately held by strategic users, where the utilities are any concave function of the allocations, and where the resource planner is not interested in maximizing revenue, but in efficient sharing of the resource. Such situations arise quite often in fair sharing of internet resources, fair sharing of funds across departments within the same parent organization, auctioning of public goods, etc. We study methods to achieve near budget balance by first collecting payments according to the celebrated VCG mechanism, and then returning as much of the collected money as rebates. Our focus on linear rebate functions allows for easy implementation. The resulting convex optimization problem is solved via relaxation to a randomized linear programming problem, for which several efficient solvers exist. This relaxation is enabled by constraint sampling. Keeping practitioners in mind, we identify the number of samples that assures a desired level of ``near-feasibility'' with the desired confidence level. Our methodology will occasionally require subsidy from outside the system. We however demonstrate via simulation that, if the mechanism is repeated several times over independent instances, then past surplus can support the subsidy requirements. We also extend our results to situations where the strategic users' utility functions are not known to the allocating entity, a common situation in the context of internet users and other problems.