351 resultados para pruning
Resumo:
This paper discusses the role of mango canopy architecture in mango productivity and orchard management and considers potential increases on production of high density orchards through improved canopy architecture. Lower tree height, reduced vigour and smaller more open canopies are recognised as important aspects of high density orchards. However, vigour management, light relations, flowering and crop load of high density orchards needs to be better understood if we are to developed sustainable highly productive canopy training and pruning systems that are easy to maintain at high planting densities.
Resumo:
Analyzing statistical dependencies is a fundamental problem in all empirical science. Dependencies help us understand causes and effects, create new scientific theories, and invent cures to problems. Nowadays, large amounts of data is available, but efficient computational tools for analyzing the data are missing. In this research, we develop efficient algorithms for a commonly occurring search problem - searching for the statistically most significant dependency rules in binary data. We consider dependency rules of the form X->A or X->not A, where X is a set of positive-valued attributes and A is a single attribute. Such rules describe which factors either increase or decrease the probability of the consequent A. A classical example are genetic and environmental factors, which can either cause or prevent a disease. The emphasis in this research is that the discovered dependencies should be genuine - i.e. they should also hold in future data. This is an important distinction from the traditional association rules, which - in spite of their name and a similar appearance to dependency rules - do not necessarily represent statistical dependencies at all or represent only spurious connections, which occur by chance. Therefore, the principal objective is to search for the rules with statistical significance measures. Another important objective is to search for only non-redundant rules, which express the real causes of dependence, without any occasional extra factors. The extra factors do not add any new information on the dependence, but can only blur it and make it less accurate in future data. The problem is computationally very demanding, because the number of all possible rules increases exponentially with the number of attributes. In addition, neither the statistical dependency nor the statistical significance are monotonic properties, which means that the traditional pruning techniques do not work. As a solution, we first derive the mathematical basis for pruning the search space with any well-behaving statistical significance measures. The mathematical theory is complemented by a new algorithmic invention, which enables an efficient search without any heuristic restrictions. The resulting algorithm can be used to search for both positive and negative dependencies with any commonly used statistical measures, like Fisher's exact test, the chi-squared measure, mutual information, and z scores. According to our experiments, the algorithm is well-scalable, especially with Fisher's exact test. It can easily handle even the densest data sets with 10000-20000 attributes. Still, the results are globally optimal, which is a remarkable improvement over the existing solutions. In practice, this means that the user does not have to worry whether the dependencies hold in future data or if the data still contains better, but undiscovered dependencies.
Resumo:
We present two discriminative language modelling techniques for Lempel-Ziv-Welch (LZW) based LID system. The previous approach to LID using LZW algorithm was to directly use the LZW pattern tables forlanguage modelling. But, since the patterns in a language pattern table are shared by other language pattern tables, confusability prevailed in the LID task. For overcoming this, we present two pruning techniques (i) Language Specific (LS-LZW)-in which patterns common to more than one pattern table are removed. (ii) Length-Frequency product based (LF-LZW)-in which patterns having their length-frequency product below a threshold are removed. These approaches reduce the classification score (Compression Ratio [LZW-CR] or the weighted discriminant score [LZW-WDS]) for non native languages and increases the LID performance considerably. Also the memory and computational requirements of these techniques are much less compared to basic LZW techniques.
Resumo:
BACKGROUND Polygenic risk scores comprising established susceptibility variants have shown to be informative classifiers for several complex diseases including prostate cancer. For prostate cancer it is unknown if inclusion of genetic markers that have so far not been associated with prostate cancer risk at a genome-wide significant level will improve disease prediction. METHODS We built polygenic risk scores in a large training set comprising over 25,000 individuals. Initially 65 established prostate cancer susceptibility variants were selected. After LD pruning additional variants were prioritized based on their association with prostate cancer. Six-fold cross validation was performed to assess genetic risk scores and optimize the number of additional variants to be included. The final model was evaluated in an independent study population including 1,370 cases and 1,239 controls. RESULTS The polygenic risk score with 65 established susceptibility variants provided an area under the curve (AUC) of 0.67. Adding an additional 68 novel variants significantly increased the AUC to 0.68 (P = 0.0012) and the net reclassification index with 0.21 (P = 8.5E-08). All novel variants were located in genomic regions established as associated with prostate cancer risk. CONCLUSIONS Inclusion of additional genetic variants from established prostate cancer susceptibility regions improves disease prediction. Prostate 75:1467–1474, 2015. © 2015 Wiley Periodicals, Inc.
Resumo:
The problem of automatic melody line identification in a MIDI file plays an important role towards taking QBH systems to the next level. We present here, a novel algorithm to identify the melody line in a polyphonic MIDI file. A note pruning and track/channel ranking method is used to identify the melody line. We use results from musicology to derive certain simple heuristics for the note pruning stage. This helps in the robustness of the algorithm, by way of discarding "spurious" notes. A ranking based on the melodic information in each track/channel enables us to choose the melody line accurately. Our algorithm makes no assumption about MIDI performer specific parameters, is simple and achieves an accuracy of 97% in identifying the melody line correctly. This algorithm is currently being used by us in a QBH system built in our lab.
Resumo:
In this paper, we present a growing and pruning radial basis function based no-reference (NR) image quality model for JPEG-coded images. The quality of the images are estimated without referring to their original images. The features for predicting the perceived image quality are extracted by considering key human visual sensitivity factors such as edge amplitude, edge length, background activity and background luminance. Image quality estimation involves computation of functional relationship between HVS features and subjective test scores. Here, the problem of quality estimation is transformed to a function approximation problem and solved using GAP-RBF network. GAP-RBF network uses sequential learning algorithm to approximate the functional relationship. The computational complexity and memory requirement are less in GAP-RBF algorithm compared to other batch learning algorithms. Also, the GAP-RBF algorithm finds a compact image quality model and does not require retraining when the new image samples are presented. Experimental results prove that the GAP-RBF image quality model does emulate the mean opinion score (MOS). The subjective test results of the proposed metric are compared with JPEG no-reference image quality index as well as full-reference structural similarity image quality index and it is observed to outperform both.
Resumo:
The impulse response of a typical wireless multipath channel can be modeled as a tapped delay line filter whose non-zero components are sparse relative to the channel delay spread. In this paper, a novel method of estimating such sparse multipath fading channels for OFDM systems is explored. In particular, Sparse Bayesian Learning (SBL) techniques are applied to jointly estimate the sparse channel and its second order statistics, and a new Bayesian Cramer-Rao bound is derived for the SBL algorithm. Further, in the context of OFDM channel estimation, an enhancement to the SBL algorithm is proposed, which uses an Expectation Maximization (EM) framework to jointly estimate the sparse channel, unknown data symbols and the second order statistics of the channel. The EM-SBL algorithm is able to recover the support as well as the channel taps more efficiently, and/or using fewer pilot symbols, than the SBL algorithm. To further improve the performance of the EM-SBL, a threshold-based pruning of the estimated second order statistics that are input to the algorithm is proposed, and its mean square error and symbol error rate performance is illustrated through Monte-Carlo simulations. Thus, the algorithms proposed in this paper are capable of obtaining efficient sparse channel estimates even in the presence of a small number of pilots.
Resumo:
We address the problem of mining targeted association rules over multidimensional market-basket data. Here, each transaction has, in addition to the set of purchased items, ancillary dimension attributes associated with it. Based on these dimensions, transactions can be visualized as distributed over cells of an n-dimensional cube. In this framework, a targeted association rule is of the form {X -> Y} R, where R is a convex region in the cube and X. Y is a traditional association rule within region R. We first describe the TOARM algorithm, based on classical techniques, for identifying targeted association rules. Then, we discuss the concepts of bottom-up aggregation and cubing, leading to the CellUnion technique. This approach is further extended, using notions of cube-count interleaving and credit-based pruning, to derive the IceCube algorithm. Our experiments demonstrate that IceCube consistently provides the best execution time performance, especially for large and complex data cubes.
Resumo:
In this paper, we study a problem of designing a multi-hop wireless network for interconnecting sensors (hereafter called source nodes) to a Base Station (BS), by deploying a minimum number of relay nodes at a subset of given potential locations, while meeting a quality of service (QoS) objective specified as a hop count bound for paths from the sources to the BS. The hop count bound suffices to ensure a certain probability of the data being delivered to the BS within a given maximum delay under a light traffic model. We observe that the problem is NP-Hard. For this problem, we propose a polynomial time approximation algorithm based on iteratively constructing shortest path trees and heuristically pruning away the relay nodes used until the hop count bound is violated. Results show that the algorithm performs efficiently in various randomly generated network scenarios; in over 90% of the tested scenarios, it gave solutions that were either optimal or were worse than optimal by just one relay. We then use random graph techniques to obtain, under a certain stochastic setting, an upper bound on the average case approximation ratio of a class of algorithms (including the proposed algorithm) for this problem as a function of the number of source nodes, and the hop count bound. To the best of our knowledge, the average case analysis is the first of its kind in the relay placement literature. Since the design is based on a light traffic model, we also provide simulation results (using models for the IEEE 802.15.4 physical layer and medium access control) to assess the traffic levels up to which the QoS objectives continue to be met. (C) 2014 Elsevier B.V. All rights reserved.
Resumo:
Frequent episode discovery is one of the methods used for temporal pattern discovery in sequential data. An episode is a partially ordered set of nodes with each node associated with an event type. For more than a decade, algorithms existed for episode discovery only when the associated partial order is total (serial episode) or trivial (parallel episode). Recently, the literature has seen algorithms for discovering episodes with general partial orders. In frequent pattern mining, the threshold beyond which a pattern is inferred to be interesting is typically user-defined and arbitrary. One way of addressing this issue in the pattern mining literature has been based on the framework of statistical hypothesis testing. This paper presents a method of assessing statistical significance of episode patterns with general partial orders. A method is proposed to calculate thresholds, on the non-overlapped frequency, beyond which an episode pattern would be inferred to be statistically significant. The method is first explained for the case of injective episodes with general partial orders. An injective episode is one where event-types are not allowed to repeat. Later it is pointed out how the method can be extended to the class of all episodes. The significance threshold calculations for general partial order episodes proposed here also generalize the existing significance results for serial episodes. Through simulations studies, the usefulness of these statistical thresholds in pruning uninteresting patterns is illustrated. (C) 2014 Elsevier Inc. All rights reserved.
Resumo:
A gradient in the density of hyperpolarization-activated cyclic-nucleotide gated (HCN) channels is necessary for the emergence of several functional maps within hippocampal pyramidal neurons. Here, we systematically analyzed the impact of dendritic atrophy on nine such functional maps, related to input resistance and local/transfer impedance properties, using conductance-based models of hippocampal pyramidal neurons. We introduced progressive dendritic atrophy in a CA1 pyramidal neuron reconstruction through a pruning algorithm, measured all functional maps in each pruned reconstruction, and arrived at functional forms for the dependence of underlying measurements on dendritic length. We found that, across frequencies, atrophied neurons responded with higher efficiency to incoming inputs, and the transfer of signals across the dendritic tree was more effective in an atrophied reconstruction. Importantly, despite the presence of identical HCN-channel density gradients, spatial gradients in input resistance, local/transfer resonance frequencies and impedance profiles were significantly constricted in reconstructions with dendrite atrophy, where these physiological measurements across dendritic locations converged to similar values. These results revealed that, in atrophied dendritic structures, the presence of an ion channel density gradient alone was insufficient to sustain homologous functional maps along the same neuronal topograph. We assessed the biophysical basis for these conclusions and found that this atrophy-induced constriction of functional maps was mediated by an enhanced spatial spread of the influence of an HCN-channel cluster in atrophied trees. These results demonstrated that the influence fields of ion channel conductances need to be localized for channel gradients to express themselves as homologous functional maps, suggesting that ion channel gradients are necessary but not sufficient for the emergence of functional maps within single neurons.
Resumo:
211 p. :il.
Resumo:
280 p. : il.
Resumo:
Microglia are largely known as the major orchestrators of the brain inflammatory response. As such, they have been traditionally studied in various contexts of disease, where their activation has been assumed to induce a wide range of detrimental effects. In the last few years, a series of discoveries have challenged the current view of microglia, showing their active and positive contribution to normal brain function. This Research Topic will review the novel physiological roles of microglia in the developing, mature and aging brain, under non-pathological conditions. In particular, this Research Topic will discuss the cellular and molecular mechanisms by which microglia contribute to the formation, pruning and plasticity of synapses; the maintenance of the blood brain barrier; the regulation of adult neurogenesis and hippocampal learning; and neuronal survival, among other important roles. Because these novel findings defy our understanding of microglial function in health as much as in disease, this Research Topic will also summarize the current view of microglial nomenclature, phenotypes, origin and differentiation, sex differences, and contribution to various brain pathologies. Additionally, novel imaging approaches and molecular tools to study microglia in their non-activated state will be discussed. In conclusion, this Research Topic seeks to emphasize how the current research in neuroscience is challenged by never-resting microglia.
Resumo:
A partir do consenso já existente, de que o desenvolvimento urbano é responsável, em parte, pelo desequilíbrio ambiental predominante nas cidades mais populosas, nas quais a administração dos resíduos gerados torna-se um grande desafio, este estudo foi realizado com a finalidade de desenvolver um modelo de gerenciamento para os resíduos de poda de árvores de espaços públicos, visando a utilização do material podado, considerado de boa qualidade, o que minimizaria a disposição de resíduos em aterros sanitários. Para tanto, foi desenvolvido um modelo diferenciado do ponto de vista de legal, gerencial, tecnológico e econômico, que pudesse servir de base à pesquisa e gerar estratégias para beneficiar o meio ambiente. A Unidade de Conservação, que pertence à Fundação Parques e Jardins da Prefeitura da Cidade do Rio de Janeiro, localizada na Taquara, foi analisada no Estudo de Caso. As espécies arbóreas que produzem maior volume de poda nessa seção foram selecionadas de modo que fosse possível o seu aproveitamento econômico-ecológico. Concluiu-se que há uma inviabilidade para segregação dos resíduos de poda por parte da Fundação Parques e Jardins e que os mesmos poderiam ser transferidos diretamente para o aterro receptor, em fase de encerramento de atividades, sem custos excedentes. Foi feita uma apreciação especial do Centro de Tratamento de Resíduos Sólidos de Gericinó, por ser grande receptor dos resíduos produzidos nas operações de manejo da área em evidência. Foi elaborada a proposta de criação de uma Usina Verde nas áreas já desativadas do aterro, como forma complementar ao processo de revitalização da área aterrada após o término de suas atividades. Esta ação contemplaria a região com um bosque, onde seriam absorvidos todos os produtos dos resíduos de poda. Haveria, também, a probabilidade de utilização operacional dos catadores nas etapas de obtenção de compostos orgânicos, cobertura morta e equipamentos paisagísticos entre outros.