39 resultados para Incremental Information-content
Resumo:
Some models of sexual selection predict that individuals vary in their genetic quality and reveal some of this variation in their secondary sexual characteristics. Alpine whitefish (Coregonus sp.) develop breeding tubercles shortly before their spawning season. These tubercles are epidermal structures that are distributed regularly along the body sides of both males and females. There is still much unexplained variation in the size of breeding tubercles within both sexes and with much overlap between the sexes. It has been suggested that breeding tubercles function to maintain body contact between the mating partners during spawning, act as weapons for defence of spawning territories, or are sexual signals that reveal aspects of genetic quality. We took two samples of whitefish from their spawning place, one at the beginning and one around the peak of spawning season. We found that females have on average smaller breeding tubercles than males, and that tubercle size partly reveals the stage of gonad maturation. Two independent full-factorial breeding experiments revealed that embryo mortality was significantly influenced by male and female effects. This finding demonstrates that the males differed in their genetic quality (because offspring get nothing but genes from their fathers). Tubercle size was negatively linked to some aspects of embryo mortality in the first breeding experiment but not significantly so in the second. This lack of consistency adds to inconsistent results that were reported before and suggests that (i) some aspects of genetic quality are not revealed in breeding tubercles while others are, or (ii) individuals vary in their signaling strategies and the information content of breeding tubercles is not always reliable. Moreover, the fact that female whitefish have breeding tubercles of significant size while males seem to have few reasons to be choosy suggests that the tubercles might also serve some functions that are not linked to sexual signaling.
Resumo:
Time-lapse geophysical data acquired during transient hydrological experiments are being increasingly employed to estimate subsurface hydraulic properties at the field scale. In particular, crosshole ground-penetrating radar (GPR) data, collected while water infiltrates into the subsurface either by natural or artificial means, have been demonstrated in a number of studies to contain valuable information concerning the hydraulic properties of the unsaturated zone. Previous work in this domain has considered a variety of infiltration conditions and different amounts of time-lapse GPR data in the estimation procedure. However, the particular benefits and drawbacks of these different strategies as well as the impact of a variety of key and common assumptions remain unclear. Using a Bayesian Markov-chain-Monte-Carlo stochastic inversion methodology, we examine in this paper the information content of time-lapse zero-offset-profile (ZOP) GPR traveltime data, collected under three different infiltration conditions, for the estimation of van Genuchten-Mualem (VGM) parameters in a layered subsurface medium. Specifically, we systematically analyze synthetic and field GPR data acquired under natural loading and two rates of forced infiltration, and we consider the value of incorporating different amounts of time-lapse measurements into the estimation procedure. Our results confirm that, for all infiltration scenarios considered, the ZOP GPR traveltime data contain important information about subsurface hydraulic properties as a function of depth, with forced infiltration offering the greatest potential for VGM parameter refinement because of the higher stressing of the hydrological system. Considering greater amounts of time-lapse data in the inversion procedure is also found to help refine VGM parameter estimates. Quite importantly, however, inconsistencies observed in the field results point to the strong possibility that posterior uncertainties are being influenced by model structural errors, which in turn underlines the fundamental importance of a systematic analysis of such errors in future related studies.
Resumo:
Some models of sexual selection predict that individuals vary in their genetic quality and reveal some of this variation in their secondary sexual characteristics. Alpine whitefish (Coregonus sp.) develop breeding tubercles shortly before their spawning season. These tubercles are epidermal structures that are distributed regularly along the body sides of both males and females. There is still much unexplained variation in the size of breeding tubercles within both sexes and with much overlap between the sexes. It has been suggested that breeding tubercles function to maintain body contact between the mating partners during spawning, act as weapons for defence of spawning territories, or are sexual signals that reveal aspects of genetic quality. We took two samples of whitefish from their spawning place, one at the beginning and one around the peak of spawning season. We found that females have on average smaller breeding tubercles than males, and that tubercle size partly reveals the stage of gonad maturation. Two independent full-factorial breeding experiments revealed that embryo mortality was significantly influenced by male and female effects. This finding demonstrates that the males differed in their genetic quality (because offspring get nothing but genes from their fathers). Tubercle size was negatively linked to some aspects of embryo mortality in the first breeding experiment but not significantly so in the second. This lack of consistency adds to inconsistent results that were reported before and suggests that (i) some aspects of genetic quality are not revealed in breeding tubercles while others are, or (ii) individuals vary in their signaling strategies and the information content of breeding tubercles is not always reliable. Moreover, the fact that female whitefish have breeding tubercles of significant size while males seem to have few reasons to be choosy suggests that the tubercles might also serve some functions that are not linked to sexual signaling.
Resumo:
Reliable information is a crucial factor influencing decision-making and, thus, fitness in all animals. A common source of information comes from inadvertent cues produced by the behavior of conspecifics. Here we use a system of experimental evolution with robots foraging in an arena containing a food source to study how communication strategies can evolve to regulate information provided by such cues. The robots could produce information by emitting blue light, which the other robots could perceive with their cameras. Over the first few generations, the robots quickly evolved to successfully locate the food, while emitting light randomly. This behavior resulted in a high intensity of light near food, which provided social information allowing other robots to more rapidly find the food. Because robots were competing for food, they were quickly selected to conceal this information. However, they never completely ceased to produce information. Detailed analyses revealed that this somewhat surprising result was due to the strength of selection on suppressing information declining concomitantly with the reduction in information content. Accordingly, a stable equilibrium with low information and considerable variation in communicative behaviors was attained by mutation selection. Because a similar coevolutionary process should be common in natural systems, this may explain why communicative strategies are so variable in many animal species.
Resumo:
Among various advantages, their small size makes model organisms preferred subjects of investigation. Yet, even in model systems detailed analysis of numerous developmental processes at cellular level is severely hampered by their scale. For instance, secondary growth of Arabidopsis hypocotyls creates a radial pattern of highly specialized tissues that comprises several thousand cells starting from a few dozen. This dynamic process is difficult to follow because of its scale and because it can only be investigated invasively, precluding comprehensive understanding of the cell proliferation, differentiation, and patterning events involved. To overcome such limitation, we established an automated quantitative histology approach. We acquired hypocotyl cross-sections from tiled high-resolution images and extracted their information content using custom high-throughput image processing and segmentation. Coupled with automated cell type recognition through machine learning, we could establish a cellular resolution atlas that reveals vascular morphodynamics during secondary growth, for example equidistant phloem pole formation. DOI: http://dx.doi.org/10.7554/eLife.01567.001.
Resumo:
BACKGROUND: Solexa/Illumina short-read ultra-high throughput DNA sequencing technology produces millions of short tags (up to 36 bases) by parallel sequencing-by-synthesis of DNA colonies. The processing and statistical analysis of such high-throughput data poses new challenges; currently a fair proportion of the tags are routinely discarded due to an inability to match them to a reference sequence, thereby reducing the effective throughput of the technology. RESULTS: We propose a novel base calling algorithm using model-based clustering and probability theory to identify ambiguous bases and code them with IUPAC symbols. We also select optimal sub-tags using a score based on information content to remove uncertain bases towards the ends of the reads. CONCLUSION: We show that the method improves genome coverage and number of usable tags as compared with Solexa's data processing pipeline by an average of 15%. An R package is provided which allows fast and accurate base calling of Solexa's fluorescence intensity files and the production of informative diagnostic plots.
Resumo:
Nowadays, the joint exploitation of images acquired daily by remote sensing instruments and of images available from archives allows a detailed monitoring of the transitions occurring at the surface of the Earth. These modifications of the land cover generate spectral discrepancies that can be detected via the analysis of remote sensing images. Independently from the origin of the images and of type of surface change, a correct processing of such data implies the adoption of flexible, robust and possibly nonlinear method, to correctly account for the complex statistical relationships characterizing the pixels of the images. This Thesis deals with the development and the application of advanced statistical methods for multi-temporal optical remote sensing image processing tasks. Three different families of machine learning models have been explored and fundamental solutions for change detection problems are provided. In the first part, change detection with user supervision has been considered. In a first application, a nonlinear classifier has been applied with the intent of precisely delineating flooded regions from a pair of images. In a second case study, the spatial context of each pixel has been injected into another nonlinear classifier to obtain a precise mapping of new urban structures. In both cases, the user provides the classifier with examples of what he believes has changed or not. In the second part, a completely automatic and unsupervised method for precise binary detection of changes has been proposed. The technique allows a very accurate mapping without any user intervention, resulting particularly useful when readiness and reaction times of the system are a crucial constraint. In the third, the problem of statistical distributions shifting between acquisitions is studied. Two approaches to transform the couple of bi-temporal images and reduce their differences unrelated to changes in land cover are studied. The methods align the distributions of the images, so that the pixel-wise comparison could be carried out with higher accuracy. Furthermore, the second method can deal with images from different sensors, no matter the dimensionality of the data nor the spectral information content. This opens the doors to possible solutions for a crucial problem in the field: detecting changes when the images have been acquired by two different sensors.
Resumo:
Résumé : Les mécanismes de contrôle des couleurs mélaniques chez les vertébrés sont encore discutés parmi les biologistes de l'évolution. Une hypothèse récente affirme que les effets pléiotropies du système des mélanocortines expliquent l'association fréquente entre la coloration eumélanique noire (due à la déposition d'eumélanine) et de nombreux traits physiologiques et comportementaux. De nombreuses études suggèrent, en effet, que des niveaux plus élevés des mélanocortines induisent l'assombrissement des téguments eumélaniques et affectent d'autres traits phénotypiques simultanément. Cependant, il n'est pas encore établi si ce mécanisme de pléiotropie peut s'appliquer aux colorations dues à la déposition de phaeomélanine, une autre forme commune de mélanine. Les antagonistes des mélanocortines déclenchent le phaeomélanogenèse et bloquent l'effet des mélanocortines ou ont un effet pharmacologique opposé. Nous nous proposons donc d'évaluer l'hypothèse que les effets pléiotropes des antagonistes des mélanocortines génèrent des covariations entre la coloration phaeomélanique et des aspects de la qualité individuelle. Comme prédit par cette hypothèse, nous constatons chez la chouette effraie (Tyto alba) que les traits phénotypiques (résistance au stress oxydatif et aux parasites) corrèlent positivement au degré d'expression d'une couleur eumélanique mais négativement au degré d'expression d'une coloration phaeomélanique. Puis, nous montrons chez la chouette hulotte (Strix aluco) que les associations génétiques entre la coloration phaeomélanique et la physiologie (immunité et la régulation de l'homéostasie) confèrent des avantages aux individus de différentes couleurs dans différents environnements caractérisés par l'abondance de nourriture et le niveau d'exposition aux parasites. Ainsi, nos études soutiennent l'hypothèse que les effets pléiotropes des antagonistes des mélanocortines génèrent des covariations entre les traits mélaniques et divers aspects de la qualité individuelle. Finalement, nous montrons chez le faucon crécerelle (Falco Tinnunculus) que l'expression des ornements mélaniques est sensible à la qualité de l'environnement dans lequel les individus grandissent. Ceci suggère que les gènes codant pour les mélanocortines et leurs antagonistes pourraient induire une expression des traits mélaniques dépendante de la condition de l'individu, un pattern d'expression rarement observé pour des traits généralement sous fort contrôle génétique. Summary : The information content and control mechanisms of melanin-based colour signals in vertebrates are still debated among evolutionary biologists. A recent hypothesis contends that pleiotropic effects of the melanocortin system accounts for the frequent association between black eumelanic coloration and physiological and behavioural traits. Accordingly, empirical evidence suggests that higher levels of melanocortins concurrently promote darker eumelanic integuments and affect other phenotypic traits. However, whether this mechanism may apply to signals relying on phaeomelanin, another common form of melanin pigments, remains to be established. Melanocortin antagonists trigger phaeomelanogenesis and block the effect of melanocortins or result in the opposite pharmacological effect. Therefore, we tested the hypothesis that pleiotropic effects of melanocortin antagonists and inverse agonists account for covariations between phaeomelanin-based coloration and aspects of individual quality. As predicted, we found that phenotypic traits (resistance to oxidative stress and parasites) correlated positively with a eumelanic trait and negatively with a phaeomelanic trait in the barn owl (Tyto alba). Then, we showed in the tawny owl (Strix aluco) that genetic associations between phaeomelanin-based coloration and physiology (immunity and regulation of energy homeostasis) confer benefits to differently coloured individuals under different levels of food abundance and parasite exposure. Altogether, our studies support the hypothesis that pleiotropic effects of melanocortins antagonists can indeed account for covariations between phaeomelanin-based traits and aspects of individual quality. Eventually, we show in the Eurasian kestrel (Falco Tinnunculus) that expression of melanin-based ornaments is sensitive to the environment in which individuals grow. This suggests that genes coding for melanocortins and their antagonists can mediate the condition-dependent component of melanin-based traits.
Resumo:
Closely related species may be very difficult to distinguish morphologically, yet sometimes morphology is the only reasonable possibility for taxonomic classification. Here we present learning-vector-quantization artificial neural networks as a powerful tool to classify specimens on the basis of geometric morphometric shape measurements. As an example, we trained a neural network to distinguish between field and root voles from Procrustes transformed landmark coordinates on the dorsal side of the skull, which is so similar in these two species that the human eye cannot make this distinction. Properly trained neural networks misclassified only 3% of specimens. Therefore, we conclude that the capacity of learning vector quantization neural networks to analyse spatial coordinates is a powerful tool among the range of pattern recognition procedures that is available to employ the information content of geometric morphometrics.
Genetic Variations and Diseases in UniProtKB/Swiss-Prot: The Ins and Outs of Expert Manual Curation.
Resumo:
During the last few years, next-generation sequencing (NGS) technologies have accelerated the detection of genetic variants resulting in the rapid discovery of new disease-associated genes. However, the wealth of variation data made available by NGS alone is not sufficient to understand the mechanisms underlying disease pathogenesis and manifestation. Multidisciplinary approaches combining sequence and clinical data with prior biological knowledge are needed to unravel the role of genetic variants in human health and disease. In this context, it is crucial that these data are linked, organized, and made readily available through reliable online resources. The Swiss-Prot section of the Universal Protein Knowledgebase (UniProtKB/Swiss-Prot) provides the scientific community with a collection of information on protein functions, interactions, biological pathways, as well as human genetic diseases and variants, all manually reviewed by experts. In this article, we present an overview of the information content of UniProtKB/Swiss-Prot to show how this knowledgebase can support researchers in the elucidation of the mechanisms leading from a molecular defect to a disease phenotype.
Resumo:
OBJECTIVES: The purpose of this study was to evaluate the association between inflammation and heart failure (HF) risk in older adults. BACKGROUND: Inflammation is associated with HF risk factors and also directly affects myocardial function. METHODS: The association of baseline serum concentrations of interleukin (IL)-6, tumor necrosis factor-alpha, and C-reactive protein (CRP) with incident HF was assessed with Cox models among 2,610 older persons without prevalent HF enrolled in the Health ABC (Health, Aging, and Body Composition) study (age 73.6 +/- 2.9 years; 48.3% men; 59.6% white). RESULTS: During follow-up (median 9.4 years), HF developed in 311 (11.9%) participants. In models controlling for clinical characteristics, ankle-arm index, and incident coronary heart disease, doubling of IL-6, tumor necrosis factor-alpha, and CRP concentrations was associated with 29% (95% confidence interval: 13% to 47%; p < 0.001), 46% (95% confidence interval: 17% to 84%; p = 0.001), and 9% (95% confidence interval: -1% to 24%; p = 0.087) increase in HF risk, respectively. In models including all 3 markers, IL-6, and tumor necrosis factor-alpha, but not CRP, remained significant. These associations were similar across sex and race and persisted in models accounting for death as a competing event. Post-HF ejection fraction was available in 239 (76.8%) cases; inflammatory markers had stronger association with HF with preserved ejection fraction. Repeat IL-6 and CRP determinations at 1-year follow-up did not provide incremental information. Addition of IL-6 to the clinical Health ABC HF model improved model discrimination (C index from 0.717 to 0.734; p = 0.001) and fit (decreased Bayes information criterion by 17.8; p < 0.001). CONCLUSIONS: Inflammatory markers are associated with HF risk among older adults and may improve HF risk stratification.
Resumo:
Il est important pour les entreprises de compresser les informations détaillées dans des sets d'information plus compréhensibles. Au chapitre 1, je résume et structure la littérature sur le sujet « agrégation d'informations » en contrôle de gestion. Je récapitule l'analyse coûts-bénéfices que les comptables internes doivent considérer quand ils décident des niveaux optimaux d'agrégation d'informations. Au-delà de la perspective fondamentale du contenu d'information, les entreprises doivent aussi prendre en considération des perspectives cogni- tives et comportementales. Je développe ces aspects en faisant la part entre la comptabilité analytique, les budgets et plans, et la mesure de la performance. Au chapitre 2, je focalise sur un biais spécifique qui se crée lorsque les informations incertaines sont agrégées. Pour les budgets et plans, des entreprises doivent estimer les espérances des coûts et des durées des projets, car l'espérance est la seule mesure de tendance centrale qui est linéaire. A la différence de l'espérance, des mesures comme le mode ou la médiane ne peuvent pas être simplement additionnés. En considérant la forme spécifique de distributions des coûts et des durées, l'addition des modes ou des médianes résultera en une sous-estimation. Par le biais de deux expériences, je remarque que les participants tendent à estimer le mode au lieu de l'espérance résultant en une distorsion énorme de l'estimati¬on des coûts et des durées des projets. Je présente également une stratégie afin d'atténuer partiellement ce biais. Au chapitre 3, j'effectue une étude expérimentale pour comparer deux approches d'esti¬mation du temps qui sont utilisées en comptabilité analytique, spécifiquement « coûts basés sur les activités (ABC) traditionnelles » et « time driven ABC » (TD-ABC). Au contraire des affirmations soutenues par les défenseurs de l'approche TD-ABC, je constate que cette dernière n'est pas nécessairement appropriée pour les calculs de capacité. Par contre, je démontre que le TD-ABC est plus approprié pour les allocations de coûts que l'approche ABC traditionnelle. - It is essential for organizations to compress detailed sets of information into more comprehensi¬ve sets, thereby, establishing sharp data compression and good decision-making. In chapter 1, I review and structure the literature on information aggregation in management accounting research. I outline the cost-benefit trade-off that management accountants need to consider when they decide on the optimal levels of information aggregation. Beyond the fundamental information content perspective, organizations also have to account for cognitive and behavi¬oral perspectives. I elaborate on these aspects differentiating between research in cost accounti¬ng, budgeting and planning, and performance measurement. In chapter 2, I focus on a specific bias that arises when probabilistic information is aggregated. In budgeting and planning, for example, organizations need to estimate mean costs and durations of projects, as the mean is the only measure of central tendency that is linear. Different from the mean, measures such as the mode or median cannot simply be added up. Given the specific shape of cost and duration distributions, estimating mode or median values will result in underestimations of total project costs and durations. In two experiments, I find that participants tend to estimate mode values rather than mean values resulting in large distortions of estimates for total project costs and durations. I also provide a strategy that partly mitigates this bias. In the third chapter, I conduct an experimental study to compare two approaches to time estimation for cost accounting, i.e., traditional activity-based costing (ABC) and time-driven ABC (TD-ABC). Contrary to claims made by proponents of TD-ABC, I find that TD-ABC is not necessarily suitable for capacity computations. However, I also provide evidence that TD-ABC seems better suitable for cost allocations than traditional ABC.
Resumo:
The GO annotation dataset provided by the UniProt Consortium (GOA: http://www.ebi.ac.uk/GOA) is a comprehensive set of evidenced-based associations between terms from the Gene Ontology resource and UniProtKB proteins. Currently supplying over 100 million annotations to 11 million proteins in more than 360,000 taxa, this resource has increased 2-fold over the last 2 years and has benefited from a wealth of checks to improve annotation correctness and consistency as well as now supplying a greater information content enabled by GO Consortium annotation format developments. Detailed, manual GO annotations obtained from the curation of peer-reviewed papers are directly contributed by all UniProt curators and supplemented with manual and electronic annotations from 36 model organism and domain-focused scientific resources. The inclusion of high-quality, automatic annotation predictions ensures the UniProt GO annotation dataset supplies functional information to a wide range of proteins, including those from poorly characterized, non-model organism species. UniProt GO annotations are freely available in a range of formats accessible by both file downloads and web-based views. In addition, the introduction of a new, normalized file format in 2010 has made for easier handling of the complete UniProt-GOA data set.
Resumo:
In Neo-Darwinism, variation and natural selection are the two evolutionary mechanisms which propel biological evolution. Our previous article presented a histogram model [1] consisting in populations of individuals whose number changed under the influence of variation and/or fitness, the total population remaining constant. Individuals are classified into bins, and the content of each bin is calculated generation after generation by an Excel spreadsheet. Here, we apply the histogram model to a stable population with fitness F(1)=1.00 in which one or two fitter mutants emerge. In a first scenario, a single mutant emerged in the population whose fitness was greater than 1.00. The simulations ended when the original population was reduced to a single individual. The histogram model was validated by excellent agreement between its predictions and those of a classical continuous function (Eqn. 1) which predicts the number of generations needed for a favorable mutation to spread throughout a population. But in contrast to Eqn. 1, our histogram model is adaptable to more complex scenarios, as demonstrated here. In the second and third scenarios, the original population was present at time zero together with two mutants which differed from the original population by two higher and distinct fitness values. In the fourth scenario, the large original population was present at time zero together with one fitter mutant. After a number of generations, when the mutant offspring had multiplied, a second mutant was introduced whose fitness was even greater. The histogram model also allows Shannon entropy (SE) to be monitored continuously as the information content of the total population decreases or increases. The results of these simulations illustrate, in a graphically didactic manner, the influence of natural selection, operating through relative fitness, in the emergence and dominance of a fitter mutant.
Resumo:
In Neo-Darwinism, variation and natural selection are the two evolutionary mechanisms which propel biological evolution. Our previous reports presented a histogram model to simulate the evolution of populations of individuals classified into bins according to an unspecified, quantifiable phenotypic character, and whose number in each bin changed generation after generation under the influence of fitness, while the total population was maintained constant. The histogram model also allowed Shannon entropy (SE) to be monitored continuously as the information content of the total population decreased or increased. Here, a simple Perl (Practical Extraction and Reporting Language) application was developed to carry out these computations, with the critical feature of an added random factor in the percent of individuals whose offspring moved to a vicinal bin. The results of the simulations demonstrate that the random factor mimicking variation increased considerably the range of values covered by Shannon entropy, especially when the percentage of changed offspring was high. This increase in information content is interpreted as facilitated adaptability of the population.