150 resultados para Selection models
Resumo:
The structure of laser glasses in the system (Y(2)O(3))(0.2){(Al(2)O(3))(x))(B(2)O(3))(0.8-x)} (0.15 <= x <= 0.40) has been investigated by means of (11)B, (27)Al, and (89)Y solid state NMR as well as electron spin echo envelope modulation (ESEEM) of Yb-doped samples. The latter technique has been applied for the first time to an aluminoborate glass system. (11)B magic-angle spinning (MAS)-NMR spectra reveal that, while the majority of the boron atoms are three-coordinated over the entire composition region, the fraction of three-coordinated boron atoms increases significantly with increasing x. Charge balance considerations as well as (11)B NMR lineshape analyses suggest that the dominant borate species are predominantly singly charged metaborate (BO(2/2)O(-)), doubly charged pyroborate (BO(1/2)(O(-))(2)), and (at x = 0.40) triply charged orthoborate groups. As x increases along this series, the average anionic charge per trigonal borate group increases from 1.38 to 2.91. (27)Al MAS-NMR spectra show that the alumina species are present in the coordination states four, five and six, and the fraction of four-coordinated Al increases markedly with increasing x. All of the Al coordination states are in intimate contact with both the three-and the four-coordinate boron species and vice versa, as indicated by (11)B/(27)Al rotational echo double resonance (REDOR) data. These results are consistent with the formation of a homogeneous, non-segregated glass structure. (89)Y solid state NMR spectra show a significant chemical shift trend, reflecting that the second coordination sphere becomes increasingly ""aluminate-like'' with increasing x. This conclusion is supported by electron spin echo envelope modulation (ESEEM) data of Yb-doped glasses, which indicate that both borate and aluminate species participate in the medium range structure of the rare-earth ions, consistent with a random spatial distribution of the glass components.
Resumo:
The parallel mutation-selection evolutionary dynamics, in which mutation and replication are independent events, is solved exactly in the case that the Malthusian fitnesses associated to the genomes are described by the random energy model (REM) and by a ferromagnetic version of the REM. The solution method uses the mapping of the evolutionary dynamics into a quantum Ising chain in a transverse field and the Suzuki-Trotter formalism to calculate the transition probabilities between configurations at different times. We find that in the case of the REM landscape the dynamics can exhibit three distinct regimes: pure diffusion or stasis for short times, depending on the fitness of the initial configuration, and a spin-glass regime for large times. The dynamic transition between these dynamical regimes is marked by discontinuities in the mean-fitness as well as in the overlap with the initial reference sequence. The relaxation to equilibrium is described by an inverse time decay. In the ferromagnetic REM, we find in addition to these three regimes, a ferromagnetic regime where the overlap and the mean-fitness are frozen. In this case, the system relaxes to equilibrium in a finite time. The relevance of our results to information processing aspects of evolution is discussed.
Resumo:
Thanks to recent advances in molecular biology, allied to an ever increasing amount of experimental data, the functional state of thousands of genes can now be extracted simultaneously by using methods such as cDNA microarrays and RNA-Seq. Particularly important related investigations are the modeling and identification of gene regulatory networks from expression data sets. Such a knowledge is fundamental for many applications, such as disease treatment, therapeutic intervention strategies and drugs design, as well as for planning high-throughput new experiments. Methods have been developed for gene networks modeling and identification from expression profiles. However, an important open problem regards how to validate such approaches and its results. This work presents an objective approach for validation of gene network modeling and identification which comprises the following three main aspects: (1) Artificial Gene Networks (AGNs) model generation through theoretical models of complex networks, which is used to simulate temporal expression data; (2) a computational method for gene network identification from the simulated data, which is founded on a feature selection approach where a target gene is fixed and the expression profile is observed for all other genes in order to identify a relevant subset of predictors; and (3) validation of the identified AGN-based network through comparison with the original network. The proposed framework allows several types of AGNs to be generated and used in order to simulate temporal expression data. The results of the network identification method can then be compared to the original network in order to estimate its properties and accuracy. Some of the most important theoretical models of complex networks have been assessed: the uniformly-random Erdos-Renyi (ER), the small-world Watts-Strogatz (WS), the scale-free Barabasi-Albert (BA), and geographical networks (GG). The experimental results indicate that the inference method was sensitive to average degree k variation, decreasing its network recovery rate with the increase of k. The signal size was important for the inference method to get better accuracy in the network identification rate, presenting very good results with small expression profiles. However, the adopted inference method was not sensible to recognize distinct structures of interaction among genes, presenting a similar behavior when applied to different network topologies. In summary, the proposed framework, though simple, was adequate for the validation of the inferred networks by identifying some properties of the evaluated method, which can be extended to other inference methods.
Resumo:
Background: The inference of gene regulatory networks (GRNs) from large-scale expression profiles is one of the most challenging problems of Systems Biology nowadays. Many techniques and models have been proposed for this task. However, it is not generally possible to recover the original topology with great accuracy, mainly due to the short time series data in face of the high complexity of the networks and the intrinsic noise of the expression measurements. In order to improve the accuracy of GRNs inference methods based on entropy (mutual information), a new criterion function is here proposed. Results: In this paper we introduce the use of generalized entropy proposed by Tsallis, for the inference of GRNs from time series expression profiles. The inference process is based on a feature selection approach and the conditional entropy is applied as criterion function. In order to assess the proposed methodology, the algorithm is applied to recover the network topology from temporal expressions generated by an artificial gene network (AGN) model as well as from the DREAM challenge. The adopted AGN is based on theoretical models of complex networks and its gene transference function is obtained from random drawing on the set of possible Boolean functions, thus creating its dynamics. On the other hand, DREAM time series data presents variation of network size and its topologies are based on real networks. The dynamics are generated by continuous differential equations with noise and perturbation. By adopting both data sources, it is possible to estimate the average quality of the inference with respect to different network topologies, transfer functions and network sizes. Conclusions: A remarkable improvement of accuracy was observed in the experimental results by reducing the number of false connections in the inferred topology by the non-Shannon entropy. The obtained best free parameter of the Tsallis entropy was on average in the range 2.5 <= q <= 3.5 (hence, subextensive entropy), which opens new perspectives for GRNs inference methods based on information theory and for investigation of the nonextensivity of such networks. The inference algorithm and criterion function proposed here were implemented and included in the DimReduction software, which is freely available at http://sourceforge.net/projects/dimreduction and http://code.google.com/p/dimreduction/.
Resumo:
Background: Feature selection is a pattern recognition approach to choose important variables according to some criteria in order to distinguish or explain certain phenomena (i.e., for dimensionality reduction). There are many genomic and proteomic applications that rely on feature selection to answer questions such as selecting signature genes which are informative about some biological state, e. g., normal tissues and several types of cancer; or inferring a prediction network among elements such as genes, proteins and external stimuli. In these applications, a recurrent problem is the lack of samples to perform an adequate estimate of the joint probabilities between element states. A myriad of feature selection algorithms and criterion functions have been proposed, although it is difficult to point the best solution for each application. Results: The intent of this work is to provide an open-source multiplataform graphical environment for bioinformatics problems, which supports many feature selection algorithms, criterion functions and graphic visualization tools such as scatterplots, parallel coordinates and graphs. A feature selection approach for growing genetic networks from seed genes ( targets or predictors) is also implemented in the system. Conclusion: The proposed feature selection environment allows data analysis using several algorithms, criterion functions and graphic visualization tools. Our experiments have shown the software effectiveness in two distinct types of biological problems. Besides, the environment can be used in different pattern recognition applications, although the main concern regards bioinformatics tasks.
Resumo:
We consider the problem of interaction neighborhood estimation from the partial observation of a finite number of realizations of a random field. We introduce a model selection rule to choose estimators of conditional probabilities among natural candidates. Our main result is an oracle inequality satisfied by the resulting estimator. We use then this selection rule in a two-step procedure to evaluate the interacting neighborhoods. The selection rule selects a small prior set of possible interacting points and a cutting step remove from this prior set the irrelevant points. We also prove that the Ising models satisfy the assumptions of the main theorems, without restrictions on the temperature, on the structure of the interacting graph or on the range of the interactions. It provides therefore a large class of applications for our results. We give a computationally efficient procedure in these models. We finally show the practical efficiency of our approach in a simulation study.
Resumo:
The application of laser induced breakdown spectrometry (LIBS) aiming the direct analysis of plant materials is a great challenge that still needs efforts for its development and validation. In this way, a series of experimental approaches has been carried out in order to show that LIBS can be used as an alternative method to wet acid digestions based methods for analysis of agricultural and environmental samples. The large amount of information provided by LIBS spectra for these complex samples increases the difficulties for selecting the most appropriated wavelengths for each analyte. Some applications have suggested that improvements in both accuracy and precision can be achieved by the application of multivariate calibration in LIBS data when compared to the univariate regression developed with line emission intensities. In the present work, the performance of univariate and multivariate calibration, based on partial least squares regression (PLSR), was compared for analysis of pellets of plant materials made from an appropriate mixture of cryogenically ground samples with cellulose as the binding agent. The development of a specific PLSR model for each analyte and the selection of spectral regions containing only lines of the analyte of interest were the best conditions for the analysis. In this particular application, these models showed a similar performance. but PLSR seemed to be more robust due to a lower occurrence of outliers in comparison to the univariate method. Data suggests that efforts dealing with sample presentation and fitness of standards for LIBS analysis must be done in order to fulfill the boundary conditions for matrix independent development and validation. (C) 2009 Elsevier B.V. All rights reserved.
Resumo:
Objective: We carry out a systematic assessment on a suite of kernel-based learning machines while coping with the task of epilepsy diagnosis through automatic electroencephalogram (EEG) signal classification. Methods and materials: The kernel machines investigated include the standard support vector machine (SVM), the least squares SVM, the Lagrangian SVM, the smooth SVM, the proximal SVM, and the relevance vector machine. An extensive series of experiments was conducted on publicly available data, whose clinical EEG recordings were obtained from five normal subjects and five epileptic patients. The performance levels delivered by the different kernel machines are contrasted in terms of the criteria of predictive accuracy, sensitivity to the kernel function/parameter value, and sensitivity to the type of features extracted from the signal. For this purpose, 26 values for the kernel parameter (radius) of two well-known kernel functions (namely. Gaussian and exponential radial basis functions) were considered as well as 21 types of features extracted from the EEG signal, including statistical values derived from the discrete wavelet transform, Lyapunov exponents, and combinations thereof. Results: We first quantitatively assess the impact of the choice of the wavelet basis on the quality of the features extracted. Four wavelet basis functions were considered in this study. Then, we provide the average accuracy (i.e., cross-validation error) values delivered by 252 kernel machine configurations; in particular, 40%/35% of the best-calibrated models of the standard and least squares SVMs reached 100% accuracy rate for the two kernel functions considered. Moreover, we show the sensitivity profiles exhibited by a large sample of the configurations whereby one can visually inspect their levels of sensitiveness to the type of feature and to the kernel function/parameter value. Conclusions: Overall, the results evidence that all kernel machines are competitive in terms of accuracy, with the standard and least squares SVMs prevailing more consistently. Moreover, the choice of the kernel function and parameter value as well as the choice of the feature extractor are critical decisions to be taken, albeit the choice of the wavelet family seems not to be so relevant. Also, the statistical values calculated over the Lyapunov exponents were good sources of signal representation, but not as informative as their wavelet counterparts. Finally, a typical sensitivity profile has emerged among all types of machines, involving some regions of stability separated by zones of sharp variation, with some kernel parameter values clearly associated with better accuracy rates (zones of optimality). (C) 2011 Elsevier B.V. All rights reserved.
Resumo:
Here, I investigate the use of Bayesian updating rules applied to modeling how social agents change their minds in the case of continuous opinion models. Given another agent statement about the continuous value of a variable, we will see that interesting dynamics emerge when an agent assigns a likelihood to that value that is a mixture of a Gaussian and a uniform distribution. This represents the idea that the other agent might have no idea about what is being talked about. The effect of updating only the first moments of the distribution will be studied, and we will see that this generates results similar to those of the bounded confidence models. On also updating the second moment, several different opinions always survive in the long run, as agents become more stubborn with time. However, depending on the probability of error and initial uncertainty, those opinions might be clustered around a central value.
Resumo:
Today several different unsupervised classification algorithms are commonly used to cluster similar patterns in a data set based only on its statistical properties. Specially in image data applications, self-organizing methods for unsupervised classification have been successfully applied for clustering pixels or group of pixels in order to perform segmentation tasks. The first important contribution of this paper refers to the development of a self-organizing method for data classification, named Enhanced Independent Component Analysis Mixture Model (EICAMM), which was built by proposing some modifications in the Independent Component Analysis Mixture Model (ICAMM). Such improvements were proposed by considering some of the model limitations as well as by analyzing how it should be improved in order to become more efficient. Moreover, a pre-processing methodology was also proposed, which is based on combining the Sparse Code Shrinkage (SCS) for image denoising and the Sobel edge detector. In the experiments of this work, the EICAMM and other self-organizing models were applied for segmenting images in their original and pre-processed versions. A comparative analysis showed satisfactory and competitive image segmentation results obtained by the proposals presented herein. (C) 2008 Published by Elsevier B.V.
Resumo:
The aims of the present study were to compare the effects of two periodization models on metabolic syndrome risk factors in obese adolescents and verify whether the angiotensin-converting enzyme (ACE) genotype is important in establishing these effects. A total of 32 postpuberty obese adolescents were submitted to aerobic training (AT) and resistance training (RT) for 14 weeks. The subjects were divided into linear periodization (LP, n = 16) or daily undulating periodization (DUP, n = 16). Body composition, visceral and subcutaneous fat, glycemia, insulinemia, homeostasis model assessment of insulin resistance (HOMA-IR), lipid profiles, blood pressure, maximal oxygen consumption (VO(2max)), resting metabolic rate (RMR), muscular endurance were analyzed at baseline and after intervention. Both groups demonstrated a significant reduction in body mass, BMI, body fat, visceral and subcutaneous fat, total and low-density lipoprotein cholesterol, blood pressure and an increase in fat-free mass, VO(2max), and muscular endurance. However, only DUP promoted a reduction in insulin concentrations and HOMA-IR. It is important to emphasize that there was no statics difference between LP and DUP groups; however, it appears that there may be bigger changes in the DUP than LP group in some of the metabolic syndrome risk factors in obese adolescents with regard to the effect size (ES). Both periodization models presented a large effect on muscular endurance. Despite the limitation of sample size, our results suggested that the ACE genotype may influence the functional and metabolic characteristics of obese adolescents and may be considered in the future strategies for massive obesity control.
Resumo:
Background: Leptin-deficient mice (Lep(ob)/Lep(ob), also known as ob/ob) are of great importance for studies of obesity, diabetes and other correlated pathologies. Thus, generation of animals carrying the Lep(ob) gene mutation as well as additional genomic modifications has been used to associate genes with metabolic diseases. However, the infertility of Lep(ob)/Lep(ob) mice impairs this kind of breeding experiment. Objective: To propose a new method for production of Lep(ob)/Lep(ob) animals and Lep(ob)/Lep(ob)-derived animal models by restoring the fertility of Lep(ob)/Lep(ob) mice in a stable way through white adipose tissue transplantations. Methods: For this purpose, 1 g of peri-gonadal adipose tissue from lean donors was used in subcutaneous transplantations of Lep(ob)/Lep(ob) animals and a crossing strategy was established to generate Lep(ob)/Lep(ob)-derived mice. Results: The presented method reduced by four times the number of animals used to generate double transgenic models (from about 20 to 5 animals per double mutant produced) and minimized the number of genotyping steps (from 3 to 1 genotyping step, reducing the number of Lep gene genotyping assays from 83 to 6). Conclusion: The application of the adipose transplantation technique drastically improves both the production of Lep(ob)/Lep(ob) animals and the generation of Lep(ob)/Lep(ob)-derived animal models. International Journal of Obesity (2009) 33, 938-944; doi: 10.1038/ijo.2009.95; published online 16 June 2009
Anthropometric characteristics and motor skills in talent selection and development in indoor soccer
Resumo:
Kick performance, anthropometric characteristics, slalom, and linear running were assessed in 49 (24 elite, 25 nonelite) postpubertal indoor soccer players in order to (a) verify whether anthropometric characteristics and physical and technical capacities can distinguish players of different competitive levels, (b) compare the kicking kinematics of these groups, with and without a defined target, and (c) compare results on the assessments and coaches` subjective rankings of the players. Thigh circumference and specific technical capacities differentiated the players by level of play; cluster analysis correctly classified 77.5% of the players. The correlation between players` standardized measures and the coaches` rankings was 0.29. Anthropometric characteristics and physical capacities do not necessarily differentiate players at post-pubertal stages and should not be overvalued during early development. Considering the coaches` rankings, performance measures outside the specific game conditions may not be useful in identification of talented players.
Resumo:
To evaluate the potential for fermentation of raspberry pulp, sixteen yeast strains (S. cerevisiae and S. bayanus) were studied. Volatile compounds were determined by GC-MS, GC-FID, and GC-PFPD. Ethanol. glycerol and organic acids were determined by HPLC. HPLC-DAD was used to analyse phenolic acids. Sensory analysis was performed by trained panellists. After a screening step, CAT-1, UFLA FW 15 and S. bayanus CBS 1505 were previously selected based on their fermentative characteristics and profile of the metabolites identified. The beverage produced with CAT-1 showed the highest volatile fatty acid concentration (1542.6 mu g/L), whereas the beverage produced with UFLA FIN 15 showed the highest concentration of acetates (2211.1 mu g/L) and total volatile compounds (5835 mu g/L). For volatile sulphur compounds. 566.5 mu g/L were found in the beverage produced with S. bayanus CBS 1505. The lowest concentration of volatile sulphur compounds (151.9 mu g/L) was found for the beverage produced with UFLA FW 15. In the sensory analysis, the beverage produced with UFLA FW 15 was characterised by the descriptors raspberry, cherry, sweet, strawberry, floral and violet. In conclusion, strain UFLA FW 15 was the yeast that produced a raspberry wine with a good chemical and sensory quality. (C) 2010 Elsevier Ltd. All rights reserved.
Resumo:
A combination of an extension of the topological instability ""lambda criterion"" and a thermodynamic criterion were applied to the Al-La system, indicating the best range of compositions for glass formation. Alloy compositions in this range were prepared by melt-spinning and casting in an arc-melting furnace with a wedge-section copper mold. The GFA of these samples was evaluated by X-ray diffraction, differential scanning calorimetry and scanning electron microscopy. The results indicated that the gamma* parameter of compositions with high GFA is higher, corresponding to a range in which the lambda parameter is greater than 0.1, which are compositions far from Al solid solution. A new alloy was identified with the best GFA reported so far for this system, showing a maximum thickness of 286 mu m in a wedge-section copper mold. Crown Copyright (C) 2009 Published by Elsevier B.V. All rights reserved.