960 resultados para Maximum-entropy selection criterion
Resumo:
In order to select superior hybrids for the concentration of favorable alleles for resistance to papaya black spot, powdery mildew and phoma spot, 67 hybrids were evaluated in two seasons, in 2007, in a randomized block design with two replications. Genetic gains were estimated from the selection indices of Smith & Hazel, Pesek & Baker, Williams, Mulamba & Mock, with selection intensity of 22.39%, corresponding to 15 hybrids. The index of Mulamba & Mock showed gains more suitable for the five traits assessed when it was used the criterion of economic weight tentatively assigned. Together, severity of black spot on leaves and on fruits, characteristics considered most relevant to the selection of resistant materials, expressed percentage gain of -44.15%. In addition, there were gains for other characteristics, with negative predicted selective percentage gain. The results showed that the index of Mulamba & Mock is the most efficient procedure for simultaneous selection of papaya hybrid resistant to black spot, powdery mildew and phoma spot.
Resumo:
Research on cluster analysis for categorical data continues to develop, new clustering algorithms being proposed. However, in this context, the determination of the number of clusters is rarely addressed. We propose a new approach in which clustering and the estimation of the number of clusters is done simultaneously for categorical data. We assume that the data originate from a finite mixture of multinomial distributions and use a minimum message length criterion (MML) to select the number of clusters (Wallace and Bolton, 1986). For this purpose, we implement an EM-type algorithm (Silvestre et al., 2008) based on the (Figueiredo and Jain, 2002) approach. The novelty of the approach rests on the integration of the model estimation and selection of the number of clusters in a single algorithm, rather than selecting this number based on a set of pre-estimated candidate models. The performance of our approach is compared with the use of Bayesian Information Criterion (BIC) (Schwarz, 1978) and Integrated Completed Likelihood (ICL) (Biernacki et al., 2000) using synthetic data. The obtained results illustrate the capacity of the proposed algorithm to attain the true number of cluster while outperforming BIC and ICL since it is faster, which is especially relevant when dealing with large data sets.
Resumo:
Background: Several researchers seek methods for the selection of homogeneous groups of animals in experimental studies, a fact justified because homogeneity is an indispensable prerequisite for casualization of treatments. The lack of robust methods that comply with statistical and biological principles is the reason why researchers use empirical or subjective methods, influencing their results. Objective: To develop a multivariate statistical model for the selection of a homogeneous group of animals for experimental research and to elaborate a computational package to use it. Methods: The set of echocardiographic data of 115 male Wistar rats with supravalvular aortic stenosis (AoS) was used as an example of model development. Initially, the data were standardized, and became dimensionless. Then, the variance matrix of the set was submitted to principal components analysis (PCA), aiming at reducing the parametric space and at retaining the relevant variability. That technique established a new Cartesian system into which the animals were allocated, and finally the confidence region (ellipsoid) was built for the profile of the animals’ homogeneous responses. The animals located inside the ellipsoid were considered as belonging to the homogeneous batch; those outside the ellipsoid were considered spurious. Results: The PCA established eight descriptive axes that represented the accumulated variance of the data set in 88.71%. The allocation of the animals in the new system and the construction of the confidence region revealed six spurious animals as compared to the homogeneous batch of 109 animals. Conclusion: The biometric criterion presented proved to be effective, because it considers the animal as a whole, analyzing jointly all parameters measured, in addition to having a small discard rate.
Resumo:
This comment corrects the errors in the estimation process that appear in Martins (2001). The first error is in the parametric probit estimation, as the previously presented results do not maximize the log-likelihood function. In the global maximum more variables become significant. As for the semiparametric estimation method, the kernel function used in Martins (2001) can take on both positive and negative values, which implies that the participation probability estimates may be outside the interval [0,1]. We have solved the problem by applying local smoothing in the kernel estimation, as suggested by Klein and Spady (1993).
Resumo:
This paper studies collective choice rules whose outcomes consist of a collection of simultaneous decisions, each one of which is the only concern of some group of individuals in society. The need for such rules arises in different contexts, including the establishment of jurisdictions, the location of multiple public facilities, or the election of representative committees. We define a notion of allocation consistency requiring that each partial aspect of the global decision taken by society as a whole should be ratified by the group of agents who are directly concerned with this particular aspect. We investigate the possibility of designing envy-free allocation consistent rules, we also explore whether such rules may also respect the Condorcet criterion.
Resumo:
We extend PML theory to account for information on the conditional moments up to order four, but without assuming a parametric model, to avoid a risk of misspecification of the conditional distribution. The key statistical tool is the quartic exponential family, which allows us to generalize the PML2 and QGPML1 methods proposed in Gourieroux et al. (1984) to PML4 and QGPML2 methods, respectively. An asymptotic theory is developed. The key numerical tool that we use is the Gauss-Freud integration scheme that solves a computational problem that has previously been raised in several fields. Simulation exercises demonstrate the feasibility and robustness of the methods [Authors]
Resumo:
We study model selection strategies based on penalized empirical loss minimization. We point out a tight relationship between error estimation and data-based complexity penalization: any good error estimate may be converted into a data-based penalty function and the performance of the estimate is governed by the quality of the error estimate. We consider several penalty functions, involving error estimates on independent test data, empirical {\sc vc} dimension, empirical {\sc vc} entropy, andmargin-based quantities. We also consider the maximal difference between the error on the first half of the training data and the second half, and the expected maximal discrepancy, a closely related capacity estimate that can be calculated by Monte Carlo integration. Maximal discrepancy penalty functions are appealing for pattern classification problems, since their computation is equivalent to empirical risk minimization over the training data with some labels flipped.
Resumo:
We present a non-equilibrium theory in a system with heat and radiative fluxes. The obtained expression for the entropy production is applied to a simple one-dimensional climate model based on the first law of thermodynamics. In the model, the dissipative fluxes are assumed to be independent variables, following the criteria of the Extended Irreversible Thermodynamics (BIT) that enlarges, in reference to the classical expression, the applicability of a macroscopic thermodynamic theory for systems far from equilibrium. We analyze the second differential of the classical and the generalized entropy as a criteria of stability of the steady states. Finally, the extreme state is obtained using variational techniques and observing that the system is close to the maximum dissipation rate
Resumo:
The development and tests of an iterative reconstruction algorithm for emission tomography based on Bayesian statistical concepts are described. The algorithm uses the entropy of the generated image as a prior distribution, can be accelerated by the choice of an exponent, and converges uniformly to feasible images by the choice of one adjustable parameter. A feasible image has been defined as one that is consistent with the initial data (i.e. it is an image that, if truly a source of radiation in a patient, could have generated the initial data by the Poisson process that governs radioactive disintegration). The fundamental ideas of Bayesian reconstruction are discussed, along with the use of an entropy prior with an adjustable contrast parameter, the use of likelihood with data increment parameters as conditional probability, and the development of the new fast maximum a posteriori with entropy (FMAPE) Algorithm by the successive substitution method. It is shown that in the maximum likelihood estimator (MLE) and FMAPE algorithms, the only correct choice of initial image for the iterative procedure in the absence of a priori knowledge about the image configuration is a uniform field.
Resumo:
A new statistical parallax method using the Maximum Likelihood principle is presented, allowing the simultaneous determination of a luminosity calibration, kinematic characteristics and spatial distribution of a given sample. This method has been developed for the exploitation of the Hipparcos data and presents several improvements with respect to the previous ones: the effects of the selection of the sample, the observational errors, the galactic rotation and the interstellar absorption are taken into account as an intrinsic part of the formulation (as opposed to external corrections). Furthermore, the method is able to identify and characterize physically distinct groups in inhomogeneous samples, thus avoiding biases due to unidentified components. Moreover, the implementation used by the authors is based on the extensive use of numerical methods, so avoiding the need for simplification of the equations and thus the bias they could introduce. Several examples of application using simulated samples are presented, to be followed by applications to real samples in forthcoming articles.
Resumo:
BACKGROUND: : A primary goal of clinical pharmacology is to understand the factors that determine the dose-effect relationship and to use this knowledge to individualize drug dose. METHODS: : A principle-based criterion is proposed for deciding among alternative individualization methods. RESULTS: : Safe and effective variability defines the maximum acceptable population variability in drug concentration around the population average. CONCLUSIONS: : A decision on whether patient covariates alone are sufficient, or whether therapeutic drug monitoring in combination with target concentration intervention is needed, can be made by comparing the remaining population variability after a particular dosing method with the safe and effective variability.
Resumo:
[spa] El índice del máximo y el mínimo nivel es una técnica muy útil, especialmente para toma de decisiones, que usa la distancia de Hamming y el coeficiente de adecuación en el mismo problema. En este trabajo, se propone una generalización a través de utilizar medias generalizadas y cuasi aritméticas. A estos operadores de agregación, se les denominará el índice del máximo y el mínimo nivel medio ponderado ordenado generalizado (GOWAIMAM) y cuasi aritmético (Quasi-OWAIMAM). Estos nuevos operadores generalizan una amplia gama de casos particulares como el índice del máximo y el mínimo nivel generalizado (GIMAM), el OWAIMAM, y otros. También se desarrolla una aplicación en la toma de decisiones sobre selección de productos.
Resumo:
[spa] El índice del máximo y el mínimo nivel es una técnica muy útil, especialmente para toma de decisiones, que usa la distancia de Hamming y el coeficiente de adecuación en el mismo problema. En este trabajo, se propone una generalización a través de utilizar medias generalizadas y cuasi aritméticas. A estos operadores de agregación, se les denominará el índice del máximo y el mínimo nivel medio ponderado ordenado generalizado (GOWAIMAM) y cuasi aritmético (Quasi-OWAIMAM). Estos nuevos operadores generalizan una amplia gama de casos particulares como el índice del máximo y el mínimo nivel generalizado (GIMAM), el OWAIMAM, y otros. También se desarrolla una aplicación en la toma de decisiones sobre selección de productos.
Resumo:
Isothermal magnetization curves up to 23 T have been measured in Gd5Si1.8Ge2.2. We show that the values of the entropy change at the first-order magnetostructural transition, obtained from the Clausius-Clapeyron equation and the Maxwell relation, are coincident, provided the Maxwell relation is evaluated only within the transition region and the maximum applied field is high enough to complete the transition. These values are also in agreement with the entropy change obtained from differential scanning calorimetry. We also show that a simple phenomenological model based on the temperature and field dependence of the magnetization accounts for these results.
Resumo:
The extended Gaussian ensemble (EGE) is introduced as a generalization of the canonical ensemble. This ensemble is a further extension of the Gaussian ensemble introduced by Hetherington [J. Low Temp. Phys. 66, 145 (1987)]. The statistical mechanical formalism is derived both from the analysis of the system attached to a finite reservoir and from the maximum statistical entropy principle. The probability of each microstate depends on two parameters ß and ¿ which allow one to fix, independently, the mean energy of the system and the energy fluctuations, respectively. We establish the Legendre transform structure for the generalized thermodynamic potential and propose a stability criterion. We also compare the EGE probability distribution with the q-exponential distribution. As an example, an application to a system with few independent spins is presented.