962 resultados para Maximum entropy method
Resumo:
The internationalization as an organizational phenomenon fundamentally strategic had as theoretical contributions some Schools that throughout the decades 60, 70, and 80 developed behavioral and economic approaches in order to explain the process. The behavioral approach deals with the perception of phenomenon as a gradual process from the perspective of the executives behavior (JOHANSON and VAHLNE, 1977; HALLÉN and WIEDERSHEIM - PAUL, 1979; CZINKOTA, 1985). This phenomenon in permanent theoretical and managerial evolution made an opportunity to build this investigation, whose goal is to analyse the impact comes from organizational capabilities and the external environment on the international performance of exporting firms. For both, were used as theoretical basis two types of analysis for the comprehension of international performance: Strategic Management - Industrial Organization and Resource-Based View and International Businesses - Current Economic and Behavioral. It was made a cross-sectional survey-based explanatory research, including 150 exporting companies with operations in the Northeast of Brazil. A conceptual model was made with eight constructs and eight research hypotheses, representative of the effects of external factors on international performance. The data were processed using the Exploratory Factor Analysis and Structural Equation Modeling. The structural equations model was reespecified and estimated through the use of the maximum-likelihood method up to achieve adequated values of indexes of adjustment. As the main theoretical contribution, were identified organizational and physical resources which shows the importance of the management skills development, of the learning capability and capability to establish strategic alliances abroad. That because the knowledge, as the operational point of view as in its strategic application, offers to organization conditions of market positioning which can create opportunities sustainable competitive advantages and which impact the performance of international companies
Resumo:
The objective of this study was to evaluate the association of visual scores of body structure, precocity and muscularity with production (body weight at 18 months and average daily gain) and reproductive (scrotal circumference) traits in Brahman cattle in order to determine the possible use of these scores as selection criteria to improve carcass quality. Covariance components were estimated by the restricted maximum likelihood method using an animal model that included contemporary group as fixed effect. A total of 1,116 observations of body structure, precocity and muscularity were used. Heritability was 0.39, 043 and 0.40 for body structure, precocity and muscularity, respectively. The genetic correlations were 0.79 between body structure and precocity, 0.87 between body structure and muscularity, and 0.91 between precocity and muscularity. The genetic correlations between visual scores and body weight at 18 months were positive (0.77, 0.57 and 0.59 for body structure, precocity and muscularity, respectively). Similar genetic correlations were observed between average daily gain and visual scores (0.60, 0.57 and 0.48, respectively), whereas the genetic correlations between scrotal circumference and these scores were low (0.13, 0.02, and 0.13). The results indicate that visual scores can be used as selection criteria in Brahman breeding programs. Favorable correlated responses should be seen in average daily gain and body weight at 18 months. However, no correlated response is expected for scrotal circumference.
Resumo:
No estudo de séries temporais, os processos estocásticos usuais assumem que as distribuições marginais são contínuas e, em geral, não são adequados para modelar séries de contagem, pois as suas características não lineares colocam alguns problemas estatísticos, principalmente na estimação dos parâmetros. Assim, investigou-se metodologias apropriadas de análise e modelação de séries com distribuições marginais discretas. Neste contexto, Al-Osh and Alzaid (1987) e McKenzie (1988) introduziram na literatura a classe dos modelos autorregressivos com valores inteiros não negativos, os processos INAR. Estes modelos têm sido frequentemente tratados em artigos científicos ao longo das últimas décadas, pois a sua importância nas aplicações em diversas áreas do conhecimento tem despertado um grande interesse no seu estudo. Neste trabalho, após uma breve revisão sobre séries temporais e os métodos clássicos para a sua análise, apresentamos os modelos autorregressivos de valores inteiros não negativos de primeira ordem INAR (1) e a sua extensão para uma ordem p, as suas propriedades e alguns métodos de estimação dos parâmetros nomeadamente, o método de Yule-Walker, o método de Mínimos Quadrados Condicionais (MQC), o método de Máxima Verosimilhança Condicional (MVC) e o método de Quase Máxima Verosimilhança (QMV). Apresentamos também um critério automático de seleção de ordem para modelos INAR, baseado no Critério de Informação de Akaike Corrigido, AICC, um dos critérios usados para determinar a ordem em modelos autorregressivos, AR. Finalmente, apresenta-se uma aplicação da metodologia dos modelos INAR em dados reais de contagem relativos aos setores dos transportes marítimos e atividades de seguros de Cabo Verde.
Resumo:
We study the problem of detecting sentences describing adverse drug reactions (ADRs) and frame the problem as binary classification. We investigate different neural network (NN) architectures for ADR classification. In particular, we propose two new neural network models, Convolutional Recurrent Neural Network (CRNN) by concatenating convolutional neural networks with recurrent neural networks, and Convolutional Neural Network with Attention (CNNA) by adding attention weights into convolutional neural networks. We evaluate various NN architectures on a Twitter dataset containing informal language and an Adverse Drug Effects (ADE) dataset constructed by sampling from MEDLINE case reports. Experimental results show that all the NN architectures outperform the traditional maximum entropy classifiers trained from n-grams with different weighting strategies considerably on both datasets. On the Twitter dataset, all the NN architectures perform similarly. But on the ADE dataset, CNN performs better than other more complex CNN variants. Nevertheless, CNNA allows the visualisation of attention weights of words when making classification decisions and hence is more appropriate for the extraction of word subsequences describing ADRs.
Resumo:
Fishing trials with monofilament gill nets and longlines using small hooks were carried out at the same fishing grounds in Cyclades (Aegean Sea) over 1 year. Four sizes of MUSTAD brand, round bent, flatted sea hooks (Quality 2316 DT, numbers 15, 13, 12 and 11) and four mesh sizes of 22, 24, 26 and 28 turn nominal bar length monofilament gill nets were used. Significant differences in the catch size frequency distributions of the two gears were found for four out of five of the most important species caught by both the gears (Diplodus annularis, Diplodus vulgaris, Pagellus erythrinus, Scorpaena porcus and Serranus cabrilla), with longlines catching larger fish and a wider size range than gill nets. Whereas longline catch size frequency distributions for most species for the different hook sizes were generally highly overlapped, suggesting little or no differences in size selectivity, gill net catch size frequency distributions clearly showed size selection, with larger mesh sizes catching larger fish. A variety of models were fitted to the gill net data, with the lognormal providing the best fit in most cases. A maximum likelihood method was also used to estimate the parameters of the logistic model for the longline data. Because of the highly overlapped longline catch size frequency distributions parameters could only be estimated for two species. This study shows that the two static gears have different impacts in terms of size selection. This information will be useful for the more effective management of these small-scale, multi-species and multi-gear fisheries. (C) 2002 Elsevier Science B.V. All rights reserved.
Resumo:
The selectivity of four hook sizes (STELL brand(1), Quality 2335, numbers 12, 9, 6 and 4) used in a semi-pelagic longline fishery was studied in the Azores. Two species were caught in sufficient numbers for modelling of selectivity: the black spot sea bream (Pagellus bogaraveo) and the bluemouth rockfish (Helicolenus dactylopterus dactylopterus). A maximum likelihood method was used to fit a versatile model which can be used to describe a wide range of selectivity curves; from bell-shaped to asymptotic. Significant differences in size selectivity between hooks were found for both species. In the case of Pagellus bogaraveo, the smallest hook (number 12) had the lowest catch rates and all hooks were characterised by logistic-type selectivity curves, with sizes at 50% selectivity of: 27.9, 30.4, and 32.8 cm for hooks numbers 12, 9 and 6, respectively. The number 9 hook was the most efficient for Helicolenus d. dactylopterus, with selectivity curves varying from strongly skewed to the right for the number 12 hook to logistic-type for the numbers 6 and 4 hooks. Sizes at 50% selectivity for this species were 16.8, 18.7, 20.7, and 22.0 cm. respectively. (C) 1999 Elsevier Science B.V. All rights reserved.
Resumo:
The objective of this study was to evaluate the effects of inclusion or non-inclusion of short lactations and cow (CGG) and/or dam (DGG) genetic group on the genetic evaluation of 305-day milk yield (MY305), age at first calving (AFC), and first calving interval (FCI) of Girolando cows. Covariance components were estimated by the restricted maximum likelihood method in an animal model of single trait analyses. The heritability estimates for MY305, AFC, and FCI ranged from 0.23 to 0.29, 0.40 to 0.44, and 0.13 to 0.14, respectively, when short lactations were not included, and from 0.23 to 0.28, 0.39 to 0.43, and 0.13 to 0.14, respectively, when short lactations were included. The inclusion of short lactations caused little variation in the variance components and heritability estimates of traits, but their non-inclusion resulted in the re-ranking of animals. Models with CGG or DGG fixed effects had higher heritability estimates for all traits compared with models that consider these two effects simultaneously. We recommend using the model with fixed effects of CGG and inclusion of short lactations for the genetic evaluation of Girolando cattle.
Resumo:
Feed efficiency and carcass characteristics are late-measured traits. The detection of molecular markers associated with them can help breeding programs to select animals early in life, and to predict breeding values with high accuracy. The objective of this study was to identify polymorphisms in the functional and positional candidate gene NEUROD1 (neurogenic differentiation 1), and investigate their associations with production traits in reference families of Nelore cattle. A total of 585 steers were used, from 34 sires chosen to represent the variability of this breed. By sequencing 14 animals with extreme residual feed intake (RFI) values, seven single nucleotide polymorphisms (SNPs) in NEUROD1 were identified. The investigation of marker effects on the target traits RFI, backfat thickness (BFT), ribeye area (REA), average body weight (ABW), and metabolic body weight (MBW) was performed with a mixed model using the restricted maximum likelihood method. SNP1062, which changes cytosine for guanine, had no significant association with RFI or REA. However, we found an additive effect on ABW (P ≤ 0.05) and MBW (P ≤ 0.05), with an estimated allele substitution effect of -1.59 and -0.93 kg0.75, respectively. A dominant effect of this SNP for BFT was also found (P ≤ 0.010). Our results are the first that identify NEUROD1 as a candidate that affects BFT, ABW, and MBW. Once confirmed, the inclusion of this SNP in dense panels may improve the accuracy of genomic selection for these traits in Nelore beef cattle as this SNP is not currently represented on SNP chips.
Resumo:
Knowledge of the geographical distribution of timber tree species in the Amazon is still scarce. This is especially true at the local level, thereby limiting natural resource management actions. Forest inventories are key sources of information on the occurrence of such species. However, areas with approved forest management plans are mostly located near access roads and the main industrial centers. The present study aimed to assess the spatial scale effects of forest inventories used as sources of occurrence data in the interpolation of potential species distribution models. The occurrence data of a group of six forest tree species were divided into four geographical areas during the modeling process. Several sampling schemes were then tested applying the maximum entropy algorithm, using the following predictor variables: elevation, slope, exposure, normalized difference vegetation index (NDVI) and height above the nearest drainage (HAND). The results revealed that using occurrence data from only one geographical area with unique environmental characteristics increased both model overfitting to input data and omission error rates. The use of a diagonal systematic sampling scheme and lower threshold values led to improved model performance. Forest inventories may be used to predict areas with a high probability of species occurrence, provided they are located in forest management plan regions representative of the environmental range of the model projection area.
Resumo:
We propose a mixed finite element method for a class of nonlinear diffusion equations, which is based on their interpretation as gradient flows in optimal transportation metrics. We introduce an appropriate linearization of the optimal transport problem, which leads to a mixed symmetric formulation. This formulation preserves the maximum principle in case of the semi-discrete scheme as well as the fully discrete scheme for a certain class of problems. In addition solutions of the mixed formulation maintain exponential convergence in the relative entropy towards the steady state in case of a nonlinear Fokker-Planck equation with uniformly convex potential. We demonstrate the behavior of the proposed scheme with 2D simulations of the porous medium equations and blow-up questions in the Patlak-Keller-Segel model.
Resumo:
We extend PML theory to account for information on the conditional moments up to order four, but without assuming a parametric model, to avoid a risk of misspecification of the conditional distribution. The key statistical tool is the quartic exponential family, which allows us to generalize the PML2 and QGPML1 methods proposed in Gourieroux et al. (1984) to PML4 and QGPML2 methods, respectively. An asymptotic theory is developed. The key numerical tool that we use is the Gauss-Freud integration scheme that solves a computational problem that has previously been raised in several fields. Simulation exercises demonstrate the feasibility and robustness of the methods [Authors]
Resumo:
In the forensic examination of DNA mixtures, the question of how to set the total number of contributors (N) presents a topic of ongoing interest. Part of the discussion gravitates around issues of bias, in particular when assessments of the number of contributors are not made prior to considering the genotypic configuration of potential donors. Further complication may stem from the observation that, in some cases, there may be numbers of contributors that are incompatible with the set of alleles seen in the profile of a mixed crime stain, given the genotype of a potential contributor. In such situations, procedures that take a single and fixed number contributors as their output can lead to inferential impasses. Assessing the number of contributors within a probabilistic framework can help avoiding such complication. Using elements of decision theory, this paper analyses two strategies for inference on the number of contributors. One procedure is deterministic and focuses on the minimum number of contributors required to 'explain' an observed set of alleles. The other procedure is probabilistic using Bayes' theorem and provides a probability distribution for a set of numbers of contributors, based on the set of observed alleles as well as their respective rates of occurrence. The discussion concentrates on mixed stains of varying quality (i.e., different numbers of loci for which genotyping information is available). A so-called qualitative interpretation is pursued since quantitative information such as peak area and height data are not taken into account. The competing procedures are compared using a standard scoring rule that penalizes the degree of divergence between a given agreed value for N, that is the number of contributors, and the actual value taken by N. Using only modest assumptions and a discussion with reference to a casework example, this paper reports on analyses using simulation techniques and graphical models (i.e., Bayesian networks) to point out that setting the number of contributors to a mixed crime stain in probabilistic terms is, for the conditions assumed in this study, preferable to a decision policy that uses categoric assumptions about N.
Resumo:
The development and tests of an iterative reconstruction algorithm for emission tomography based on Bayesian statistical concepts are described. The algorithm uses the entropy of the generated image as a prior distribution, can be accelerated by the choice of an exponent, and converges uniformly to feasible images by the choice of one adjustable parameter. A feasible image has been defined as one that is consistent with the initial data (i.e. it is an image that, if truly a source of radiation in a patient, could have generated the initial data by the Poisson process that governs radioactive disintegration). The fundamental ideas of Bayesian reconstruction are discussed, along with the use of an entropy prior with an adjustable contrast parameter, the use of likelihood with data increment parameters as conditional probability, and the development of the new fast maximum a posteriori with entropy (FMAPE) Algorithm by the successive substitution method. It is shown that in the maximum likelihood estimator (MLE) and FMAPE algorithms, the only correct choice of initial image for the iterative procedure in the absence of a priori knowledge about the image configuration is a uniform field.
Resumo:
Due to the lack of information concerning maximum rainfall equations for most locations in Mato Grosso do Sul State, the alternative for carrying out hydraulic work projects has been information from meteorological stations closest to the location in which the project is carried out. Alternative methods, such as 24 hours rain disaggregation method from rainfall data due to greater availability of stations and longer observations can work. Based on this approach, the objective of this study was to estimate maximum rainfall equations for Mato Grosso do Sul State by adjusting the 24 hours rain disaggregation method, depending on data obtained from rain gauge stations from Dourado and Campo Grande. For this purpose, data consisting of 105 rainfall stations were used, which are available in the ANA (Water Resources Management National Agency) database. Based on the results we concluded: the intense rainfall equations obtained by pluviogram analysis showed determination coefficient above 99%; and the performance of 24 hours rain disaggregation method was classified as excellent, based on relative average error WILMOTT concordance index (1982).
Resumo:
It has been generally accepted that the method of moments (MoM) variogram, which has been widely applied in soil science, requires about 100 sites at an appropriate interval apart to describe the variation adequately. This sample size is often larger than can be afforded for soil surveys of agricultural fields or contaminated sites. Furthermore, it might be a much larger sample size than is needed where the scale of variation is large. A possible alternative in such situations is the residual maximum likelihood (REML) variogram because fewer data appear to be required. The REML method is parametric and is considered reliable where there is trend in the data because it is based on generalized increments that filter trend out and only the covariance parameters are estimated. Previous research has suggested that fewer data are needed to compute a reliable variogram using a maximum likelihood approach such as REML, however, the results can vary according to the nature of the spatial variation. There remain issues to examine: how many fewer data can be used, how should the sampling sites be distributed over the site of interest, and how do different degrees of spatial variation affect the data requirements? The soil of four field sites of different size, physiography, parent material and soil type was sampled intensively, and MoM and REML variograms were calculated for clay content. The data were then sub-sampled to give different sample sizes and distributions of sites and the variograms were computed again. The model parameters for the sets of variograms for each site were used for cross-validation. Predictions based on REML variograms were generally more accurate than those from MoM variograms with fewer than 100 sampling sites. A sample size of around 50 sites at an appropriate distance apart, possibly determined from variograms of ancillary data, appears adequate to compute REML variograms for kriging soil properties for precision agriculture and contaminated sites. (C) 2007 Elsevier B.V. All rights reserved.