976 resultados para Probability and Statistics
Resumo:
In this paper, we investigate the joint effects of high-power amplifier (HPA) nonlinearity, in-phase/quadrature-phase (I/Q) imbalance and crosstalk, on the performance of multiple-input multiple-output (MIMO) transmit beamforming (TB) systems, and propose a compensation method for the three impairments together. The performance of the MIMO TB system equipped with the proposed compensation scheme is evaluated in terms of average symbol error probability and capacity when transmissions are performed over uncorrelated Rayleigh fading channels. Numerical results are provided and show the effects on performance of several system parameters, namely, the HPA parameters, image-leakage ratio, crosstalk, numbers of antennas, length of pilot symbols and phase-shift keying modulation order.
Resumo:
Objective: To identify food acquisition patterns in Brazil and relate them to the sociodemographic characteristics of the household. Design: A cross-sectional national Household Budget Survey (HBS). Principal component factor analysis was used to derive food patterns (factors) on the basis of the acquisition of food classified into thirty-two food groups. Setting: The source of data originates from the 2002-2003 HBS carried out by the Brazilian Institute of Geography and Statistics between June 2002 and July 2003 using a representative sample of all Brazilian households. Subject: A total of 48 470 households allocated into 443 strata of households that were geographically and socio-economically homogeneous as a study unit. Results: We identified two patterns of food acquisition. The first, named `dual`, was characterized by dairy, fruit, fruit juice, vegetables, processed meat, soft drinks, sweets, bread and margarine, and by inverse correlations with Brazilian staple foods. In contrast, the second pattern, named `traditional`, was characterized by rice, beans, manioc, flour, milk and sugar. The `dual` pattern was associated with higher household educational level, income and the average age of adults on the strata, whereas the `traditional` presented higher loadings in less-educated households and in the rural setting. Conclusions: Dietary patterns described here suggest that policies and programmes to promote healthy eating need to consider that healthy and non-healthy foods may be integral in the same pattern.
Resumo:
We study random walks systems on Z whose general description follows. At time zero, there is a number N >= 1 of particles at each vertex of N, all being inactive, except for those placed at the vertex one. Each active particle performs a simple random walk on Z and, up to the time it dies, it activates all inactive particles that it meets along its way. An active particle dies at the instant it reaches a certain fixed total of jumps (L >= 1) without activating any particle, so that its lifetime depends strongly on the past of the process. We investigate how the probability of survival of the process depends on L and on the jumping probabilities of the active particles.
Resumo:
Energy efficiency and renewable energy use are two main priorities leading to industrial sustainability nowadays according to European Steel Technology Platform (ESTP). Modernization efforts can be done by industries to improve energy consumptions of the production lines. These days, steel making industrial applications are energy and emission intensive. It was estimated that over the past years, energy consumption and corresponding CO2 generation has increased steadily reaching approximately 338.15 parts per million in august 2010 [1]. These kinds of facts and statistics have introduced a lot of room for improvement in energy efficiency for industrial applications through modernization and use of renewable energy sources such as solar Photovoltaic Systems (PV).The purpose of this thesis work is to make a preliminary design and simulation of the solar photovoltaic system which would attempt to cover the energy demand of the initial part of the pickling line hydraulic system at the SSAB steel plant. For this purpose, the energy consumptions of this hydraulic system would be studied and evaluated and a general analysis of the hydraulic and control components performance would be done which would yield a proper set of guidelines contributing towards future energy savings. The results of the energy efficiency analysis showed that the initial part of the pickling line hydraulic system worked with a low efficiency of 3.3%. Results of general analysis showed that hydraulic accumulators of 650 liter size should be used by the initial part pickling line system in combination with a one pump delivery of 100 l/min. Based on this, one PV system can deliver energy to an AC motor-pump set covering 17.6% of total energy and another PV system can supply a DC hydraulic pump substituting 26.7% of the demand. The first system used 290 m2 area of the roof and was sized as 40 kWp, the second used 109 m2 and was sized as 15.2 kWp. It was concluded that the reason for the low efficiency was the oversized design of the system. Incremental modernization efforts could help to improve the hydraulic system energy efficiency and make the design of the solar photovoltaic system realistically possible. Two types of PV systems where analyzed in the thesis work. A method was found calculating the load simulation sequence based on the energy efficiency studies to help in the PV system simulations. Hydraulic accumulators integrated into the pickling line worked as energy storage when being charged by the PV system as well.
Resumo:
Solutions to combinatorial optimization problems, such as problems of locating facilities, frequently rely on heuristics to minimize the objective function. The optimum is sought iteratively and a criterion is needed to decide when the procedure (almost) attains it. Pre-setting the number of iterations dominates in OR applications, which implies that the quality of the solution cannot be ascertained. A small, almost dormant, branch of the literature suggests using statistical principles to estimate the minimum and its bounds as a tool to decide upon stopping and evaluating the quality of the solution. In this paper we examine the functioning of statistical bounds obtained from four different estimators by using simulated annealing on p-median test problems taken from Beasley’s OR-library. We find the Weibull estimator and the 2nd order Jackknife estimator preferable and the requirement of sample size to be about 10 being much less than the current recommendation. However, reliable statistical bounds are found to depend critically on a sample of heuristic solutions of high quality and we give a simple statistic useful for checking the quality. We end the paper with an illustration on using statistical bounds in a problem of locating some 70 distribution centers of the Swedish Post in one Swedish region.
Resumo:
MAPfastR is a software package developed to analyze QTL data from inbred and outbred line-crosses. The package includes a number of modules for fast and accurate QTL analyses. It has been developed in the R language for fast and comprehensive analyses of large datasets. MAPfastR is freely available at: http://www.computationalgenetics.se/?page_id=7.
Resumo:
We present the hglm package for fitting hierarchical generalized linear models. It can be used for linear mixed models and generalized linear mixed models with random effects for a variety of links and a variety of distributions for both the outcomes and the random effects. Fixed effects can also be fitted in the dispersion part of the model.
Resumo:
Background: The sensitivity to microenvironmental changes varies among animals and may be under genetic control. It is essential to take this element into account when aiming at breeding robust farm animals. Here, linear mixed models with genetic effects in the residual variance part of the model can be used. Such models have previously been fitted using EM and MCMC algorithms. Results: We propose the use of double hierarchical generalized linear models (DHGLM), where the squared residuals are assumed to be gamma distributed and the residual variance is fitted using a generalized linear model. The algorithm iterates between two sets of mixed model equations, one on the level of observations and one on the level of variances. The method was validated using simulations and also by re-analyzing a data set on pig litter size that was previously analyzed using a Bayesian approach. The pig litter size data contained 10,060 records from 4,149 sows. The DHGLM was implemented using the ASReml software and the algorithm converged within three minutes on a Linux server. The estimates were similar to those previously obtained using Bayesian methodology, especially the variance components in the residual variance part of the model. Conclusions: We have shown that variance components in the residual variance part of a linear mixed model can be estimated using a DHGLM approach. The method enables analyses of animal models with large numbers of observations. An important future development of the DHGLM methodology is to include the genetic correlation between the random effects in the mean and residual variance parts of the model as a parameter of the DHGLM.
Resumo:
This paper presents the techniques of likelihood prediction for the generalized linear mixed models. Methods of likelihood prediction is explained through a series of examples; from a classical one to more complicated ones. The examples show, in simple cases, that the likelihood prediction (LP) coincides with already known best frequentist practice such as the best linear unbiased predictor. The paper outlines a way to deal with the covariate uncertainty while producing predictive inference. Using a Poisson error-in-variable generalized linear model, it has been shown that in complicated cases LP produces better results than already know methods.
Resumo:
The p-median problem is often used to locate P service facilities in a geographically distributed population. Important for the performance of such a model is the distance measure. Distance measure can vary if the accuracy of the road network varies. The rst aim in this study is to analyze how the optimal location solutions vary, using the p-median model, when the road network is alternated. It is hard to nd an exact optimal solution for p-median problems. Therefore, in this study two heuristic solutions are applied, simulating annealing and a classic heuristic. The secondary aim is to compare the optimal location solutions using dierent algorithms for large p-median problem. The investigation is conducted by the means of a case study in a rural region with an asymmetrically distributed population, Dalecarlia. The study shows that the use of more accurate road networks gives better solutions for optimal location, regardless what algorithm that is used and regardless how many service facilities that is optimized for. It is also shown that the simulated annealing algorithm not just is much faster than the classic heuristic used here, but also in most cases gives better location solutions.
Resumo:
Generalized linear mixed models are flexible tools for modeling non-normal data and are useful for accommodating overdispersion in Poisson regression models with random effects. Their main difficulty resides in the parameter estimation because there is no analytic solution for the maximization of the marginal likelihood. Many methods have been proposed for this purpose and many of them are implemented in software packages. The purpose of this study is to compare the performance of three different statistical principles - marginal likelihood, extended likelihood, Bayesian analysis-via simulation studies. Real data on contact wrestling are used for illustration.
Resumo:
This paper studies the impact of HIV/AIDS on per capita income and education. It explores two channels from HIV/AIDS to income that have not been sufficiently stressed by the literature: the reduction of the incentives to study due to shorter expected longevity and the reduction of productivity of experienced workers. In the model individuals live for three periods, may get infected in the second period and with some probability die of Aids before reaching the third period of their life. Parents care for the welfare of the future generations so that they will maximize lifetime utility of their dynasty. The simulations predict that the most affected countries in Sub-Saharan Africa will be in the future, on average, thirty percent poorer than they would be without AIDS. Schooling will decline in some cases by forty percent. These figures are dramatically reduced with widespread medical treatment, as it increases the survival probability and productivity of infected individuals.
Resumo:
This paper provides a systematic and unified treatment of the developments in the area of kernel estimation in econometrics and statistics. Both the estimation and hypothesis testing issues are discussed for the nonparametric and semiparametric regression models. A discussion on the choice of windowwidth is also presented.
Resumo:
The inflationary stabilization recently observed in Brazil brings a lot of changes in all aspects of the country’s economic life. In this work we look at the impacts on the stock market, specifically at Bovespa - the São Paulo Stock Exchange. We analyze the leading variables and statistics that describe Bovespa’s behavior, such as volatility and systematic risk, comparing the four years preceding and the four years after 1994, when the Real Plan was implemented. In order to eliminate exogenous influences, we use control series made with international Stock Exchanges Indexes. The results show that after 1994 there was reduced volatility, increased trade volume, reduced efficiency of the Bovespa Index and no changes in systematic risk.
Resumo:
Currently, one of the biggest challenges for the field of data mining is to perform cluster analysis on complex data. Several techniques have been proposed but, in general, they can only achieve good results within specific areas providing no consensus of what would be the best way to group this kind of data. In general, these techniques fail due to non-realistic assumptions about the true probability distribution of the data. Based on this, this thesis proposes a new measure based on Cross Information Potential that uses representative points of the dataset and statistics extracted directly from data to measure the interaction between groups. The proposed approach allows us to use all advantages of this information-theoretic descriptor and solves the limitations imposed on it by its own nature. From this, two cost functions and three algorithms have been proposed to perform cluster analysis. As the use of Information Theory captures the relationship between different patterns, regardless of assumptions about the nature of this relationship, the proposed approach was able to achieve a better performance than the main algorithms in literature. These results apply to the context of synthetic data designed to test the algorithms in specific situations and to real data extracted from problems of different fields