1000 resultados para Distribuições de probabilidade


Relevância:

20.00% 20.00%

Publicador:

Resumo:

Currently great emphasis is given for seed metering that assist rigorous demands in relation to longitudinal distribution of seeds, as well as to the index of fails in spacing laws, breaks and double seeds. The evaluation of these variable demands much time and work of attainment of data and processing. The objective of this work went propose to use of graphs of normal probability, facilitating the treatment of the data and decreasing the time of processing. The evaluation methodology consists in the counting of broken seeds, fail spacing and double seeds through the measure of the spacing among seeds, preliminary experiments through combinations of treatments had been carried through whose factors of variation were the level of the reservoir of seeds, the leveling of the seed metering, the speed of displacement and dosage of seeds. The evaluation was carried through in two parts, first through preliminary experiments for elaboration of the graphs of normal probability and later in experiments with bigger sampling for evaluation of the influence of the factors most important. It was done the evaluation of seed metering of rotating internal ring, and the amount of necessary data for the evaluation was very decreased through of the graphs of normal probability that facilitated to prioritize only the significant factors. The dosage of seeds was factor that more important because factor (D) have greater significance.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The static and cyclic assays are common to test materials in structures.. For cycling assays to assess the fatigue behavior of the material and thereby obtain the S-N curves and these are used to construct the diagrams of living constant. However, these diagrams, when constructed with small amounts of S-N curves underestimate or overestimate the actual behavior of the composite, there is increasing need for more testing to obtain more accurate results. Therewith, , a way of reducing costs is the statistical analysis of the fatigue behavior. The aim of this research was evaluate the probabilistic fatigue behavior of composite materials. The research was conducted in three parts. The first part consists of associating the equation of probability Weilbull equations commonly used in modeling of composite materials S-N curve, namely the exponential equation and power law and their generalizations. The second part was used the results obtained by the equation which best represents the S-N curves of probability and trained a network to the modular 5% failure. In the third part, we carried out a comparative study of the results obtained using the nonlinear model by parts (PNL) with the results of a modular network architecture (MN) in the analysis of fatigue behavior. For this we used a database of ten materials obtained from the literature to assess the ability of generalization of the modular network as well as its robustness. From the results it was found that the power law of probability generalized probabilistic behavior better represents the fatigue and composites that although the generalization ability of the MN that was not robust training with 5% failure rate, but for values mean the MN showed more accurate results than the PNL model

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The state of Rio Grande do Norte, possessor of an extremely irregular regime of rains, has the necessity of enlarge and specify the researches about its own hydro-climatic conditions, to achieve trustworthy results that are able to minimize the adversities imposed by these conditions and make possible the implementation of a better planning in the economic activities and of subsistence that somehow utilize of the multiple uses of hydro resources of the State. This way, the daily values observed from the pluviometric series of 166 posts, with 45 years uninterrupted of historic data, were adjusted to the incomplete gamma function to the determination of the probability of rain in the 36 period of ten days in which the year was divided. To the attainment of the α and β parameters of this function it was applied the method of the maximum verisimilitude allowing, in the end, to analyze the temporal and spatial distribution of the rain in the level of 75% of probability. The values of potential evapo-transpiration were calculated by the Linacre method that, through the SURFER software, were confronted with the dependant rain, obtaining, in this way, the spatialization of the potential hydro availability, which the values can be known to any period of ten days of the year, city and/or region of the state of Rio Grande do Norte. With the identification of the main meteorological systems that act in the State, we sought to better comprehend how this systems interfere, in the irregular regime of rain, in the situations of several clime in the major part of Rio Grande do Norte and in the hydro regional balance. And, finally, with these data in hand and with the generated maps, we verified that space-temporal distribution of the rain and of the potential hydro availability were heterogeneous in the whole State, mainly in the West and Central regions, inserted in potiguar s semi-arid, which, after the period of the rains station, suffers with dry season and length drought during the rest of the year

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The complex behavior of a wide variety of phenomena that are of interest to physicists, chemists, and engineers has been quantitatively characterized by using the ideas of fractal and multifractal distributions, which correspond in a unique way to the geometrical shape and dynamical properties of the systems under study. In this thesis we present the Space of Fractals and the methods of Hausdorff-Besicovitch, box-counting and Scaling to calculate the fractal dimension of a set. In this Thesis we investigate also percolation phenomena in multifractal objects that are built in a simple way. The central object of our analysis is a multifractal object that we call Qmf . In these objects the multifractality comes directly from the geometric tiling. We identify some differences between percolation in the proposed multifractals and in a regular lattice. There are basically two sources of these differences. The first is related to the coordination number, c, which changes along the multifractal. The second comes from the way the weight of each cell in the multifractal affects the percolation cluster. We use many samples of finite size lattices and draw the histogram of percolating lattices against site occupation probability p. Depending on a parameter, ρ, characterizing the multifractal and the lattice size, L, the histogram can have two peaks. We observe that the probability of occupation at the percolation threshold, pc, for the multifractal is lower than that for the square lattice. We compute the fractal dimension of the percolating cluster and the critical exponent β. Despite the topological differences, we find that the percolation in a multifractal support is in the same universality class as standard percolation. The area and the number of neighbors of the blocks of Qmf show a non-trivial behavior. A general view of the object Qmf shows an anisotropy. The value of pc is a function of ρ which is related to its anisotropy. We investigate the relation between pc and the average number of neighbors of the blocks as well as the anisotropy of Qmf. In this Thesis we study likewise the distribution of shortest paths in percolation systems at the percolation threshold in two dimensions (2D). We study paths from one given point to multiple other points

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The complex behavior of a wide variety of phenomena that are of interest to physicists, chemists, and engineers has been quantitatively characterized by using the ideas of fractal and multifractal distributions, which correspond in a unique way to the geometrical shape and dynamical properties of the systems under study. In this thesis we present the Space of Fractals and the methods of Hausdorff-Besicovitch, box-counting and Scaling to calculate the fractal dimension of a set. In this Thesis we investigate also percolation phenomena in multifractal objects that are built in a simple way. The central object of our analysis is a multifractal object that we call Qmf . In these objects the multifractality comes directly from the geometric tiling. We identify some differences between percolation in the proposed multifractals and in a regular lattice. There are basically two sources of these differences. The first is related to the coordination number, c, which changes along the multifractal. The second comes from the way the weight of each cell in the multifractal affects the percolation cluster. We use many samples of finite size lattices and draw the histogram of percolating lattices against site occupation probability p. Depending on a parameter, ρ, characterizing the multifractal and the lattice size, L, the histogram can have two peaks. We observe that the probability of occupation at the percolation threshold, pc, for the multifractal is lower than that for the square lattice. We compute the fractal dimension of the percolating cluster and the critical exponent β. Despite the topological differences, we find that the percolation in a multifractal support is in the same universality class as standard percolation. The area and the number of neighbors of the blocks of Qmf show a non-trivial behavior. A general view of the object Qmf shows an anisotropy. The value of pc is a function of ρ which is related to its anisotropy. We investigate the relation between pc and the average number of neighbors of the blocks as well as the anisotropy of Qmf. In this Thesis we study likewise the distribution of shortest paths in percolation systems at the percolation threshold in two dimensions (2D). We study paths from one given point to multiple other points. In oil recovery terminology, the given single point can be mapped to an injection well (injector) and the multiple other points to production wells (producers). In the previously standard case of one injection well and one production well separated by Euclidean distance r, the distribution of shortest paths l, P(l|r), shows a power-law behavior with exponent gl = 2.14 in 2D. Here we analyze the situation of one injector and an array A of producers. Symmetric arrays of producers lead to one peak in the distribution P(l|A), the probability that the shortest path between the injector and any of the producers is l, while the asymmetric configurations lead to several peaks in the distribution. We analyze configurations in which the injector is outside and inside the set of producers. The peak in P(l|A) for the symmetric arrays decays faster than for the standard case. For very long paths all the studied arrays exhibit a power-law behavior with exponent g ∼= gl.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

In Survival Analysis, long duration models allow for the estimation of the healing fraction, which represents a portion of the population immune to the event of interest. Here we address classical and Bayesian estimation based on mixture models and promotion time models, using different distributions (exponential, Weibull and Pareto) to model failure time. The database used to illustrate the implementations is described in Kersey et al. (1987) and it consists of a group of leukemia patients who underwent a certain type of transplant. The specific implementations used were numeric optimization by BFGS as implemented in R (base::optim), Laplace approximation (own implementation) and Gibbs sampling as implemented in Winbugs. We describe the main features of the models used, the estimation methods and the computational aspects. We also discuss how different prior information can affect the Bayesian estimates

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This paper we study a random strategy called MOSES, which was introduced in 1996 by Fran¸cois. Asymptotic results of this strategy; behavior of the stationary distributions of the chain associated to strategy, were derived by Fran¸cois, in 1998, of the theory of Freidlin and Wentzell [8]. Detailings of these results are in this work. Moreover, we noted that an alternative approach the convergence of this strategy is possible without making use of theory of Freidlin and Wentzell, yielding the visit almost certain of the strategy to uniform populations which contain the minimum. Some simulations in Matlab are presented in this work

Relevância:

20.00% 20.00%

Publicador:

Resumo:

O estudo teve o propósito de verificar a influência da textura do recurso pedagógico no tempo despendido e no índice de retidão para execução de uma atividade de encaixe realizada por indivíduos com paralisia cerebral. Participaram do estudo 6 alunos com sequelas de paralisia cerebral do tipo espástica, com idade entre 7 anos e 8 meses e 28 anos e 1 mês. O participante deveria segurar o objeto e levá-lo a um ponto demarcado previamente para o encaixe e, após, retorná-lo à posição inicial. O objeto foi apresentado nas texturas lisa, intermediária e áspera, de forma aleatória. Para a coleta de dados, foi utilizada filmagem para posterior leitura pelo programa Kavideo, que possibilitou cálculos para observar o tempo despendido na tarefa e o índice de retidão. Foi realizado o estudo da normalidade das distribuições, verificada por meio do teste de normalidade de Kolmogorov-Smirnov (Teste KS) e a comparação entre as texturas lisa, intermediária e áspera, no movimento de encaixe, por meio da análise de variância de medidas repetidas. Adotou-se para todos os testes nível de significância de 5% de probabilidade para a rejeição da hipótese de normalidade. O resultado do teste de análise de variância de medidas repetidas mostrou que não houve significância para as variáveis estudadas com o recurso apresentado nas três texturas. Porém, os dados da estatística descritiva mostraram que o encaixe do recurso com a textura lisa despendeu um menor tempo para execução da tarefa e, também, determinou um melhor desempenho com relação à trajetória do movimento.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The segmentation of an image aims to subdivide it into constituent regions or objects that have some relevant semantic content. This subdivision can also be applied to videos. However, in these cases, the objects appear in various frames that compose the videos. The task of segmenting an image becomes more complex when they are composed of objects that are defined by textural features, where the color information alone is not a good descriptor of the image. Fuzzy Segmentation is a region-growing segmentation algorithm that uses affinity functions in order to assign to each element in an image a grade of membership for each object (between 0 and 1). This work presents a modification of the Fuzzy Segmentation algorithm, for the purpose of improving the temporal and spatial complexity. The algorithm was adapted to segmenting color videos, treating them as 3D volume. In order to perform segmentation in videos, conventional color model or a hybrid model obtained by a method for choosing the best channels were used. The Fuzzy Segmentation algorithm was also applied to texture segmentation by using adaptive affinity functions defined for each object texture. Two types of affinity functions were used, one defined using the normal (or Gaussian) probability distribution and the other using the Skew Divergence. This latter, a Kullback-Leibler Divergence variation, is a measure of the difference between two probability distributions. Finally, the algorithm was tested in somes videos and also in texture mosaic images composed by images of the Brodatz album

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Two-level factorial designs are widely used in industrial experimentation. However, many factors in such a design require a large number of runs to perform the experiment, and too many replications of the treatments may not be feasible, considering limitations of resources and of time, making it expensive. In these cases, unreplicated designs are used. But, with only one replicate, there is no internal estimate of experimental error to make judgments about the significance of the observed efects. One of the possible solutions for this problem is to use normal plots or half-normal plots of the efects. Many experimenters use the normal plot, while others prefer the half-normal plot and, often, for both cases, without justification. The controversy about the use of these two graphical techniques motivates this work, once there is no register of formal procedure or statistical test that indicates \which one is best". The choice between the two plots seems to be a subjective issue. The central objective of this master's thesis is, then, to perform an experimental comparative study of the normal plot and half-normal plot in the context of the analysis of the 2k unreplicated factorial experiments. This study involves the construction of simulated scenarios, in which the graphics performance to detect significant efects and to identify outliers is evaluated in order to verify the following questions: Can be a plot better than other? In which situations? What kind of information does a plot increase to the analysis of the experiment that might complement those provided by the other plot? What are the restrictions on the use of graphics? Herewith, this work intends to confront these two techniques; to examine them simultaneously in order to identify similarities, diferences or relationships that contribute to the construction of a theoretical reference to justify or to aid in the experimenter's decision about which of the two graphical techniques to use and the reason for this use. The simulation results show that the half-normal plot is better to assist in the judgement of the efects, while the normal plot is recommended to detect outliers in the data

Relevância:

20.00% 20.00%

Publicador:

Resumo:

In survival analysis, the response is usually the time until the occurrence of an event of interest, called failure time. The main characteristic of survival data is the presence of censoring which is a partial observation of response. Associated with this information, some models occupy an important position by properly fit several practical situations, among which we can mention the Weibull model. Marshall-Olkin extended form distributions other a basic generalization that enables greater exibility in adjusting lifetime data. This paper presents a simulation study that compares the gradient test and the likelihood ratio test using the Marshall-Olkin extended form Weibull distribution. As a result, there is only a small advantage for the likelihood ratio test

Relevância:

20.00% 20.00%

Publicador:

Resumo:

In the work reported here we present theoretical and numerical results about a Risk Model with Interest Rate and Proportional Reinsurance based on the article Inequalities for the ruin probability in a controlled discrete-time risk process by Ros ario Romera and Maikol Diasparra (see [5]). Recursive and integral equations as well as upper bounds for the Ruin Probability are given considering three di erent approaches, namely, classical Lundberg inequality, Inductive approach and Martingale approach. Density estimation techniques (non-parametrics) are used to derive upper bounds for the Ruin Probability and the algorithms used in the simulation are presented

Relevância:

20.00% 20.00%

Publicador:

Resumo:

We present a dependent risk model to describe the surplus of an insurance portfolio, based on the article "A ruin model with dependence between claim sizes and claim intervals"(Albrecher and Boxma [1]). An exact expression for the Laplace transform of the survival function of the surplus is derived. The results obtained are illustrated by several numerical examples and the case when we ignore the dependence structure present in the model is investigated. For the phase type claim sizes, we study by the survival probability, considering this is a class of distributions computationally tractable and more general

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The central objective of a study Non-Homogeneous Markov Chains is the concept of weak and strong ergodicity. A chain is weak ergodic if the dependence on the initial distribution vanishes with time, and it is strong ergodic if it is weak ergodic and converges in distribution. Most theoretical results on strong ergodicity assume some knowledge of the limit behavior of the stationary distributions. In this work, we collect some general results on weak and strong ergodicity for chains with space enumerable states, and also study the asymptotic behavior of the stationary distributions of a particular type of Markov Chains with finite state space, called Markov Chains with Rare Transitions

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Coordenação de Aperfeiçoamento de Pessoal de Nível Superior