876 resultados para Probabilistic Finite Automata
Resumo:
Fixed-point roundoff noise in digital implementation of linear systems arises due to overflow, quantization of coefficients and input signals, and arithmetical errors. In uniform white-noise models, the last two types of roundoff errors are regarded as uniformly distributed independent random vectors on cubes of suitable size. For input signal quantization errors, the heuristic model is justified by a quantization theorem, which cannot be directly applied to arithmetical errors due to the complicated input-dependence of errors. The complete uniform white-noise model is shown to be valid in the sense of weak convergence of probabilistic measures as the lattice step tends to zero if the matrices of realization of the system in the state space satisfy certain nonresonance conditions and the finite-dimensional distributions of the input signal are absolutely continuous.
Resumo:
Numerical modeling of the eddy currents induced in the human body by the pulsed field gradients in MRI presents a difficult computational problem. It requires an efficient and accurate computational method for high spatial resolution analyses with a relatively low input frequency. In this article, a new technique is described which allows the finite difference time domain (FDTD) method to be efficiently applied over a very large frequency range, including low frequencies. This is not the case in conventional FDTD-based methods. A method of implementing streamline gradients in FDTD is presented, as well as comparative analyses which show that the correct source injection in the FDTD simulation plays a crucial rule in obtaining accurate solutions. In particular, making use of the derivative of the input source waveform is shown to provide distinct benefits in accuracy over direct source injection. In the method, no alterations to the properties of either the source or the transmission media are required. The method is essentially frequency independent and the source injection method has been verified against examples with analytical solutions. Results are presented showing the spatial distribution of gradient-induced electric fields and eddy currents in a complete body model.
Resumo:
The role of PACs (primary adsorption centers) in the mesopore (i.e., transport) region of activated carbons during adsorption of polar species, such as water, is unclear. A classical model of three-dimensional adsorption on finite PACs is presented. The model is a preliminary, theoretical investigation into adsorption on mesopore PACs and is intended to give some insight into the energetic and physical processes at work. Work processes are developed to obtain isotherms and three-dimensional sorbate growth on PACs of varying size and energetic characteristics. The work processes allow two forms of adsorbed phase growth: densification at constant boundary and boundary growth at constant density. Relatively strong sorbate-sorbent interactions and strong surface tension favor adsorbed phase densification over boundary growth. Conversely, relatively weak sorbate-sorbent interactions and weak surface tension favor boundary growth over densification. If sorbate-sorbate interactions are strong compared to sorbate-sorbent interactions, condensation with hysteresis occurs. This can also give rise to delayed boundary growth, where all initial adsorption occurs in the monolayer only. The results indicate that adsorbed phase growth on PACs may be quite complex.
Resumo:
The purpose of this study was threefold: first, the study was designed to illustrate the use of data and information collected in food safety surveys in a quantitative risk assessment. In this case, the focus was on the food service industry; however, similar data from other parts of the food chain could be similarly incorporated. The second objective was to quantitatively describe and better understand the role that the food service industry plays in the safety of food. The third objective was to illustrate the additional decision-making information that is available when uncertainty and variability are incorporated into the modelling of systems. (C) 2002 Elsevier Science B.V. All rights reserved.
Resumo:
Error condition detected We consider discrete two-point boundary value problems of the form D-2 y(k+1) = f (kh, y(k), D y(k)), for k = 1,...,n - 1, (0,0) = G((y(0),y(n));(Dy-1,Dy-n)), where Dy-k = (y(k) - Yk-I)/h and h = 1/n. This arises as a finite difference approximation to y" = f(x,y,y'), x is an element of [0,1], (0,0) = G((y(0),y(1));(y'(0),y'(1))). We assume that f and G = (g(0), g(1)) are continuous and fully nonlinear, that there exist pairs of strict lower and strict upper solutions for the continuous problem, and that f and G satisfy additional assumptions that are known to yield a priori bounds on, and to guarantee the existence of solutions of the continuous problem. Under these assumptions we show that there are at least three distinct solutions of the discrete approximation which approximate solutions to the continuous problem as the grid size, h, goes to 0. (C) 2003 Elsevier Science Ltd. All rights reserved.
Resumo:
A high definition, finite difference time domain (HD-FDTD) method is presented in this paper. This new method allows the FDTD method to be efficiently applied over a very large frequency range including low frequencies, which are problematic for conventional FDTD methods. In the method, no alterations to the properties of either the source or the transmission media are required. The method is essentially frequency independent and has been verified against analytical solutions within the frequency range 50 Hz-1 GHz. As an example of the lower frequency range, the method has been applied to the problem of induced eddy currents in the human body resulting from the pulsed magnetic field gradients of an MRI system. The new method only requires approximately 0.3% of the source period to obtain an accurate solution. (C) 2003 Elsevier Science Inc. All rights reserved.
Resumo:
Regional commodity forecasts are being used increasingly in agricultural industries to enhance their risk management and decision-making processes. These commodity forecasts are probabilistic in nature and are often integrated with a seasonal climate forecast system. The climate forecast system is based on a subset of analogue years drawn from the full climatological distribution. In this study we sought to measure forecast quality for such an integrated system. We investigated the quality of a commodity (i.e. wheat and sugar) forecast based on a subset of analogue years in relation to a standard reference forecast based on the full climatological set. We derived three key dimensions of forecast quality for such probabilistic forecasts: reliability, distribution shift, and change in dispersion. A measure of reliability was required to ensure no bias in the forecast distribution. This was assessed via the slope of the reliability plot, which was derived from examination of probability levels of forecasts and associated frequencies of realizations. The other two dimensions related to changes in features of the forecast distribution relative to the reference distribution. The relationship of 13 published accuracy/skill measures to these dimensions of forecast quality was assessed using principal component analysis in case studies of commodity forecasting using seasonal climate forecasting for the wheat and sugar industries in Australia. There were two orthogonal dimensions of forecast quality: one associated with distribution shift relative to the reference distribution and the other associated with relative distribution dispersion. Although the conventional quality measures aligned with these dimensions, none measured both adequately. We conclude that a multi-dimensional approach to assessment of forecast quality is required and that simple measures of reliability, distribution shift, and change in dispersion provide a means for such assessment. The analysis presented was also relevant to measuring quality of probabilistic seasonal climate forecasting systems. The importance of retaining a focus on the probabilistic nature of the forecast and avoiding simplifying, but erroneous, distortions was discussed in relation to applying this new forecast quality assessment paradigm to seasonal climate forecasts. Copyright (K) 2003 Royal Meteorological Society.
Resumo:
This paper describes the modification of a two-dimensional finite element long wave hydrodynamic model in order to predict the net current and water levels attributable to the influences of waves. Tests examine the effects of the application of wave induced forces, including comparisons to a physical experiment. An example of a real river system is presented with comparisons to measured data, which demonstrate the importance of simulating the combined effects of tides and waves upon hydrodynamic behavior. (C) 2002 Elsevier Science Ltd. All rights reserved.
Resumo:
We analyze the sequences of round-off errors of the orbits of a discretized planar rotation, from a probabilistic angle. It was shown [Bosio & Vivaldi, 2000] that for a dense set of parameters, the discretized map can be embedded into an expanding p-adic dynamical system, which serves as a source of deterministic randomness. For each parameter value, these systems can generate infinitely many distinct pseudo-random sequences over a finite alphabet, whose average period is conjectured to grow exponentially with the bit-length of the initial condition (the seed). We study some properties of these symbolic sequences, deriving a central limit theorem for the deviations between round-off and exact orbits, and obtain bounds concerning repetitions of words. We also explore some asymptotic problems computationally, verifying, among other things, that the occurrence of words of a given length is consistent with that of an abstract Bernoulli sequence.
Resumo:
Within the development of motor vehicles, crash safety (e.g. occupant protection, pedestrian protection, low speed damageability), is one of the most important attributes. In order to be able to fulfill the increased requirements in the framework of shorter cycle times and rising pressure to reduce costs, car manufacturers keep intensifying the use of virtual development tools such as those in the domain of Computer Aided Engineering (CAE). For crash simulations, the explicit finite element method (FEM) is applied. The accuracy of the simulation process is highly dependent on the accuracy of the simulation model, including the midplane mesh. One of the roughest approximations typically made is the actual part thickness which, in reality, can vary locally. However, almost always a constant thickness value is defined throughout the entire part due to complexity reasons. On the other hand, for precise fracture analysis within FEM, the correct thickness consideration is one key enabler. Thus, availability of per element thickness information, which does not exist explicitly in the FEM model, can significantly contribute to an improved crash simulation quality, especially regarding fracture prediction. Even though the thickness is not explicitly available from the FEM model, it can be inferred from the original CAD geometric model through geometric calculations. This paper proposes and compares two thickness estimation algorithms based on ray tracing and nearest neighbour 3D range searches. A systematic quantitative analysis of the accuracy of both algorithms is presented, as well as a thorough identification of particular geometric arrangements under which their accuracy can be compared. These results enable the identification of each technique’s weaknesses and hint towards a new, integrated, approach to the problem that linearly combines the estimates produced by each algorithm.
Resumo:
Pectus carinatum (PC) is a chest deformity caused by a disproportionate growth of the costal cartilages compared to the bony thoracic skeleton, pulling the sternum towards, which leads to its protrusion. There has been a growing interest on using the ‘reversed Nuss’ technique as minimally invasive procedure for PC surgical correction. A corrective bar is introduced between the skin and the thoracic cage and positioned on top of the sternum highest protrusion area for continuous pressure. Then, it is fixed to the ribs and kept implanted for about 2–3 years. The purpose of this work was to (a) assess the stresses distribution on the thoracic cage that arise from the procedure, and (b) investigate the impact of different positioning of the corrective bar along the sternum. The higher stresses were generated on the 4th, 5th and 6th ribs backend, supporting the hypothesis of pectus deformities correction-induced scoliosis. The different bar positioning originated different stresses on the ribs’ backend. The bar position that led to lower stresses generated on the ribs backend was the one that also led to the smallest sternum displacement. However, this may be preferred, as the risk of induced scoliosis is lowered.
Resumo:
A package of B-spline finite strip models is developed for the linear analysis of piezolaminated plates and shells. This package is associated to a global optimization technique in order to enhance the performance of these types of structures, subjected to various types of objective functions and/or constraints, with discrete and continuous design variables. The models considered are based on a higher-order displacement field and one can apply them to the static, free vibration and buckling analyses of laminated adaptive structures with arbitrary lay-ups, loading and boundary conditions. Genetic algorithms, with either binary or floating point encoding of design variables, were considered to find optimal locations of piezoelectric actuators as well as to determine the best voltages applied to them in order to obtain a desired structure shape. These models provide an overall economy of computing effort for static and vibration problems.
Resumo:
The main properties of strangelets, namely their energy per baryon, radius and electric charge, are studied in the unpaired magnetized strange quark matter (MSQM) and paired magnetized colour flavour locked (MCFL) phases. Temperature effects are taken into account in order to study their stability compared to the Fe-56 isotope and nonmagnetized strangelets within the framework of the MIT bag model. We conclude that the presence of a magnetic field tends to stabilize the strangelets more, even when temperature is considered. It is also shown that MCFL strangelets are more stable than ordinary MSQM strangelets for typical gap values of the order of O(100) MeV. A distinctive feature in the detection of strangelets either in cosmic rays or in heavy-ion collider experiments could be their electric charge. We find that the electric charge is modified in the presence of the magnetic field, leading to higher (lower) charge values for MSQM (MCFL) strangelets, when compared to the nonmagnetized case.