211 resultados para prediction error


Relevância:

20.00% 20.00%

Publicador:

Resumo:

Designing and optimizing high performance microprocessors is an increasingly difficult task due to the size and complexity of the processor design space, high cost of detailed simulation and several constraints that a processor design must satisfy. In this paper, we propose the use of empirical non-linear modeling techniques to assist processor architects in making design decisions and resolving complex trade-offs. We propose a procedure for building accurate non-linear models that consists of the following steps: (i) selection of a small set of representative design points spread across processor design space using latin hypercube sampling, (ii) obtaining performance measures at the selected design points using detailed simulation, (iii) building non-linear models for performance using the function approximation capabilities of radial basis function networks, and (iv) validating the models using an independently and randomly generated set of design points. We evaluate our model building procedure by constructing non-linear performance models for programs from the SPEC CPU2000 benchmark suite with a microarchitectural design space that consists of 9 key parameters. Our results show that the models, built using a relatively small number of simulations, achieve high prediction accuracy (only 2.8% error in CPI estimates on average) across a large processor design space. Our models can potentially replace detailed simulation for common tasks such as the analysis of key microarchitectural trends or searches for optimal processor design points.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

With the introduction of 2D flat-panel X-ray detectors, 3D image reconstruction using helical cone-beam tomography is fast replacing the conventional 2D reconstruction techniques. In 3D image reconstruction, the source orbit or scanning geometry should satisfy the data sufficiency or completeness condition for exact reconstruction. The helical scan geometry satisfies this condition and hence can give exact reconstruction. The theoretically exact helical cone-beam reconstruction algorithm proposed by Katsevich is a breakthrough and has attracted interest in the 3D reconstruction using helical cone-beam Computed Tomography.In many practical situations, the available projection data is incomplete. One such case is where the detector plane does not completely cover the full extent of the object being imaged in lateral direction resulting in truncated projections. This result in artifacts that mask small features near to the periphery of the ROI when reconstructed using the convolution back projection (CBP) method assuming that the projection data is complete. A number of techniques exist which deal with completion of missing data followed by the CBP reconstruction. In 2D, linear prediction (LP)extrapolation has been shown to be efficient for data completion, involving minimal assumptions on the nature of the data, producing smooth extensions of the missing projection data.In this paper, we propose to extend the LP approach for extrapolating helical cone beam truncated data. The projection on the multi row flat panel detectors has missing columns towards either ends in the lateral direction in truncated data situation. The available data from each detector row is modeled using a linear predictor. The available data is extrapolated and this completed projection data is backprojected using the Katsevich algorithm. Simulation results show the efficacy of the proposed method.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Equations are developed for predicting the activity coefficients of oxygen dissolved in ternary liquid alloys. These are extensions of earlier treatments, and are based on a model in which each oxygen atom is assumed to make four bonds with neighboring metal atoms. It is also postulated that the strong oxygen-metal bonds distort the electronic configuration around the metal atoms bonded to oxygen, and that the quantitative reduction of the strength of bonds made by these atoms with all of the adjacent metal atoms is equivalent to a factor of approximately two. The predictions of the quasichemical equation which is derived agree satisfactorily with the partial molar free energies of oxygen in Ag-Cu-Sn solutions at 1200°C reported in literature. An extension of this treatment to multicomponent solutions is also indicated.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The contemporary methods for source characterization rely mainly on experiments. These methods produce inaccurate results in the low‐frequency band, where the characteristics are all the more important. Moreover, the experimental methods cannot be used at the design stage. Hence, a numerical technique to obtain the source characteristics is desirable. In this paper, the pressure‐time history and the mass‐flux‐time history obtained by means of the time‐domain analysis have been used, along with the two‐load method to compute the source characteristics. Two new computational methods for obtaining the source characteristics have been described. These are much simpler, and computationally more economical than the complete time‐domain simulation, which makes use of the method of characteristics.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

He propose a new time domain method for efficient representation of the KCG and delineation of its component waves. The method is based on the multipulse Linear prediction (LP) coding which is being widely used in speech processing. The excitation to the LP synthesis filter consists of a few pulses defined by their locations and amplitudes. Based on the amplitudes and their distribution, the pulses are suitably combined to delineate the component waves. Beat to beat correlation in the ECG signal is used in QRS periodicity prediction. The method entails a data compression of 1 in 6. The method reconstructs the signal with an NMSE of less than 5%.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Prediction of variable bit rate compressed video traffic is critical to dynamic allocation of resources in a network. In this paper, we propose a technique for preprocessing the dataset used for training a video traffic predictor. The technique involves identifying the noisy instances in the data using a fuzzy inference system. We focus on three prediction techniques, namely, linear regression, neural network and support vector regression and analyze their performance on H.264 video traces. Our experimental results reveal that data preprocessing greatly improves the performance of linear regression and neural network, but is not effective on support vector regression.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Evaluation of the probability of error in decision feedback equalizers is difficult due to the presence of a hard limiter in the feedback path. This paper derives the upper and lower bounds on the probability of a single error and multiple error patterns. The bounds are fairly tight. The bounds can also be used to select proper tap gains of the equalizer.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Upper bounds on the probability of error due to co-channel interference are proposed in this correspondence. The bounds are easy to compute and can be fairly tight.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Breakout noise from HVAC ducts is important at low frequencies, and the coupling between the acoustic waves and the structural waves plays a critical role in the prediction of the transverse transmission loss. This paper describes the analytical calculation of breakout noise by incorporating three-dimensional effects along with the acoustical and structural wave coupling phenomena. The first step in the breakout noise prediction is to calculate the inside duct pressure field and the normal duct wall vibration by using the solution of the governing differential equations in terms of Green's function. The resultant equations are rearranged in terms of impedance and mobility, which results in a compact matrix formulation. The Green's function selected for the current problem is the cavity Green's function with modification of wave number in the longitudinal direction in order to incorporate the terminal impedance. The second step is to calculate the radiated sound power from the compliant duct walls by means of an ``equivalent unfolded plate'' model. The transverse transmission loss from the duct walls is calculated using the ratio of the incident power due to surface source inside the duct to the acoustic power radiated from the compliant duct walls. Analytical results are validated with the FE-BE numerical models.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The questions that one should answer in engineering computations - deterministic, probabilistic/randomized, as well as heuristic - are (i) how good the computed results/outputs are and (ii) how much the cost in terms of amount of computation and the amount of storage utilized in getting the outputs is. The absolutely errorfree quantities as well as the completely errorless computations done in a natural process can never be captured by any means that we have at our disposal. While the computations including the input real quantities in nature/natural processes are exact, all the computations that we do using a digital computer or are carried out in an embedded form are never exact. The input data for such computations are also never exact because any measuring instrument has inherent error of a fixed order associated with it and this error, as a matter of hypothesis and not as a matter of assumption, is not less than 0.005 per cent. Here by error we imply relative error bounds. The fact that exact error is never known under any circumstances and any context implies that the term error is nothing but error-bounds. Further, in engineering computations, it is the relative error or, equivalently, the relative error-bounds (and not the absolute error) which is supremely important in providing us the information regarding the quality of the results/outputs. Another important fact is that inconsistency and/or near-consistency in nature, i.e., in problems created from nature is completely nonexistent while in our modelling of the natural problems we may introduce inconsistency or near-inconsistency due to human error or due to inherent non-removable error associated with any measuring device or due to assumptions introduced to make the problem solvable or more easily solvable in practice. Thus if we discover any inconsistency or possibly any near-inconsistency in a mathematical model, it is certainly due to any or all of the three foregoing factors. We do, however, go ahead to solve such inconsistent/near-consistent problems and do get results that could be useful in real-world situations. The talk considers several deterministic, probabilistic, and heuristic algorithms in numerical optimisation, other numerical and statistical computations, and in PAC (probably approximately correct) learning models. It highlights the quality of the results/outputs through specifying relative error-bounds along with the associated confidence level, and the cost, viz., amount of computations and that of storage through complexity. It points out the limitation in error-free computations (wherever possible, i.e., where the number of arithmetic operations is finite and is known a priori) as well as in the usage of interval arithmetic. Further, the interdependence among the error, the confidence, and the cost is discussed.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

In the present paper, the constitutive model is proposed for cemented soils, in which the cementation component and frictional component are treated separately and then added together to get overall response. The modified Cam clay is used to predict the frictional resistance and an elasto-plastic strain softening model is proposed for the cementation component. The rectangular isotropic yield curve proposed by Vatsala (1995) for the bond component has been modified in order to account for the anisotropy generally observed in the case of natural soft cemented soils. In this paper, the model proposed is used to predict the experimental results of extension tests on the soft cemented soils whereas compression test results are presented elsewhere. The model predictions compare quite satisfactorily with the observed response. A few input parameters are required which are well defined and easily determinable and the model uses associated flow rule.