234 resultados para relative utility models
Resumo:
Calculating the potentials on the heart’s epicardial surface from the body surface potentials constitutes one form of inverse problems in electrocardiography (ECG). Since these problems are ill-posed, one approach is to use zero-order Tikhonov regularization, where the squared norms of both the residual and the solution are minimized, with a relative weight determined by the regularization parameter. In this paper, we used three different methods to choose the regularization parameter in the inverse solutions of ECG. The three methods include the L-curve, the generalized cross validation (GCV) and the discrepancy principle (DP). Among them, the GCV method has received less attention in solutions to ECG inverse problems than the other methods. Since the DP approach needs knowledge of norm of noises, we used a model function to estimate the noise. The performance of various methods was compared using a concentric sphere model and a real geometry heart-torso model with a distribution of current dipoles placed inside the heart model as the source. Gaussian measurement noises were added to the body surface potentials. The results show that the three methods all produce good inverse solutions with little noise; but, as the noise increases, the DP approach produces better results than the L-curve and GCV methods, particularly in the real geometry model. Both the GCV and L-curve methods perform well in low to medium noise situations.
Resumo:
We investigate here a modification of the discrete random pore model [Bhatia SK, Vartak BJ, Carbon 1996;34:1383], by including an additional rate constant which takes into account the different reactivity of the initial pore surface having attached functional groups and hydrogens, relative to the subsequently exposed surface. It is observed that the relative initial reactivity has a significant effect on the conversion and structural evolution, underscoring the importance of initial surface chemistry. The model is tested against experimental data on chemically controlled char oxidation and steam gasification at various temperatures. It is seen that the variations of the reaction rate and surface area with conversion are better represented by the present approach than earlier random pore models. The results clearly indicate the improvement of model predictions in the low conversion region, where the effect of the initially attached functional groups and hydrogens is more significant, particularly for char oxidation. It is also seen that, for the data examined, the initial surface chemistry is less important for steam gasification as compared to the oxidation reaction. Further development of the approach must also incorporate the dynamics of surface complexation, which is not considered here.
Resumo:
Granule impact deformation has long been recognised as important in determining whether or not two colliding granules will coalesce. Work in the last 10 years has highlighted the fact that viscous effects are significant in granulation. The relative strengths of different formulations can vary with strain rate. Therefore, traditional strength measurements made at pseudo-static conditions give no indication, even qualitatively, of how materials will behave at high strain rates, and hence are actually misleading when used to model granule coalescence. This means that new standard methods need to be developed for determining the strain rates encountered by granules inside industrial equipment and also for measuring the mechanical properties of granules at these strain rates. The constitutive equations used in theoretical models of granule coalescence also need to be extended to include strain-rate dependent components.
Resumo:
Holden utility carrying members of the Federated Ship Painters and Dockers Union during the Labour Day march in 1965, Brisbane, Australia. Anti conscription banners can be seen in the background, and the facade of the Pearl Assurance Building.
Resumo:
The relative importance of factors that may promote genetic differentiation in marine organisms is largely unknown. Here, contributions to population structure from biogeography, habitat distribution, and isolation by distance were investigated in Axoclinus nigricaudus, a small subtidal rock reef fish, throughout its range in the Gulf of California. A 408 basepair fragment of the mitochondrial control region was sequenced from 105 individuals. Variation was significantly partitioned between many pairs of populations. Phylogenetic analyses, hierarchical analyses of variance, and general linear models substantiated a major break between two putative biogeographic regions. This genetic discontinuity coincides with an abrupt change in ecological characteristics (including temperature and salinity) but does not coincide with known oceanographic circulation patterns. Geographic distance and the nature of habitat separating populations (continuous habitat along a shoreline, discontinuous habitat along a shoreline, and open water) also contributed to population structure in general linear model analyses. To verify that local populations are genetically stable over time, one population was resampled on four occasions over eighteen months; it showed no evidence of a temporal component to diversity. These results indicate that having a planktonic life stage does not preclude geographically partitioned genetic variation over relatively small geographic distances in marine environments. Moreover, levels of genetic differentiation among populations of Axoclinus nigricaudus cannot be explained by a single factor, but are due to the combined influences of a biogeographic boundary, habitat, and geographic distance.
Resumo:
The Gaudin models based on the face-type elliptic quantum groups and the XYZ Gaudin models are studied. The Gaudin model Hamiltonians are constructed and are diagonalized by using the algebraic Bethe ansatz method. The corresponding face-type Knizhnik–Zamolodchikov equations and their solutions are given.
Resumo:
In this review we demonstrate how the algebraic Bethe ansatz is used for the calculation of the-energy spectra and form factors (operator matrix elements in the basis of Hamiltonian eigenstates) in exactly solvable quantum systems. As examples we apply the theory to several models of current interest in the study of Bose-Einstein condensates, which have been successfully created using ultracold dilute atomic gases. The first model we introduce describes Josephson tunnelling between two coupled Bose-Einstein condensates. It can be used not only for the study of tunnelling between condensates of atomic gases, but for solid state Josephson junctions and coupled Cooper pair boxes. The theory is also applicable to models of atomic-molecular Bose-Einstein condensates, with two examples given and analysed. Additionally, these same two models are relevant to studies in quantum optics; Finally, we discuss the model of Bardeen, Cooper and Schrieffer in this framework, which is appropriate for systems of ultracold fermionic atomic gases, as well as being applicable for the description of superconducting correlations in metallic grains with nanoscale dimensions.; In applying all the above models to. physical situations, the need for an exact analysis of small-scale systems is established due to large quantum fluctuations which render mean-field approaches inaccurate.
Resumo:
Many images consist of two or more 'phases', where a phase is a collection of homogeneous zones. For example, the phases may represent the presence of different sulphides in an ore sample. Frequently, these phases exhibit very little structure, though all connected components of a given phase may be similar in some sense. As a consequence, random set models are commonly used to model such images. The Boolean model and models derived from the Boolean model are often chosen. An alternative approach to modelling such images is to use the excursion sets of random fields to model each phase. In this paper, the properties of excursion sets will be firstly discussed in terms of modelling binary images. Ways of extending these models to multi-phase images will then be explored. A desirable feature of any model is to be able to fit it to data reasonably well. Different methods for fitting random set models based on excursion sets will be presented and some of the difficulties with these methods will be discussed.
Resumo:
While a number of studies have shown that object-extracted relative clauses are more difficult to understand than subject-extracted counterparts for second language (L2) English learners (e.g., Izumi, 2003), less is known about why this is the case and how they process these complex sentences. This exploratory study examines the potential applicability of Gibson's (1998, 2000) Syntactic Prediction Locality Theory (SPLT), a theory proposed to predict first language (L1) processing difficulty, to L2 processing and considers whether the theory might also account for the processing difficulties of subject- and object-extracted relative clauses encountered by L2 learners. Results of a self-paced reading time experiment from 15 Japanese learners of English are mainly consistent with the reading time profile predicted by the SPLT and thus suggest that the L1 processing theory might also be able to account for L2 processing difficulty.
Resumo:
Except for a few large scale projects, language planners have tended to talk and argue among themselves rather than to see language policy development as an inherently political process. A comparison with a social policy example, taken from the United States, suggests that it is important to understand the problem and to develop solutions in the context of the political process, as this is where decisions will ultimately be made.
Resumo:
Polytomous Item Response Theory Models provides a unified, comprehensive introduction to the range of polytomous models available within item response theory (IRT). It begins by outlining the primary structural distinction between the two major types of polytomous IRT models. This focuses on the two types of response probability that are unique to polytomous models and their associated response functions, which are modeled differently by the different types of IRT model. It describes, both conceptually and mathematically, the major specific polytomous models, including the Nominal Response Model, the Partial Credit Model, the Rating Scale model, and the Graded Response Model. Important variations, such as the Generalized Partial Credit Model are also described as are less common variations, such as the Rating Scale version of the Graded Response Model. Relationships among the models are also investigated and the operation of measurement information is described for each major model. Practical examples of major models using real data are provided, as is a chapter on choosing an appropriate model. Figures are used throughout to illustrate important elements as they are described.
Resumo:
This paper discusses a multi-layer feedforward (MLF) neural network incident detection model that was developed and evaluated using field data. In contrast to published neural network incident detection models which relied on simulated or limited field data for model development and testing, the model described in this paper was trained and tested on a real-world data set of 100 incidents. The model uses speed, flow and occupancy data measured at dual stations, averaged across all lanes and only from time interval t. The off-line performance of the model is reported under both incident and non-incident conditions. The incident detection performance of the model is reported based on a validation-test data set of 40 incidents that were independent of the 60 incidents used for training. The false alarm rates of the model are evaluated based on non-incident data that were collected from a freeway section which was video-taped for a period of 33 days. A comparative evaluation between the neural network model and the incident detection model in operation on Melbourne's freeways is also presented. The results of the comparative performance evaluation clearly demonstrate the substantial improvement in incident detection performance obtained by the neural network model. The paper also presents additional results that demonstrate how improvements in model performance can be achieved using variable decision thresholds. Finally, the model's fault-tolerance under conditions of corrupt or missing data is investigated and the impact of loop detector failure/malfunction on the performance of the trained model is evaluated and discussed. The results presented in this paper provide a comprehensive evaluation of the developed model and confirm that neural network models can provide fast and reliable incident detection on freeways. (C) 1997 Elsevier Science Ltd. All rights reserved.
Resumo:
The conventional analysis for the estimation of the tortuosity factor for transport in porous media is modified here to account for the effect of pore aspect ratio. Structural models of the porous medium are also constructed for calculating the aspect ratio as a function of porosity. Comparison of the model predictions with the extensive data of Currie (1960) for the effective diffusivity of hydrogen in packed beds shows good agreement with a network model of randomly oriented intersecting pores for porosities upto about 50 percent, which is the region of practical interest. The predictions based on this network model are also found to be in better agreement with the data of Currie than earlier expressions developed for unconsolidated and grainy media.
Resumo:
Traditional waste stabilisation pond (WSP) models encounter problems predicting pond performance because they cannot account for the influence of pond features, such as inlet structure or pond geometry, on fluid hydrodynamics. In this study, two dimensional (2-D) computational fluid dynamics (CFD) models were compared to experimental residence time distributions (RTD) from literature. In one of the-three geometries simulated, the 2-D CFD model successfully predicted the experimental RTD. However, flow patterns in the other two geometries were not well described due to the difficulty of representing the three dimensional (3-D) experimental inlet in the 2-D CFD model, and the sensitivity of the model results to the assumptions used to characterise the inlet. Neither a velocity similarity nor geometric similarity approach to inlet representation in 2-D gave results correlating with experimental data. However. it was shown that 2-D CFD models were not affected by changes in values of model parameters which are difficult to predict, particularly the turbulent inlet conditions. This work suggests that 2-D CFD models cannot be used a priori to give an adequate description of the hydrodynamic patterns in WSP. (C) 1998 Elsevier Science Ltd. All rights reserved.