866 resultados para split-operator scheme


Relevância:

30.00% 30.00%

Publicador:

Resumo:

Ozone and temperature profiles from the Michelson Interferometer for Passive Atmospheric Sounding (MIPAS) have been assimilated, using three-dimensional variational assimilation, into a stratosphere troposphere version of the Met Office numerical weather-prediction system. Analyses are made for the month of September 2002, when there was an unprecedented split in the southern hemisphere polar vortex. The analyses are validated against independent ozone observations from sondes, limb-occultation and total column ozone satellite instruments. Through most of the stratosphere, precision varies from 5 to 15%, and biases are 15% or less of the analysed field. Problems remain in the vortex and below the 60 hPa. level, especially at the tropopause where the analyses have too much ozone and poor agreement with independent data. Analysis problems are largely a result of the model rather than the data, giving confidence in the MIPAS ozone retrievals, though there may be a small high bias in MIPAS ozone in the lower stratosphere. Model issues include an excessive Brewer-Dobson circulation, which results both from known problems with the tracer transport scheme and from the data assimilation of dynamical variables. The extreme conditions of the vortex split reveal large differences between existing linear ozone photochemistry schemes. Despite these issues, the ozone analyses are able to successfully describe the ozone hole split and compare well to other studies of this event. Recommendations are made for the further development of the ozone assimilation system.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

An efficient algorithm is presented for the solution of the steady Euler equations of gas dynamics. The scheme is based on solving linearised Riemann problems approximately and in more than one dimension incorporates operator splitting. The scheme is applied to a standard test problem of flow down a channel containing a circular arc bump for three different mesh sizes.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

A finite difference scheme based on flux difference splitting is presented for the solution of the two-dimensional shallow water equations of ideal fluid flow. A linearised problem, analogous to that of Riemann for gas dynamics is defined, and a scheme, based on numerical characteristic decomposition is presented for obtaining approximate solutions to the linearised problem, and incorporates the technique of operator splitting. An average of the flow variables across the interface between cells is required, and this average is chosen to be the arithmetic mean for computational efficiency leading to arithmetic averaging. This is in contrast to usual ‘square root’ averages found in this type of Riemann solver, where the computational expense can be prohibitive. The method of upwind differencing is used for the resulting scalar problems, together with a flux limiter for obtaining a second order scheme which avoids nonphysical, spurious oscillations. An extension to the two-dimensional equations with source terms is included. The scheme is applied to the one-dimensional problems of a breaking dam and reflection of a bore, and in each case the approximate solution is compared to the exact solution of ideal fluid flow. The scheme is also applied to a problem of stationary bore generation in a channel of variable cross-section. Finally, the scheme is applied to two other dam-break problems, this time in two dimensions with one having cylindrical symmetry. Each approximate solution compares well with those given by other authors.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The scheme is based on Ami Harten's ideas (Harten, 1994), the main tools coming from wavelet theory, in the framework of multiresolution analysis for cell averages. But instead of evolving cell averages on the finest uniform level, we propose to evolve just the cell averages on the grid determined by the significant wavelet coefficients. Typically, there are few cells in each time step, big cells on smooth regions, and smaller ones close to irregularities of the solution. For the numerical flux, we use a simple uniform central finite difference scheme, adapted to the size of each cell. If any of the required neighboring cell averages is not present, it is interpolated from coarser scales. But we switch to ENO scheme in the finest part of the grids. To show the feasibility and efficiency of the method, it is applied to a system arising in polymer-flooding of an oil reservoir. In terms of CPU time and memory requirements, it outperforms Harten's multiresolution algorithm.The proposed method applies to systems of conservation laws in 1Dpartial derivative(t)u(x, t) + partial derivative(x)f(u(x, t)) = 0, u(x, t) is an element of R-m. (1)In the spirit of finite volume methods, we shall consider the explicit schemeupsilon(mu)(n+1) = upsilon(mu)(n) - Deltat/hmu ((f) over bar (mu) - (f) over bar (mu)-) = [Dupsilon(n)](mu), (2)where mu is a point of an irregular grid Gamma, mu(-) is the left neighbor of A in Gamma, upsilon(mu)(n) approximate to 1/mu-mu(-) integral(mu-)(mu) u(x, t(n))dx are approximated cell averages of the solution, (f) over bar (mu) = (f) over bar (mu)(upsilon(n)) are the numerical fluxes, and D is the numerical evolution operator of the scheme.According to the definition of (f) over bar (mu), several schemes of this type have been proposed and successfully applied (LeVeque, 1990). Godunov, Lax-Wendroff, and ENO are some of the popular names. Godunov scheme resolves well the shocks, but accuracy (of first order) is poor in smooth regions. Lax-Wendroff is of second order, but produces dangerous oscillations close to shocks. ENO schemes are good alternatives, with high order and without serious oscillations. But the price is high computational cost.Ami Harten proposed in (Harten, 1994) a simple strategy to save expensive ENO flux calculations. The basic tools come from multiresolution analysis for cell averages on uniform grids, and the principle is that wavelet coefficients can be used for the characterization of local smoothness.. Typically, only few wavelet coefficients are significant. At the finest level, they indicate discontinuity points, where ENO numerical fluxes are computed exactly. Elsewhere, cheaper fluxes can be safely used, or just interpolated from coarser scales. Different applications of this principle have been explored by several authors, see for example (G-Muller and Muller, 1998).Our scheme also uses Ami Harten's ideas. But instead of evolving the cell averages on the finest uniform level, we propose to evolve the cell averages on sparse grids associated with the significant wavelet coefficients. This means that the total number of cells is small, with big cells in smooth regions and smaller ones close to irregularities. This task requires improved new tools, which are described next.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Objective: To evaluate the velocity of canine retraction, anchorage loss and changes on canine and first molar inclinations using self-ligating and conventional brackets.Materials and Methods: Twenty-five adults with Class I malocclusion and a treatment plan involving extractions of four first premolars were selected for this randomized split-mouth control trial. Patients had either conventional or self-ligating brackets bonded to maxillary canines randomly. Retraction was accomplished using 100-g nickel-titanium closed coil springs, which were reactivated every 4 weeks. Oblique radiographs were taken before and after canine retraction was completed, and the cephalograms were superimposed on stable structures of the maxilla. Cephalometric points were digitized twice by a blinded operator for error control, and the following landmarks were collected: canine cusp and apex horizontal changes, molar cusp and apex horizontal changes, and angulation changes in canines and molars. The blinded data, which were normally distributed, were analyzed through paired t-tests for group differences.Results: No differences were found between the two groups for all variables tested.Conclusions: Both brackets showed the same velocity of canine retraction and loss of anteroposterior anchorage of the molars. No changes were found between brackets regarding the inclination of canines and first molars.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Different tools have been used to set up and adopt the model for the fulfillment of the objective of this research. 1. The Model The base model that has been used is the Analytical Hierarchy Process (AHP) adapted with the aim to perform a Benefit Cost Analysis. The AHP developed by Thomas Saaty is a multicriteria decision - making technique which decomposes a complex problem into a hierarchy. It is used to derive ratio scales from both discreet and continuous paired comparisons in multilevel hierarchic structures. These comparisons may be taken from actual measurements or from a fundamental scale that reflects the relative strength of preferences and feelings. 2. Tools and methods 2.1. The Expert Choice Software The software Expert Choice is a tool that allows each operator to easily implement the AHP model in every stage of the problem. 2.2. Personal Interviews to the farms For this research, the farms of the region Emilia Romagna certified EMAS have been detected. Information has been given by EMAS center in Wien. Personal interviews have been carried out to each farm in order to have a complete and realistic judgment of each criteria of the hierarchy. 2.3. Questionnaire A supporting questionnaire has also been delivered and used for the interviews . 3. Elaboration of the data After data collection, the data elaboration has taken place. The software support Expert Choice has been used . 4. Results of the Analysis The result of the figures above (vedere altro documento) gives a series of numbers which are fractions of the unit. This has to be interpreted as the relative contribution of each element to the fulfillment of the relative objective. So calculating the Benefits/costs ratio for each alternative the following will be obtained: Alternative One: Implement EMAS Benefits ratio: 0, 877 Costs ratio: 0, 815 Benfit/Cost ratio: 0,877/0,815=1,08 Alternative Two: Not Implement EMAS Benefits ratio: 0,123 Costs ration: 0,185 Benefit/Cost ratio: 0,123/0,185=0,66 As stated above, the alternative with the highest ratio will be the best solution for the organization. This means that the research carried out and the model implemented suggests that EMAS adoption in the agricultural sector is the best alternative. It has to be noted that the ratio is 1,08 which is a relatively low positive value. This shows the fragility of this conclusion and suggests a careful exam of the benefits and costs for each farm before adopting the scheme. On the other part, the result needs to be taken in consideration by the policy makers in order to enhance their intervention regarding the scheme adoption on the agricultural sector. According to the AHP elaboration of judgments we have the following main considerations on Benefits: - Legal compliance seems to be the most important benefit for the agricultural sector since its rank is 0,471 - The next two most important benefits are Improved internal organization (ranking 0,230) followed by Competitive advantage (ranking 0, 221) mostly due to the sub-element Improved image (ranking 0,743) Finally, even though Incentives are not ranked among the most important elements, the financial ones seem to have been decisive on the decision making process. According to the AHP elaboration of judgments we have the following main considerations on Costs: - External costs seem to be largely more important than the internal ones (ranking 0, 857 over 0,143) suggesting that Emas costs over consultancy and verification remain the biggest obstacle. - The implementation of the EMS is the most challenging element regarding the internal costs (ranking 0,750).

Relevância:

30.00% 30.00%

Publicador:

Resumo:

X-ray absorption spectroscopy (XAS) is a powerful means of investigation of structural and electronic properties in condensed -matter physics. Analysis of the near edge part of the XAS spectrum, the so – called X-ray Absorption Near Edge Structure (XANES), can typically provide the following information on the photoexcited atom: - Oxidation state and coordination environment. - Speciation of transition metal compounds. - Conduction band DOS projected on the excited atomic species (PDOS). Analysis of XANES spectra is greatly aided by simulations; in the most common scheme the multiple scattering framework is used with the muffin tin approximation for the scattering potential and the spectral simulation is based on a hypothetical, reference structure. This approach has the advantage of requiring relatively little computing power but in many cases the assumed structure is quite different from the actual system measured and the muffin tin approximation is not adequate for low symmetry structures or highly directional bonds. It is therefore very interesting and justified to develop alternative methods. In one approach, the spectral simulation is based on atomic coordinates obtained from a DFT (Density Functional Theory) optimized structure. In another approach, which is the object of this thesis, the XANES spectrum is calculated directly based on an ab – initio DFT calculation of the atomic and electronic structure. This method takes full advantage of the real many-electron final wavefunction that can be computed with DFT algorithms that include a core-hole in the absorbing atom to compute the final cross section. To calculate the many-electron final wavefunction the Projector Augmented Wave method (PAW) is used. In this scheme, the absorption cross section is written in function of several contributions as the many-electrons function of the finale state; it is calculated starting from pseudo-wavefunction and performing a reconstruction of the real-wavefunction by using a transform operator which contains some parameters, called partial waves and projector waves. The aim of my thesis is to apply and test the PAW methodology to the calculation of the XANES cross section. I have focused on iron and silicon structures and on some biological molecules target (myoglobin and cytochrome c). Finally other inorganic and biological systems could be taken into account for future applications of this methodology, which could become an important improvement with respect to the multiscattering approach.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

OBJECTIVES: The aim of the present split-mouth study is to assess the peri-implant conditions around early-loaded sandblasted and acid-etched (SLA) implants, 5 years after abutment connection and to compare, in the same patients, the results obtained with a standard protocol using identical implants with a TPS surface. MATERIAL AND METHODS: Surgical procedure was performed by the same operator and was identical at test (SLA) and control (TPS) sites, in 32 healthy patients. Abutment connection was carried out at 35 N cm 6 weeks postsurgery for test sites and 12 weeks for the controls. Patients were seen regularly, for control and professional cleaning. At 60 months, clinical measures and radiographic bone changes were recorded by the same operator, blind to the type of surface of the implant, on 27 patients, as five patients were lost to follow-up. RESULTS: A total number of 106 implants were examined. No implant was lost. No significant differences were found with respect to the presence of plaque [modified plaque index (mPI) 0.27+/-0.56 vs. 0.32+/-0.54], bleeding on probing (29% vs. 32%), mean pocket depth (3.2+/-1 vs. 3.2+/-1 mm) or mean marginal bone loss (0.32+/-1.04 vs. 0.44+/-1.12 mm) between test and control. Four implants that presented 'spinning' at the time of abutment connection presented no significant differences from the rest of the test sites. CONCLUSION: The results of this prospective study confirm that SLA implants, under defined conditions, are suitable for early loading at 6 weeks in both the mandible and the maxilla. Limited implant spinning, occasionally found at abutment connection, produces no detrimental effect on the clinical outcome when properly handled.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

En este proyecto se han analizado distintas imágenes de fragmentos de rocas de distintas granulometrías correspondientes a una serie de voladuras de una misma cantera. Cada una de las voladuras se componen de 20 imágenes. A posteriori utilizando el programa Split Desktop en su versión 3.1, se delimitaron los fragmentos de roca de los que está compuesta la imagen, obteniéndose posteriormente la curva granulométrica correspondiente a dicha imagen. Una vez se calculan las curvas granulométricas correspondientes a cada imagen, se calcula la curva media de todas ellas, pudiéndose considerar por tanto la curva media de cada voladura. Se han utilizado las distintas soluciones del software, manual, online y automático, para realizar los análisis de dichas imágenes y a posteriori comparar sus resultados. Dichos resultados se muestran a través de una serie de gráficos y tablas que se explican con detalle para la comprensión del estudio. De dichos resultados es posible afirmar que, el tratamiento de imágenes realizado de manera online y automático por Split, desemboca en el mismo resultado, al no haber una diferencia estadística significativa. Por el contrario, el sistema manual es diferente de los otros dos, no pudiéndose afirmar cual es mejor de los dos. El manual depende del operario que trabaje las imágenes y el online de los ajustes realizados y por tanto, ambos tienen ciertas incertidumbres difíciles de solucionar. Abstract In this project, different images of rock fragments of different grain sizes corresponding to a series of blasts from the same quarry have been analyzed. To study each blast, 20 images has been used and studied with the software Split Desktop 3.1. Rock fragments from each image has been delimitated with the software, obtaining a grading curve of each one. Once these curves are calculated, the mean curve of these data set is obtained and can be considered the mean curve of each blast. Different software solutions as manual, online and automatic, has been used for the analysis of these images. Then the results has been compared between them. These results are shown through a series of graphs and tables, that are explained in detail, to enhance the understanding of the study. From these results, it can be said that the image processing with online and automatic options from Split, leads to the same result, after an statistical study. On the contrary, the manual Split mode is different from the others; however is not possible to assert what will be the best. The manual Split mode depends on the operator ability and dedication, although the online mode depends on the software settings, so therefore, both have some uncertainties that are difficult to solve.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The use of telerobotic systems is essential for remote handling (RH) operations in radioactive areas of scientific facilities that generate high doses of radiation. Recent developments in remote handling technology has seen a great deal of effort being directed towards the design of modular remote handling control rooms equipped with a standard master arm which will be used to separately control a range of different slave devices. This application thus requires a kinematically dissimilar master-slave control scheme. In order to avoid drag and other effects such as friction or other non-linear and unmodelled slave arm effects of the common position-position architecture in nonbackdrivable slaves, this research has implemented a force-position control scheme. End-effector force is derived from motor torque values which, to avoid the use of radiation intolerant and costly sensing devices, are inferred from motor current measurement. This has been demonstrated on a 1-DOF test-rig with a permanent magnet synchronous motor teleoperated by a Sensable Phantom Omni® haptic master. This has been shown to allow accurate control while realistically conveying dynamic force information back to the operator.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

A high resolution, second-order central difference method for incompressible flows is presented. The method is based on a recent second-order extension of the classic Lax–Friedrichs scheme introduced for hyperbolic conservation laws (Nessyahu H. & Tadmor E. (1990) J. Comp. Physics. 87, 408-463; Jiang G.-S. & Tadmor E. (1996) UCLA CAM Report 96-36, SIAM J. Sci. Comput., in press) and augmented by a new discrete Hodge projection. The projection is exact, yet the discrete Laplacian operator retains a compact stencil. The scheme is fast, easy to implement, and readily generalizable. Its performance was tested on the standard periodic double shear-layer problem; no spurious vorticity patterns appear when the flow is underresolved. A short discussion of numerical boundary conditions is also given, along with a numerical example.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This paper is an expanded and more detailed version of the work [1] in which the Operator Quantum Error Correction formalism was introduced. This is a new scheme for the error correction of quantum operations that incorporates the known techniques - i.e. the standard error correction model, the method of decoherence-free subspaces, and the noiseless subsystem method - as special cases, and relies on a generalized mathematical framework for noiseless subsystems that applies to arbitrary quantum operations. We also discuss a number of examples and introduce the notion of unitarily noiseless subsystems.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Computer networks produce tremendous amounts of event-based data that can be collected and managed to support an increasing number of new classes of pervasive applications. Examples of such applications are network monitoring and crisis management. Although the problem of distributed event-based management has been addressed in the non-pervasive settings such as the Internet, the domain of pervasive networks has its own characteristics that make these results non-applicable. Many of these applications are based on time-series data that possess the form of time-ordered series of events. Such applications also embody the need to handle large volumes of unexpected events, often modified on-the-fly, containing conflicting information, and dealing with rapidly changing contexts while producing results with low-latency. Correlating events across contextual dimensions holds the key to expanding the capabilities and improving the performance of these applications. This dissertation addresses this critical challenge. It establishes an effective scheme for complex-event semantic correlation. The scheme examines epistemic uncertainty in computer networks by fusing event synchronization concepts with belief theory. Because of the distributed nature of the event detection, time-delays are considered. Events are no longer instantaneous, but duration is associated with them. Existing algorithms for synchronizing time are split into two classes, one of which is asserted to provide a faster means for converging time and hence better suited for pervasive network management. Besides the temporal dimension, the scheme considers imprecision and uncertainty when an event is detected. A belief value is therefore associated with the semantics and the detection of composite events. This belief value is generated by a consensus among participating entities in a computer network. The scheme taps into in-network processing capabilities of pervasive computer networks and can withstand missing or conflicting information gathered from multiple participating entities. Thus, this dissertation advances knowledge in the field of network management by facilitating the full utilization of characteristics offered by pervasive, distributed and wireless technologies in contemporary and future computer networks.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Computer networks produce tremendous amounts of event-based data that can be collected and managed to support an increasing number of new classes of pervasive applications. Examples of such applications are network monitoring and crisis management. Although the problem of distributed event-based management has been addressed in the non-pervasive settings such as the Internet, the domain of pervasive networks has its own characteristics that make these results non-applicable. Many of these applications are based on time-series data that possess the form of time-ordered series of events. Such applications also embody the need to handle large volumes of unexpected events, often modified on-the-fly, containing conflicting information, and dealing with rapidly changing contexts while producing results with low-latency. Correlating events across contextual dimensions holds the key to expanding the capabilities and improving the performance of these applications. This dissertation addresses this critical challenge. It establishes an effective scheme for complex-event semantic correlation. The scheme examines epistemic uncertainty in computer networks by fusing event synchronization concepts with belief theory. Because of the distributed nature of the event detection, time-delays are considered. Events are no longer instantaneous, but duration is associated with them. Existing algorithms for synchronizing time are split into two classes, one of which is asserted to provide a faster means for converging time and hence better suited for pervasive network management. Besides the temporal dimension, the scheme considers imprecision and uncertainty when an event is detected. A belief value is therefore associated with the semantics and the detection of composite events. This belief value is generated by a consensus among participating entities in a computer network. The scheme taps into in-network processing capabilities of pervasive computer networks and can withstand missing or conflicting information gathered from multiple participating entities. Thus, this dissertation advances knowledge in the field of network management by facilitating the full utilization of characteristics offered by pervasive, distributed and wireless technologies in contemporary and future computer networks.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

As the universe of knowledge and subjects change over time, indexing languages like classification schemes, accommodate that change by restructuring. Restructuring indexing languages affects indexer and cataloguer work. Subjects may split or lump together. They may disappear only to reappear later. And new subjects may emerge that were assumed to be already present, but not clearly articulated (Miksa, 1998). In this context we have the complex relationship between the indexing language, the text being described, and the already described collection (Tennis, 2007). It is possible to imagine indexers placing a document into an outdated class, because it is the one they have already used for their collection. However, doing this erases the semantics in the present indexing language. Given this range of choice in the context of indexing language change, the question arises, what does this look like in practice? How often does this occur? Further, what does this phenomenon tell us about subjects in indexing languages? Does the practice we observe in the reaction to indexing language change provide us evidence of conceptual models of subjects and subject creation? If it is incomplete, but gets us close, what evidence do we still require?