95 resultados para Discretisation,


Relevância:

10.00% 10.00%

Publicador:

Resumo:

The Boundary Element Method (BEM) is a discretisation technique for solving partial differential equations, which offers, for certain problems, important advantages over domain techniques. Despite the high CPU time reduction that can be achieved, some 3D problems remain today untreatable because the extremely large number of degrees of freedom—dof—involved in the boundary description. Model reduction seems to be an appealing choice for both, accurate and efficient numerical simulations. However, in the BEM the reduction in the number of degrees of freedom does not imply a significant reduction in the CPU time, because in this technique the more important part of the computing time is spent in the construction of the discrete system of equations. In this way, a reduction also in the number of weighting functions, seems to be a key point to render efficient boundary element simulations.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

No último século, houve grande avanço no entendimento das interações das radiações com a matéria. Essa compreensão se faz necessária para diversas aplicações, entre elas o uso de raios X no diagnóstico por imagens. Neste caso, imagens são formadas pelo contraste resultante da diferença na atenuação dos raios X pelos diferentes tecidos do corpo. Entretanto, algumas das interações dos raios X com a matéria podem levar à redução da qualidade destas imagens, como é o caso dos fenômenos de espalhamento. Muitas abordagens foram propostas para estimar a distribuição espectral de fótons espalhados por uma barreira, ou seja, como no caso de um feixe de campo largo, ao atingir um plano detector, tais como modelos que utilizam métodos de Monte Carlo e modelos que utilizam aproximações analíticas. Supondo-se um espectro de um feixe primário que não interage com nenhum objeto após sua emissão pelo tubo de raios X, este espectro é, essencialmente representado pelos modelos propostos anteriormente. Contudo, considerando-se um feixe largo de radiação X, interagindo com um objeto, a radiação a ser detectada por um espectrômetro, passa a ser composta pelo feixe primário, atenuado pelo material adicionado, e uma fração de radiação espalhada. A soma destas duas contribuições passa a compor o feixe resultante. Esta soma do feixe primário atenuado, com o feixe de radiação espalhada, é o que se mede em um detector real na condição de feixe largo. O modelo proposto neste trabalho visa calcular o espectro de um tubo de raios X, em situação de feixe largo, o mais fidedigno possível ao que se medem em condições reais. Neste trabalho se propõe a discretização do volume de interação em pequenos elementos de volume, nos quais se calcula o espalhamento Compton, fazendo uso de um espectro de fótons gerado pelo Modelo de TBC, a equação de Klein-Nishina e considerações geométricas. Por fim, o espectro de fótons espalhados em cada elemento de volume é somado ao espalhamento dos demais elementos de volume, resultando no espectro total espalhado. O modelo proposto foi implementado em ambiente computacional MATLAB® e comparado com medições experimentais para sua validação. O modelo proposto foi capaz de produzir espectros espalhados em diferentes condições, apresentando boa conformidade com os valores medidos, tanto em termos quantitativos, nas quais a diferença entre kerma no ar calculado e kerma no ar medido é menor que 10%, quanto qualitativos, com fatores de mérito superiores a 90%.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Three sets of laboratory column experimental results concerning the hydrogeochemistry of seawater intrusion have been modelled using two codes: ACUAINTRUSION (Chemical Engineering Department, University of Alicante) and PHREEQC (U.S.G.S.). These reactive models utilise the hydrodynamic parameters determined using the ACUAINTRUSION TRANSPORT software and fit the chloride breakthrough curves perfectly. The ACUAINTRUSION code was improved, and the instabilities were studied relative to the discretisation. The relative square errors were obtained using different combinations of the spatial and temporal steps: the global error for the total experimental data and the partial error for each element. Good simulations for the three experiments were obtained using the ACUAINTRUSION software with slight variations in the selectivity coefficients for both sediments determined in batch experiments with fresh water. The cation exchange parameters included in ACUAINTRUSION are those reported by the Gapon convention with modified exponents for the Ca/Mg exchange. PHREEQC simulations performed using the Gains-Thomas convention were unsatisfactory, with the exchange coefficients from the database of PHREEQC (or range), but those determined with fresh water – natural sediment allowed only an approximation to be obtained. For the treated sediment, the adjusted exchange coefficients were determined to improve the simulation and are vastly different from those from the database of PHREEQC or batch experiment values; however, these values fall in an order similar to the others determined under dynamic conditions. Different cation concentrations were simulated using two different software packages; this disparity could be attributed to the defined selectivity coefficients that affect the gypsum equilibrium. Consequently, different calculated sulphate concentrations are obtained using each type of software; a smaller mismatch was predicted using ACUAINTRUSION. In general, the presented simulations by ACUAINTRUSION and PHREEQC produced similar results, making predictions consistent with the experimental data. However, the simulated results are not identical to the experimental data; sulphate (total S) is overpredicted by both models, most likely due to such factors as the kinetics of gypsum, the possible variations in the exchange coefficients due to salinity and the neglect of other processes.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

This research work analyses techniques for implementing a cell-centred finite-volume time-domain (ccFV-TD) computational methodology for the purpose of studying microwave heating. Various state-of-the-art spatial and temporal discretisation methods employed to solve Maxwell's equations on multidimensional structured grid networks are investigated, and the dispersive and dissipative errors inherent in those techniques examined. Both staggered and unstaggered grid approaches are considered. Upwind schemes using a Riemann solver and intensity vector splitting are studied and evaluated. Staggered and unstaggered Leapfrog and Runge-Kutta time integration methods are analysed in terms of phase and amplitude error to identify which method is the most accurate and efficient for simulating microwave heating processes. The implementation and migration of typical electromagnetic boundary conditions. from staggered in space to cell-centred approaches also is deliberated. In particular, an existing perfectly matched layer absorbing boundary methodology is adapted to formulate a new cell-centred boundary implementation for the ccFV-TD solvers. Finally for microwave heating purposes, a comparison of analytical and numerical results for standard case studies in rectangular waveguides allows the accuracy of the developed methods to be assessed. © 2004 Elsevier Inc. All rights reserved.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The planning and management of water resources in the Pioneer Valley, north-eastern Australia requires a tool for assessing the impact of groundwater and stream abstractions on water supply reliabilities and environmental flows in Sandy Creek (the main surface water system studied). Consequently, a fully coupled stream-aquifer model has been constructed using the code MODHMS, calibrated to near-stream observations of watertable behaviour and multiple components of gauged stream flow. This model has been tested using other methods of estimation, including stream depletion analysis and radon isotope tracer sampling. The coarseness of spatial discretisation, which is required for practical reasons of computational efficiency, limits the model's capacity to simulate small-scale processes (e.g., near-stream groundwater pumping, bank storage effects), and alternative approaches are required to complement the model's range of applicability. Model predictions of groundwater influx to Sandy Creek are compared with baseflow estimates from three different hydrograph separation techniques, which were found to be unable to reflect the dynamics of Sandy Creek stream-aquifer interactions. The model was also used to infer changes in the water balance of the system caused by historical land use change. This led to constraints on the recharge distribution which can be implemented to improve model calibration performance. (c) 2006 Elsevier B.V. All rights reserved.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Plastic yield criteria for porous ductile materials are explored numerically using the finite-element technique. The cases of spherical voids arranged in simple cubic, body-centred cubic and face-centred cubic arrays are investigated with void volume fractions ranging from 2 % through to the percolation limit (over 90 %). Arbitrary triaxial macroscopic stress states and two definitions of yield are explored. The numerical data demonstrates that the yield criteria depend linearly on the determinant of the macroscopic stress tensor for the case of simple-cubic and body-centred cubic arrays - in contrast to the famous Gurson-Tvergaard-Needleman (GTN) formula - while there is no such dependence for face-centred cubic arrays within the accuracy of the finite-element discretisation. The data are well fit by a simple extension of the GTN formula which is valid for all void volume fractions, with yield-function convexity constraining the form of the extension in terms of parameters in the original formula. Simple cubic structures are more resistant to shear, while body-centred and face-centred structures are more resistant to hydrostatic pressure. The two yield surfaces corresponding to the two definitions of yield are not related by a simple scaling.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Sea-water intrusion is actively contaminating fresh groundwater reserves in the coastal aquifers of the Pioneer Valley,north-eastern Australia. A three-dimensional sea-water intrusion model has been developed using the MODHMS code to explore regional-scale processes and to aid assessment of management strategies for the system. A sea-water intrusion potential map, produced through analyses of the hydrochemistry, hydrology and hydrogeology, offsets model limitations by providing an alternative appraisal of susceptibility. Sea-water intrusion in the Pioneer Valley is not in equilibrium, and a potential exists for further landward shifts in the extent of saline groundwater. The model required consideration of tidal over-height (the additional hydraulic head at the coast produced by the action of tides), with over-height values in the range 0.5-0.9 m giving improved water-table predictions. The effect of the initial water-table condition dominated the sensitivity of the model to changes in the coastal hydraulic boundary condition. Several salination processes are probably occurring in the Pioneer Valley, rather than just simple landward sea-water advancement from modern sources of marine salts. The method of vertical discretisation (i.e. model-layer subdivision) was shown to introduce some errors in the prediction of watertable behaviour.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The ability to grow microscopic spherical birefringent crystals of vaterite, a calcium carbonate mineral, has allowed the development of an optical microrheometer based on optical tweezers. However, since these crystals are birefringent, and worse, are expected to have non-uniform birefringence, computational modeling of the microrheometer is a highly challenging task. Modeling the microrheometer - and optical tweezers in general - typically requires large numbers of repeated calculations for the same trapped particle. This places strong demands on the efficiency of computational methods used. While our usual method of choice for computational modelling of optical tweezers - the T-matrix method - meets this requirement of efficiency, it is restricted to homogeneous isotropic particles. General methods that can model complex structures such as the vaterite particles, such as finite-difference time-domain (FDTD) or finite-difference frequency-domain (FDFD) methods, are inefficient. Therefore, we have developed a hybrid FDFD/T-matrix method that combines the generality of volume-discretisation methods such as FDFD with the efficiency of the T-matrix method. We have used this hybrid method to calculate optical forces and torques on model vaterite spheres in optical traps. We present and compare the results of computational modelling and experimental measurements.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

This paper introduces a new technique in the investigation of limited-dependent variable models. This paper illustrates that variable precision rough set theory (VPRS), allied with the use of a modern method of classification, or discretisation of data, can out-perform the more standard approaches that are employed in economics, such as a probit model. These approaches and certain inductive decision tree methods are compared (through a Monte Carlo simulation approach) in the analysis of the decisions reached by the UK Monopolies and Mergers Committee. We show that, particularly in small samples, the VPRS model can improve on more traditional models, both in-sample, and particularly in out-of-sample prediction. A similar improvement in out-of-sample prediction over the decision tree methods is also shown.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

In this paper, we present a framework for Bayesian inference in continuous-time diffusion processes. The new method is directly related to the recently proposed variational Gaussian Process approximation (VGPA) approach to Bayesian smoothing of partially observed diffusions. By adopting a basis function expansion (BF-VGPA), both the time-dependent control parameters of the approximate GP process and its moment equations are projected onto a lower-dimensional subspace. This allows us both to reduce the computational complexity and to eliminate the time discretisation used in the previous algorithm. The new algorithm is tested on an Ornstein-Uhlenbeck process. Our preliminary results show that BF-VGPA algorithm provides a reasonably accurate state estimation using a small number of basis functions.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

A mathematical model is developed for the general pneumatic tyre. The model will permit the investigations of tyre deformations produced by arbitrary external loading, and will enable estimates to be made of the distributions of applied and reactive forces. The principle of Finite Elements is used to idealise the composite tyre structure, each element consisting of a triangle of double curvature with varying thickness. Large deflections of' the structure are accomodated by the use of an iterative sequence of small incremental steps, each of' which obeys the laws of linear mechanics. The theoretical results are found to compare favourably with the experimental test data obtained from two different types of ttye construction. However, limitations in the discretisation process has prohibited accurate assessments to be made of stress distributions in the regions of high stress gradients ..

Relevância:

10.00% 10.00%

Publicador:

Resumo:

2000 Mathematics Subject Classification: 26A33 (primary), 35S15 (secondary)

Relevância:

10.00% 10.00%

Publicador:

Resumo:

2000 Mathematics Subject Classification: 26A33 (primary), 35S15

Relevância:

10.00% 10.00%

Publicador:

Resumo:

A numerical method for the Dirichlet initial boundary value problem for the heat equation in the exterior and unbounded region of a smooth closed simply connected 3-dimensional domain is proposed and investigated. This method is based on a combination of a Laguerre transformation with respect to the time variable and an integral equation approach in the spatial variables. Using the Laguerre transformation in time reduces the parabolic problem to a sequence of stationary elliptic problems which are solved by a boundary layer approach giving a sequence of boundary integral equations of the first kind to solve. Under the assumption that the boundary surface of the solution domain has a one-to-one mapping onto the unit sphere, these integral equations are transformed and rewritten over this sphere. The numerical discretisation and solution are obtained by a discrete projection method involving spherical harmonic functions. Numerical results are included.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The Model for Prediction Across Scales (MPAS) is a novel set of Earth system simulation components and consists of an atmospheric model, an ocean model and a land-ice model. Its distinct features are the use of unstructured Voronoi meshes and C-grid discretisation to address shortcomings of global models on regular grids and the use of limited area models nested in a forcing data set, with respect to parallel scalability, numerical accuracy and physical consistency. This concept allows one to include the feedback of regional land use information on weather and climate at local and global scales in a consistent way, which is impossible to achieve with traditional limited area modelling approaches. Here, we present an in-depth evaluation of MPAS with regards to technical aspects of performing model runs and scalability for three medium-size meshes on four different high-performance computing (HPC) sites with different architectures and compilers. We uncover model limitations and identify new aspects for the model optimisation that are introduced by the use of unstructured Voronoi meshes. We further demonstrate the model performance of MPAS in terms of its capability to reproduce the dynamics of the West African monsoon (WAM) and its associated precipitation in a pilot study. Constrained by available computational resources, we compare 11-month runs for two meshes with observations and a reference simulation from the Weather Research and Forecasting (WRF) model. We show that MPAS can reproduce the atmospheric dynamics on global and local scales in this experiment, but identify a precipitation excess for the West African region. Finally, we conduct extreme scaling tests on a global 3?km mesh with more than 65 million horizontal grid cells on up to half a million cores. We discuss necessary modifications of the model code to improve its parallel performance in general and specific to the HPC environment. We confirm good scaling (70?% parallel efficiency or better) of the MPAS model and provide numbers on the computational requirements for experiments with the 3?km mesh. In doing so, we show that global, convection-resolving atmospheric simulations with MPAS are within reach of current and next generations of high-end computing facilities.