986 resultados para Conditional simulation


Relevância:

100.00% 100.00%

Publicador:

Resumo:

This paper presents a new approach to the LU decomposition method for the simulation of stationary and ergodic random fields. The approach overcomes the size limitations of LU and is suitable for any size simulation. The proposed approach can facilitate fast updating of generated realizations with new data, when appropriate, without repeating the full simulation process. Based on a novel column partitioning of the L matrix, expressed in terms of successive conditional covariance matrices, the approach presented here demonstrates that LU simulation is equivalent to the successive solution of kriging residual estimates plus random terms. Consequently, it can be used for the LU decomposition of matrices of any size. The simulation approach is termed conditional simulation by successive residuals as at each step, a small set (group) of random variables is simulated with a LU decomposition of a matrix of updated conditional covariance of residuals. The simulated group is then used to estimate residuals without the need to solve large systems of equations.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Simulated-annealing-based conditional simulations provide a flexible means of quantitatively integrating diverse types of subsurface data. Although such techniques are being increasingly used in hydrocarbon reservoir characterization studies, their potential in environmental, engineering and hydrological investigations is still largely unexploited. Here, we introduce a novel simulated annealing (SA) algorithm geared towards the integration of high-resolution geophysical and hydrological data which, compared to more conventional approaches, provides significant advancements in the way that large-scale structural information in the geophysical data is accounted for. Model perturbations in the annealing procedure are made by drawing from a probability distribution for the target parameter conditioned to the geophysical data. This is the only place where geophysical information is utilized in our algorithm, which is in marked contrast to other approaches where model perturbations are made through the swapping of values in the simulation grid and agreement with soft data is enforced through a correlation coefficient constraint. Another major feature of our algorithm is the way in which available geostatistical information is utilized. Instead of constraining realizations to match a parametric target covariance model over a wide range of spatial lags, we constrain the realizations only at smaller lags where the available geophysical data cannot provide enough information. Thus we allow the larger-scale subsurface features resolved by the geophysical data to have much more due control on the output realizations. Further, since the only component of the SA objective function required in our approach is a covariance constraint at small lags, our method has improved convergence and computational efficiency over more traditional methods. Here, we present the results of applying our algorithm to the integration of porosity log and tomographic crosshole georadar data to generate stochastic realizations of the local-scale porosity structure. Our procedure is first tested on a synthetic data set, and then applied to data collected at the Boise Hydrogeophysical Research Site.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Gaussian random field (GRF) conditional simulation is a key ingredient in many spatial statistics problems for computing Monte-Carlo estimators and quantifying uncertainties on non-linear functionals of GRFs conditional on data. Conditional simulations are known to often be computer intensive, especially when appealing to matrix decomposition approaches with a large number of simulation points. This work studies settings where conditioning observations are assimilated batch sequentially, with one point or a batch of points at each stage. Assuming that conditional simulations have been performed at a previous stage, the goal is to take advantage of already available sample paths and by-products to produce updated conditional simulations at mini- mal cost. Explicit formulae are provided, which allow updating an ensemble of sample paths conditioned on n ≥ 0 observations to an ensemble conditioned on n + q observations, for arbitrary q ≥ 1. Compared to direct approaches, the proposed formulae proveto substantially reduce computational complexity. Moreover, these formulae explicitly exhibit how the q new observations are updating the old sample paths. Detailed complexity calculations highlighting the benefits of this approach with respect to state-of-the-art algorithms are provided and are complemented by numerical experiments.

Relevância:

70.00% 70.00%

Publicador:

Resumo:

Geophysical techniques can help to bridge the inherent gap with regard to spatial resolution and the range of coverage that plagues classical hydrological methods. This has lead to the emergence of the new and rapidly growing field of hydrogeophysics. Given the differing sensitivities of various geophysical techniques to hydrologically relevant parameters and their inherent trade-off between resolution and range the fundamental usefulness of multi-method hydrogeophysical surveys for reducing uncertainties in data analysis and interpretation is widely accepted. A major challenge arising from such endeavors is the quantitative integration of the resulting vast and diverse database in order to obtain a unified model of the probed subsurface region that is internally consistent with all available data. To address this problem, we have developed a strategy towards hydrogeophysical data integration based on Monte-Carlo-type conditional stochastic simulation that we consider to be particularly suitable for local-scale studies characterized by high-resolution and high-quality datasets. Monte-Carlo-based optimization techniques are flexible and versatile, allow for accounting for a wide variety of data and constraints of differing resolution and hardness and thus have the potential of providing, in a geostatistical sense, highly detailed and realistic models of the pertinent target parameter distributions. Compared to more conventional approaches of this kind, our approach provides significant advancements in the way that the larger-scale deterministic information resolved by the hydrogeophysical data can be accounted for, which represents an inherently problematic, and as of yet unresolved, aspect of Monte-Carlo-type conditional simulation techniques. We present the results of applying our algorithm to the integration of porosity log and tomographic crosshole georadar data to generate stochastic realizations of the local-scale porosity structure. Our procedure is first tested on pertinent synthetic data and then applied to corresponding field data collected at the Boise Hydrogeophysical Research Site near Boise, Idaho, USA.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Orebody modelling, support effects and the estimation of recoverable reserves are key parts of open pit optimization studies. A case study is presented on the estimation of recoverable reserves using an implementation of indicator kriging where metal quantity is used to select cutoffs, and support corrections founded on a conditional simulation approach. Mining selectivity is explored in the subsequent optimization study to compare results from indicator kriging of grade estimates on a regular size blocks and indicator kriging estimates on small size blocks. The use of indicator kriging models adjusted for a given selectivity and the use of grade proportions in each block for the optimization study, provide a presentation of the expected ore recovery for a predefined level of selectivity. The case study shows that indicator kriging estimation with full accounting of block grade distributions generates substantially better results in the pit optimization study. In addition, the adverse effects of small blocks and over-smoothing on optimization results are illustrated.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Abstract Accurate characterization of the spatial distribution of hydrological properties in heterogeneous aquifers at a range of scales is a key prerequisite for reliable modeling of subsurface contaminant transport, and is essential for designing effective and cost-efficient groundwater management and remediation strategies. To this end, high-resolution geophysical methods have shown significant potential to bridge a critical gap in subsurface resolution and coverage between traditional hydrological measurement techniques such as borehole log/core analyses and tracer or pumping tests. An important and still largely unresolved issue, however, is how to best quantitatively integrate geophysical data into a characterization study in order to estimate the spatial distribution of one or more pertinent hydrological parameters, thus improving hydrological predictions. Recognizing the importance of this issue, the aim of the research presented in this thesis was to first develop a strategy for the assimilation of several types of hydrogeophysical data having varying degrees of resolution, subsurface coverage, and sensitivity to the hydrologic parameter of interest. In this regard a novel simulated annealing (SA)-based conditional simulation approach was developed and then tested in its ability to generate realizations of porosity given crosshole ground-penetrating radar (GPR) and neutron porosity log data. This was done successfully for both synthetic and field data sets. A subsequent issue that needed to be addressed involved assessing the potential benefits and implications of the resulting porosity realizations in terms of groundwater flow and contaminant transport. This was investigated synthetically assuming first that the relationship between porosity and hydraulic conductivity was well-defined. Then, the relationship was itself investigated in the context of a calibration procedure using hypothetical tracer test data. Essentially, the relationship best predicting the observed tracer test measurements was determined given the geophysically derived porosity structure. Both of these investigations showed that the SA-based approach, in general, allows much more reliable hydrological predictions than other more elementary techniques considered. Further, the developed calibration procedure was seen to be very effective, even at the scale of tomographic resolution, for predictions of transport. This also held true at locations within the aquifer where only geophysical data were available. This is significant because the acquisition of hydrological tracer test measurements is clearly more complicated and expensive than the acquisition of geophysical measurements. Although the above methodologies were tested using porosity logs and GPR data, the findings are expected to remain valid for a large number of pertinent combinations of geophysical and borehole log data of comparable resolution and sensitivity to the hydrological target parameter. Moreover, the obtained results allow us to have confidence for future developments in integration methodologies for geophysical and hydrological data to improve the 3-D estimation of hydrological properties.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Recently, a lot of effort has been spent in the efficient computation of kriging predictors when observations are assimilated sequentially. In particular, kriging update formulae enabling significant computational savings were derived. Taking advantage of the previous kriging mean and variance computations helps avoiding a costly matrix inversion when adding one observation to the TeX already available ones. In addition to traditional update formulae taking into account a single new observation, Emery (2009) proposed formulae for the batch-sequential case, i.e. when TeX new observations are simultaneously assimilated. However, the kriging variance and covariance formulae given in Emery (2009) for the batch-sequential case are not correct. In this paper, we fix this issue and establish correct expressions for updated kriging variances and covariances when assimilating observations in parallel. An application in sequential conditional simulation finally shows that coupling update and residual substitution approaches may enable significant speed-ups.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Os empreendimentos de mineração comumente demandam grande quantidade de investimentos financeiros e, na maioria das vezes, longos períodos de implantação, o que os torna altamente sujeitos a diversas fontes de incertezas. Tais incertezas comumente tendem a diminuir conforme a evolução do projeto. O objetivo deste estudo é correlacionar as incertezas associadas ao modelo de teores de cobre do depósito Sequeirinho com o volume de investimentos realizados ao longo de distintas fases da pesquisa geológica. Este depósito insere-se no contexto do Complexo de Mineração Sossego, localizado no município de Canaã dos Carajás (PA). Primeiramente, foram realizadas 100 simulações para cada domínio litológico em cada campanha de sondagem (pré-1998, 1999, 2000, 2002 e 2003) a partir do método de simulação sequencial gaussiana condicionada aos dados amostrais, totalizando 1.400 possíveis cenários. Para a avaliação das incertezas foram calculados três índices: variância condicional, coeficiente de variação condicional e intervalo interquartil. Por fim, a avaliação dos investimentos foi elaborada a partir dos custos estimados para o desenvolvimento de sondagens e análises químicas. Desde a campanha pré-1998, houve uma tendência de os teores médios do depósito aproximarem-se dos prováveis valores reais observados nas fases finais da pesquisa. No ano de 2000 ocorreu o maior investimento (cerca de 28 milhões de Reais) e a redução das incertezas atingiu o patamar de 15%. Os investimentos desenvolvidos em sondagens posteriores à campanha de 2000 foram da ordem de 9 milhões de Reais (cerca de 12 mil metros de sondagem), porém, não foram constatadas reduções significativas das incertezas. Este investimento seria melhor aproveitado caso fosse redirecionado a novas áreas de prospecção. Além do montante financeiro necessário para a redução das incertezas, foco deste estudo, as variações na interpretação geológica e a locação dos furos de sondagem são variáveis importantes na análise de incertezas associadas aos investimentos em pesquisa geológica.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

A condutividade hidráulica (K) é um dos parâmetros controladores da magnitude da velocidade da água subterrânea, e consequentemente, é um dos mais importantes parâmetros que afetam o fluxo subterrâneo e o transporte de solutos, sendo de suma importância o conhecimento da distribuição de K. Esse trabalho visa estimar valores de condutividade hidráulica em duas áreas distintas, uma no Sistema Aquífero Guarani (SAG) e outra no Sistema Aquífero Bauru (SAB) por meio de três técnicas geoestatísticas: krigagem ordinária, cokrigagem e simulação condicional por bandas rotativas. Para aumentar a base de dados de valores de K, há um tratamento estatístico dos dados conhecidos. O método de interpolação matemática (krigagem ordinária) e o estocástico (simulação condicional por bandas rotativas) são aplicados para estimar os valores de K diretamente, enquanto que os métodos de krigagem ordinária combinada com regressão linear e cokrigagem permitem incorporar valores de capacidade específica (Q/s) como variável secundária. Adicionalmente, a cada método geoestatístico foi aplicada a técnica de desagrupamento por célula para comparar a sua capacidade de melhorar a performance dos métodos, o que pode ser avaliado por meio da validação cruzada. Os resultados dessas abordagens geoestatísticas indicam que os métodos de simulação condicional por bandas rotativas com a técnica de desagrupamento e de krigagem ordinária combinada com regressão linear sem a técnica de desagrupamento são os mais adequados para as áreas do SAG (rho=0.55) e do SAB (rho=0.44), respectivamente. O tratamento estatístico e a técnica de desagrupamento usados nesse trabalho revelaram-se úteis ferramentas auxiliares para os métodos geoestatísticos.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Traditionally, geostatistical algorithms are contained within specialist GIS and spatial statistics software. Such packages are often expensive, with relatively complex user interfaces and steep learning curves, and cannot be easily integrated into more complex process chains. In contrast, Service Oriented Architectures (SOAs) promote interoperability and loose coupling within distributed systems, typically using XML (eXtensible Markup Language) and Web services. Web services provide a mechanism for a user to discover and consume a particular process, often as part of a larger process chain, with minimal knowledge of how it works. Wrapping current geostatistical algorithms with a Web service layer would thus increase their accessibility, but raises several complex issues. This paper discusses a solution to providing interoperable, automatic geostatistical processing through the use of Web services, developed in the INTAMAP project (INTeroperability and Automated MAPping). The project builds upon Open Geospatial Consortium standards for describing observations, typically used within sensor webs, and employs Geography Markup Language (GML) to describe the spatial aspect of the problem domain. Thus the interpolation service is extremely flexible, being able to support a range of observation types, and can cope with issues such as change of support and differing error characteristics of sensors (by utilising descriptions of the observation process provided by SensorML). XML is accepted as the de facto standard for describing Web services, due to its expressive capabilities which allow automatic discovery and consumption by ‘naive’ users. Any XML schema employed must therefore be capable of describing every aspect of a service and its processes. However, no schema currently exists that can define the complex uncertainties and modelling choices that are often present within geostatistical analysis. We show a solution to this problem, developing a family of XML schemata to enable the description of a full range of uncertainty types. These types will range from simple statistics, such as the kriging mean and variances, through to a range of probability distributions and non-parametric models, such as realisations from a conditional simulation. By employing these schemata within a Web Processing Service (WPS) we show a prototype moving towards a truly interoperable geostatistical software architecture.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Apresenta-se uma metodologia para caracterizar a transmissividade dos Granitos Hercínicos e Metasedimentos do Complexo Xisto-Grauváquico do maciço envolvente e subjacente à antiga área mineira de urânio da Quinta do Bispo. Inicia-se com a modelação das litologias e grau de alteração a que se segue a simulação condicional da densidade de fracturação. No final, a densidade de fracturação é convertida num modelo 3D de transmissividade por relação com os resultados dos ensaios de bombagem. The purpose of this work is to present a methodology for characterizing the transmissivity of the Hercynian granites and complex schist–greywacke metasediment rocks surrounding and underlying the old Quinta do Bispo uranium mining site. The methodology encompasses modelling of lithologies and weathering levels, followed by a conditional simulation of fracture density. Fracture density is then converted into a 3D model of transmissivity via a relationship with pumping tests.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Abstract: Asthma prevalence in children and adolescents in Spain is 10-17%. It is the most common chronic illness during childhood. Prevalence has been increasing over the last 40 years and there is considerable evidence that, among other factors, continued exposure to cigarette smoke results in asthma in children. No statistical or simulation model exist to forecast the evolution of childhood asthma in Europe. Such a model needs to incorporate the main risk factors that can be managed by medical authorities, such as tobacco (OR = 1.44), to establish how they affect the present generation of children. A simulation model using conditional probability and discrete event simulation for childhood asthma was developed and validated by simulating realistic scenario. The parameters used for the model (input data) were those found in the bibliography, especially those related to the incidence of smoking in Spain. We also used data from a panel of experts from the Hospital del Mar (Barcelona) related to actual evolution and asthma phenotypes. The results obtained from the simulation established a threshold of a 15-20% smoking population for a reduction in the prevalence of asthma. This is still far from the current level in Spain, where 24% of people smoke. We conclude that more effort must be made to combat smoking and other childhood asthma risk factors, in order to significantly reduce the number of cases. Once completed, this simulation methodology can realistically be used to forecast the evolution of childhood asthma as a function of variation in different risk factors.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Control of a chaotic system by homogeneous nonlinear driving, when a conditional Lyapunov exponent is zero, may give rise to special and interesting synchronizationlike behaviors in which the response evolves in perfect correlation with the drive. Among them, there are the amplification of the drive attractor and the shift of it to a different region of phase space. In this paper, these synchronizationlike behaviors are discussed, and demonstrated by computer simulation of the Lorentz model [E. N. Lorenz, J. Atmos. Sci. 20 130 (1963)] and the double scroll [T. Matsumoto, L. O. Chua, and M. Komuro, IEEE Trans. CAS CAS-32, 798 (1985)].