942 resultados para Climatic data simulation
Resumo:
The triple- and quadruple-escape peaks of 6.128 MeV photons from the (19)F(p,alpha gamma)(16)O nuclear reaction were observed in an HPGe detector. The experimental peak areas, measured in spectra projected with a restriction function that allows quantitative comparison of data from different multiplicities, are in reasonably good agreement with those predicted by Monte Carlo simulations done with the general-purpose radiation-transport code PENELOPE. The behaviour of the escape intensities was simulated for some gamma-ray energies and detector dimensions; the results obtained can be extended to other energies using an empirical function and statistical properties related to the phenomenon. (C) 2010 Elsevier B.V. All rights reserved.
Resumo:
A time efficient optical model is proposed for GATE simulation of a LYSO scintillation matrix coupled to a photomultiplier. The purpose is to avoid the excessively long computation time when activating the optical processes in GATE. The usefulness of the model is demonstrated by comparing the simulated and experimental energy spectra obtained with the dual planar head equipment for dosimetry with a positron emission tomograph ( DoPET). The procedure to apply the model is divided in two steps. Firstly, a simplified simulation of a single crystal element of DoPET is used to fit an analytic function that models the optical attenuation inside the crystal. In a second step, the model is employed to calculate the influence of this attenuation in the energy registered by the tomograph. The use of the proposed optical model is around three orders of magnitude faster than a GATE simulation with optical processes enabled. A good agreement was found between the experimental and simulated data using the optical model. The results indicate that optical interactions inside the crystal elements play an important role on the energy resolution and induce a considerable degradation of the spectra information acquired by DoPET. Finally, the same approach employed by the proposed optical model could be useful to simulate a scintillation matrix coupled to a photomultiplier using single or dual readout scheme.
Resumo:
The issue of smoothing in kriging has been addressed either by estimation or simulation. The solution via estimation calls for postprocessing kriging estimates in order to correct the smoothing effect. Stochastic simulation provides equiprobable images presenting no smoothing and reproducing the covariance model. Consequently, these images reproduce both the sample histogram and the sample semivariogram. However, there is still a problem, which is the lack of local accuracy of simulated images. In this paper, a postprocessing algorithm for correcting the smoothing effect of ordinary kriging estimates is compared with sequential Gaussian simulation realizations. Based on samples drawn from exhaustive data sets, the postprocessing algorithm is shown to be superior to any individual simulation realization yet, at the expense of providing one deterministic estimate of the random function.
Resumo:
Predictors of random effects are usually based on the popular mixed effects (ME) model developed under the assumption that the sample is obtained from a conceptual infinite population; such predictors are employed even when the actual population is finite. Two alternatives that incorporate the finite nature of the population are obtained from the superpopulation model proposed by Scott and Smith (1969. Estimation in multi-stage surveys. J. Amer. Statist. Assoc. 64, 830-840) or from the finite population mixed model recently proposed by Stanek and Singer (2004. Predicting random effects from finite population clustered samples with response error. J. Amer. Statist. Assoc. 99, 1119-1130). Predictors derived under the latter model with the additional assumptions that all variance components are known and that within-cluster variances are equal have smaller mean squared error (MSE) than the competitors based on either the ME or Scott and Smith`s models. As population variances are rarely known, we propose method of moment estimators to obtain empirical predictors and conduct a simulation study to evaluate their performance. The results suggest that the finite population mixed model empirical predictor is more stable than its competitors since, in terms of MSE, it is either the best or the second best and when second best, its performance lies within acceptable limits. When both cluster and unit intra-class correlation coefficients are very high (e.g., 0.95 or more), the performance of the empirical predictors derived under the three models is similar. (c) 2007 Elsevier B.V. All rights reserved.
Resumo:
We review some issues related to the implications of different missing data mechanisms on statistical inference for contingency tables and consider simulation studies to compare the results obtained under such models to those where the units with missing data are disregarded. We confirm that although, in general, analyses under the correct missing at random and missing completely at random models are more efficient even for small sample sizes, there are exceptions where they may not improve the results obtained by ignoring the partially classified data. We show that under the missing not at random (MNAR) model, estimates on the boundary of the parameter space as well as lack of identifiability of the parameters of saturated models may be associated with undesirable asymptotic properties of maximum likelihood estimators and likelihood ratio tests; even in standard cases the bias of the estimators may be low only for very large samples. We also show that the probability of a boundary solution obtained under the correct MNAR model may be large even for large samples and that, consequently, we may not always conclude that a MNAR model is misspecified because the estimate is on the boundary of the parameter space.
Resumo:
Mathematical modeling has been extensively applied to the study and development of fuel cells. In this work, the objective is to characterize a mechanistic model for the anode of a direct ethanol fuel cell and perform appropriate simulations. The software Comsol Multiphysics (R) (and the Chemical Engineering Module) was used in this work. The software Comsol Multiphysics (R) is an interactive environment for modeling scientific and engineering applications using partial differential equations (PDEs). Based on the finite element method, it provides speed and accuracy for several applications. The mechanistic model developed here can supply details of the physical system, such as the concentration profiles of the components within the anode and the coverage of the adsorbed species on the electrode surface. Also, the anode overpotential-current relationship can be obtained. To validate the anode model presented in this paper, experimental data obtained with a single fuel cell operating with an ethanol solution at the anode were used. (C) 2008 Elsevier B.V. All rights reserved.
Resumo:
The aim of this work was to design a set of rules for levodopa infusion dose adjustment in Parkinson’s disease based on a simulation experiments. Using this simulator, optimal infusions dose in different conditions were calculated. There are seven conditions (-3 to +3)appearing in a rating scale for Parkinson’s disease patients. By finding mean of the differences between conditions and optimal dose, two sets of rules were designed. The set of rules was optimized by several testing. Usefulness for optimizing the titration procedure of new infusion patients based on rule-based reasoning was investigated. Results show that both of the number of the steps and the errors for finding optimal dose was shorten by new rules. At last, the dose predicted with new rules well on each single occasion of majority of patients in simulation experiments.
Resumo:
This thesis evaluates different sites for a weather measurement system and a suitable PV- simulation for University of Surabaya (UBAYA) in Indonesia/Java. The weather station is able to monitor all common weather phenomena including solar insolation. It is planned to use the data for scientific and educational purposes in the renewable energy studies. During evaluation and installation it falls into place that official specifications from global meteorological organizations could not be meet for some sensors caused by the conditions of UBAYA campus. After arranging the hardware the weather at the site was monitored for period of time. A comparison with different official sources from ground based and satellite bases measurements showed differences in wind and solar radiation. In some cases the monthly average solar insolation was deviating 42 % for satellite-based measurements. For the ground based it was less than 10 %. The average wind speed has a difference of 33 % compared to a source, which evaluated the wind power in Surabaya. The wind direction shows instabilities towards east compared with data from local weather station at the airport. PSET has the chance to get some investments to investigate photovoltaic on there own roof. With several simulations a suitable roof direction and the yearly and monthly outputs are shown. With a 7.7 kWpeak PV installation with the latest crystalline technology on the market 8.82 MWh/year could be achieved with weather data from 2012. Thin film technology could increase the value up to 9.13 MWh/year. However, the roofs have enough area to install PV. Finally the low price of electricity in Indonesia makes it not worth to feed in the energy into the public grid.
Resumo:
Generalized linear mixed models are flexible tools for modeling non-normal data and are useful for accommodating overdispersion in Poisson regression models with random effects. Their main difficulty resides in the parameter estimation because there is no analytic solution for the maximization of the marginal likelihood. Many methods have been proposed for this purpose and many of them are implemented in software packages. The purpose of this study is to compare the performance of three different statistical principles - marginal likelihood, extended likelihood, Bayesian analysis-via simulation studies. Real data on contact wrestling are used for illustration.
Resumo:
Solar plus heat pump systems are often very complex in design, with sometimes special heat pump arrangements and control. Therefore detailed heat pump models can give very slow system simulations and still not so accurate results compared to real heat pump performance in a system. The idea here is to start from a standard measured performance map of test points for a heat pump according to EN 14825 and then determine characteristic parameters for a simplified correlation based model of the heat pump. By plotting heat pump test data in different ways including power input and output form and not only as COP, a simplified relation could be seen. By using the same methodology as in the EN 12975 QDT part in the collector test standard it could be shown that a very simple model could describe the heat pump test data very accurately, by identifying 4 parameters in the correlation equation found. © 2012 The Authors.
Resumo:
Bin planning (arrangements) is a key factor in the timber industry. Improper planning of the storage bins may lead to inefficient transportation of resources, which threaten the overall efficiency and thereby limit the profit margins of sawmills. To address this challenge, a simulation model has been developed. However, as numerous alternatives are available for arranging bins, simulating all possibilities will take an enormous amount of time and it is computationally infeasible. A discrete-event simulation model incorporating meta-heuristic algorithms has therefore been investigated in this study. Preliminary investigations indicate that the results achieved by GA based simulation model are promising and better than the other meta-heuristic algorithm. Further, a sensitivity analysis has been done on the GA based optimal arrangement which contributes to gaining insights and knowledge about the real system that ultimately leads to improved and enhanced efficiency in sawmill yards. It is expected that the results achieved in the work will support timber industries in making optimal decisions with respect to arrangement of storage bins in a sawmill yard.
Numerical Simulation Of Sediment Transport And Bedmorphology Around A Hydraulic Structure On A River
Resumo:
Scour around hydraulic structures is a critical problem in hydraulic engineering. Under prediction of scour depth may lead to costly failures of the structure, while over prediction might result in unnecessary costs. Unfortunately, up-to-date empirical scour prediction formulas are based on laboratory experiments that are not always able to reproduce field conditions due to complicated geometry of rivers and temporal and spatial scales of a physical model. However, computational fluid dynamics (CFD) tools can perform using real field dimensions and operating conditions to predict sediment scour around hydraulic structures. In Korea, after completing the Four Major Rivers Restoration Project, several new weirs have been built across Han, Nakdong, Geum and Yeongsan Rivers. Consequently, sediment deposition and bed erosion around such structures have became a major issue in these four rivers. In this study, an application of an open source CFD software package, the TELEMAC-MASCARET, to simulate sediment transport and bed morphology around Gangjeong weir, which is the largest multipurpose weir built on Nakdong River. A real bathymetry of the river and a geometry of the weir have been implemented into the numerical model. The numerical simulation is carried out with a real hydrograph at the upstream boundary. The bedmorphology obtained from the numerical results has been validated against field observation data, and a maximum of simulated scour depth is compared with the results obtained by empirical formulas of Hoffmans. Agreement between numerical computations, observed data and empirical formulas is judged to be satisfactory on all major comparisons. The outcome of this study does not only point out the locations where deposition and erosion might take place depending on the weir gate operation, but also analyzes the mechanism of formation and evolution of scour holes after the weir gates.
Resumo:
The Enriquillo and Azuei are saltwater lakes located in a closed water basin in the southwestern region of the island of La Hispaniola, these have been experiencing dramatic changes in total lake-surface area coverage during the period 1980-2012. The size of Lake Enriquillo presented a surface area of approximately 276 km2 in 1984, gradually decreasing to 172 km2 in 1996. The surface area of the lake reached its lowest point in the satellite observation record in 2004, at 165 km2. Then the recent growth of the lake began reaching its 1984 size by 2006. Based on surface area measurement for June and July 2013, Lake Enriquillo has a surface area of ~358 km2. Sumatra sizes at both ends of the record are 116 km2 in 1984 and 134 km2in 2013, an overall 15.8% increase in 30 years. Determining the causes of lake surface area changes is of extreme importance due to its environmental, social, and economic impacts. The overall goal of this study is to quantify the changing water balance in these lakes and their catchment area using satellite and ground observations and a regional atmospheric-hydrologic modeling approach. Data analyses of environmental variables in the region reflect a hydrological unbalance of the lakes due to changing regional hydro-climatic conditions. Historical data show precipitation, land surface temperature and humidity, and sea surface temperature (SST), increasing over region during the past decades. Salinity levels have also been decreasing by more than 30% from previously reported baseline levels. Here we present a summary of the historical data obtained, new sensors deployed in the sourrounding sierras and the lakes, and the integrated modeling exercises. As well as the challenges of gathering, storing, sharing, and analyzing this large volumen of data in a remote location from such a diverse number of sources.
Resumo:
In a reccnt paper. Bai and Perron (1998) considcrccl theoretical issues relatec\ lo lhe limiting distriblltion of estimators and test. statist.ics in the linear model \\'ith multiplc struct ural changes. \Ve assess. via simulations, the adequacy of the \'arious I1Iethods suggested. These CO\'er the size and power of tests for structural changes. the cO\'erage rates of the confidence Íntervals for the break dates and the relat.Í\'e merits of methods to select the I1umber of breaks. The \'arious data generating processes considered alIo,,' for general conditions OIl the data and the errors including differellces across segmcll(s. Yarious practical recommendations are made.
Resumo:
We investigate the issue of whether there was a stable money demand function for Japan in 1990's using both aggregate and disaggregate time series data. The aggregate data appears to support the contention that there was no stable money demand function. The disaggregate data shows that there was a stable money demand function. Neither was there any indication of the presence of liquidity trapo Possible sources of discrepancy are explored and the diametrically opposite results between the aggregate and disaggregate analysis are attributed to the neglected heterogeneity among micro units. We also conduct simulation analysis to show that when heterogeneity among micro units is present. The prediction of aggregate outcomes, using aggregate data is less accurate than the prediction based on micro equations. Moreover. policy evaluation based on aggregate data can be grossly misleading.