971 resultados para Monte-Carlo simulation
Resumo:
The dependency of the blood oxygenation level dependent (BOLD) signal on underlying hemodynamics is not well understood. Building a forward biophysical model of this relationship is important for the quantitative estimation of the hemodynamic changes and neural activity underlying functional magnetic resonance imaging (fMRI) signals. We have developed a general model of the BOLD signal which can model both intra- and extravascular signals for an arbitrary tissue model across a wide range of imaging parameters. The model of the BOLD signal was instantiated as a look-up-table (LuT), and was verified against concurrent fMRI and optical imaging measurements of activation induced hemodynamics. Magn Reson Med, 2008. © 2008 Wiley-Liss, Inc.
Resumo:
Monte Carlo field-theoretic simulations (MCFTS) are performed on melts of symmetric diblock copolymer for invariant polymerization indexes extending down to experimentally relevant values of N̅ ∼ 10^4. The simulations are performed with a fluctuating composition field, W_−(r), and a pressure field, W_+(r), that follows the saddle-point approximation. Our study focuses on the disordered-state structure function, S(k), and the order−disorder transition (ODT). Although shortwavelength fluctuations cause an ultraviolet (UV) divergence in three dimensions, this is readily compensated for with the use of an effective Flory−Huggins interaction parameter, χ_e. The resulting S(k) matches the predictions of renormalized one-loop (ROL) calculations over the full range of χ_eN and N̅ examined in our study, and agrees well with Fredrickson−Helfand (F−H) theory near the ODT. Consistent with the F−H theory, the ODT is discontinuous for finite N̅ and the shift in (χ_eN)_ODT follows the predicted N̅^−1/3 scaling over our range of N̅.
Resumo:
This paper employs an extensive Monte Carlo study to test the size and power of the BDS and close return methods of testing for departures from independent and identical distribution. It is found that the finite sample properties of the BDS test are far superior and that the close return method cannot be recommended as a model diagnostic. Neither test can be reliably used for very small samples, while the close return test has low power even at large sample sizes
Resumo:
Monte Carlo algorithms often aim to draw from a distribution π by simulating a Markov chain with transition kernel P such that π is invariant under P. However, there are many situations for which it is impractical or impossible to draw from the transition kernel P. For instance, this is the case with massive datasets, where is it prohibitively expensive to calculate the likelihood and is also the case for intractable likelihood models arising from, for example, Gibbs random fields, such as those found in spatial statistics and network analysis. A natural approach in these cases is to replace P by an approximation Pˆ. Using theory from the stability of Markov chains we explore a variety of situations where it is possible to quantify how ’close’ the chain given by the transition kernel Pˆ is to the chain given by P . We apply these results to several examples from spatial statistics and network analysis.
Resumo:
The Monte Carlo Independent Column Approximation (McICA) is a flexible method for representing subgrid-scale cloud inhomogeneity in radiative transfer schemes. It does, however, introduce conditional random errors but these have been shown to have little effect on climate simulations, where spatial and temporal scales of interest are large enough for effects of noise to be averaged out. This article considers the effect of McICA noise on a numerical weather prediction (NWP) model, where the time and spatial scales of interest are much closer to those at which the errors manifest themselves; this, as we show, means that noise is more significant. We suggest methods for efficiently reducing the magnitude of McICA noise and test these methods in a global NWP version of the UK Met Office Unified Model (MetUM). The resultant errors are put into context by comparison with errors due to the widely used assumption of maximum-random-overlap of plane-parallel homogeneous cloud. For a simple implementation of the McICA scheme, forecasts of near-surface temperature are found to be worse than those obtained using the plane-parallel, maximum-random-overlap representation of clouds. However, by applying the methods suggested in this article, we can reduce noise enough to give forecasts of near-surface temperature that are an improvement on the plane-parallel maximum-random-overlap forecasts. We conclude that the McICA scheme can be used to improve the representation of clouds in NWP models, with the provision that the associated noise is sufficiently small.
Resumo:
We discuss the basic hydrodynamics that determines the density structure of the disks around hot stars. Observational evidence supports the idea that these disks are Keplerian (rotationally supported) gaseous disks. A popular scenario in the literature, which naturally leads to the formation of Keplerian disks, is the viscous decretion model. According to this scenario, the disks are hydrostatically supported in the vertical direction, while the radial structure is governed by the viscous transport. This suggests that the temperature is one primary factor that governs the disk density structure. In a previous study we demonstrated, using three-dimensional non-LTE Monte Carlo simulations, that viscous Keplerian disks can be highly nonisothermal. In this paper we build on our previous work and solve the full problem of the steady state nonisothermal viscous diffusion and vertical hydrostatic equilibrium. We find that the self-consistent solution departs significantly from the analytic isothermal density, with potentially large effects on the emergent spectrum. This implies that nonisothermal disk models must be used for a detailed modeling of Be star disks.
Resumo:
We describe the canonical and microcanonical Monte Carlo algorithms for different systems that can be described by spin models. Sites of the lattice, chosen at random, interchange their spin values, provided they are different. The canonical ensemble is generated by performing exchanges according to the Metropolis prescription whereas in the microcanonical ensemble, exchanges are performed as long as the total energy remains constant. A systematic finite size analysis of intensive quantities and a comparison with results obtained from distinct ensembles are performed and the quality of results reveal that the present approach may be an useful tool for the study of phase transitions, specially first-order transitions. (C) 2009 Elsevier B.V. All rights reserved.
Resumo:
Neste trabalho, analisamos utilização da metodologia CreditRLsk+ do Credit Suisse sua adequação ao mercado brasileiro, com objetivo de calcular risco de uma carteira de crédito. Certas hipóteses assumidas na formulação do modelo CreditRisk+ não valem para o mercado brasileiro, caracterizado, por exemplo, por uma elevada probabilidade de defcnilt. Desenvolvemos, então, uma metodologia para cálculo da distribuição de perdas através do método de Simulação de Monte Cario, alterando algumas hipóteses originais do modelo com objetivo de adaptá-lo ao nosso mercado. utilização de simulações também oferece resultados mais precisos em situações onde as carteiras possuem uma pequena população de contratos, além de eliminar possíveis problemas de convergência do método analítico, mesmo considerando as hipóteses do modelo original. Verifica-se ainda que tempo computacional pode ser menor que da metodologia original, principalmente em carteiras com elevado número de devedores de perfis distintos com alocações em diversos setores da economia. Tendo em vista as restrições acima, acreditamos que metodologia proposta seja uma alternativa para forma analítica do modelo CreditRisk+. Apresentamos exemplos de utilização resultados providos por estas simulações. ponto central deste trabalho realçar importância da utilização de metodologias alternativas de medição de risco de crédito que incorporem as particularidades do mercado brasileiro.
Resumo:
O presente trabalho tem por objetivo descrever, avaliar comparar as metodologias analítica da simulação Monte Cario para cálculo do Value at Risk (Valor em Risco) de instituições financeiras de empresas. Para comparar as vantagens desvantagens de cada metodologia, efetuaremos comparações algébricas realizamos diversos testes empíricos com instituições hipotéticas que apresentassem diferentes níveis de alavancagem de composição em seus balanços, que operassem em diferentes mercados (consideramos os mercados de ações, de opções de compra de títulos de renda fixa prefixados).
Resumo:
A atividade de crédito consiste, em termos financeiros, na entrega de um valor presente mediante uma promessa de pagamento em data futura. Uma vez que se trata de promessa de pagamento, existe a possibilidade de que a mesma não seja cumprida, dando origem ao risco de crédito. Dada a existência desse risco, cabe às instituições que concedem crédito realizar uma análise do proponente, buscando aferir se o mesmo possui condições de cumprir o compromisso assumido. Um dos aspectos a ser levado em consideração nessa análise diz respeito à capacidade de pagamento da empresa, realizada através de projeções do fluxo de caixa da mesma. Com base nisso, o presente trabalho teve por objetivo a construção de um modelo de simulação para avaliar a capacidade de pagamento de empresas em financiamentos de longo prazo. Entende-se que a aplicação de um modelo dessa natureza pode trazer informações adicionais relevantes para o problema em análise. A avaliação do modelo construído, realizada através de consulta a experts no problema de decisão de crédito, aponta que tanto a técnica utilizada quanto o modelo construído mostram-se adequados e aderentes à realidade, propiciando uma maior segurança para a tomada de decisão.
Resumo:
Trata da aplicabilidade da Simulação de Monte Carlo para a análise de riscos e, conseqüentemente, o apoio à decisão de investir ou não em um projeto. São abordados métodos de análise de riscos e seleção de projetos, bem como a natureza, vantagens e limitações da Simulação de Monte Carío. Por fim este instrumento tem sua viabilidade analisada sob a luz do processo de análise de riscos de uma empresa brasileira.
Resumo:
Despite the commonly held belief that aggregate data display short-run comovement, there has been little discussion about the econometric consequences of this feature of the data. We use exhaustive Monte-Carlo simulations to investigate the importance of restrictions implied by common-cyclical features for estimates and forecasts based on vector autoregressive models. First, we show that the ìbestî empirical model developed without common cycle restrictions need not nest the ìbestî model developed with those restrictions. This is due to possible differences in the lag-lengths chosen by model selection criteria for the two alternative models. Second, we show that the costs of ignoring common cyclical features in vector autoregressive modelling can be high, both in terms of forecast accuracy and efficient estimation of variance decomposition coefficients. Third, we find that the Hannan-Quinn criterion performs best among model selection criteria in simultaneously selecting the lag-length and rank of vector autoregressions.
Resumo:
Despite the belief, supported byrecentapplied research, thataggregate datadisplay short-run comovement, there has been little discussion about the econometric consequences ofthese data “features.” W e use exhaustive M onte-Carlo simulations toinvestigate theimportance ofrestrictions implied by common-cyclicalfeatures for estimates and forecasts based on vectorautoregressive and errorcorrection models. First, weshowthatthe“best” empiricalmodeldevelopedwithoutcommoncycles restrictions neednotnestthe“best” modeldevelopedwiththoserestrictions, duetothe use ofinformation criteria forchoosingthe lagorderofthe twoalternative models. Second, weshowthatthecosts ofignoringcommon-cyclicalfeatures inV A R analysis may be high in terms offorecastingaccuracy and e¢ciency ofestimates ofvariance decomposition coe¢cients. A lthough these costs are more pronounced when the lag orderofV A R modelsareknown, theyarealsonon-trivialwhenitis selectedusingthe conventionaltoolsavailabletoappliedresearchers. T hird, we…ndthatifthedatahave common-cyclicalfeatures andtheresearcherwants touseaninformationcriterium to selectthelaglength, theH annan-Q uinn criterium is themostappropriate, sincethe A kaike and theSchwarz criteriahave atendency toover- and under-predictthe lag lengthrespectivelyinoursimulations.
Resumo:
Neste trabalho investigamos as propriedades em pequena amostra e a robustez das estimativas dos parâmetros de modelos DSGE. Tomamos o modelo de Smets and Wouters (2007) como base e avaliamos a performance de dois procedimentos de estimação: Método dos Momentos Simulados (MMS) e Máxima Verossimilhança (MV). Examinamos a distribuição empírica das estimativas dos parâmetros e sua implicação para as análises de impulso-resposta e decomposição de variância nos casos de especificação correta e má especificação. Nossos resultados apontam para um desempenho ruim de MMS e alguns padrões de viés nas análises de impulso-resposta e decomposição de variância com estimativas de MV nos casos de má especificação considerados.