979 resultados para kirkwood superposition approximation
Resumo:
In this paper we address the importance of distributive effects in the social valuation of QALY's. We propose a social welfarefunction that generalises the functions traditionally used in the health economic literature. The novelty is that, depending on the individual health gains, our function can representeither preferences for concentrating or preferences for spreading total gain or both together, an issue which has notbeen addressed until now. Based on an experiment, we observe that this generalisation provides a suitable approximation tothe sampled social preferences.
Resumo:
The paper develops a method to solve higher-dimensional stochasticcontrol problems in continuous time. A finite difference typeapproximation scheme is used on a coarse grid of low discrepancypoints, while the value function at intermediate points is obtainedby regression. The stability properties of the method are discussed,and applications are given to test problems of up to 10 dimensions.Accurate solutions to these problems can be obtained on a personalcomputer.
Resumo:
That individuals contribute in social dilemma interactions even when contributing is costly is a well-established observation in the experimental literature. Since a contributor is always strictly worse off than a non-contributor the question is raised if an intrinsic motivation to contribute can survive in an evolutionary setting. Using recent results on deterministic approximation of stochastic evolutionary dynamics we give conditions for equilibria with a positive number of contributors to be selected in the long run.
Resumo:
This work aims the applicability of the Transient electromagnetic method at an arid and semiarid environmental condition in the Santiago Island – Cape Verde. Some seashore areas of this island show an increasing salt contamination of the groundwater. The main objective of present work is to relate this water-quality condition with parameters taken from the transient sounding’s data. In this context, transient soundings have been acquired from 2005 through 2009, at several chosen valleys near the sea, in a mean rate of one field campaign each year. The first phase of this work was the understanding of the geophysical method details, problems and applicability, as the chosen and acquired equipment was the first one to be permanently available to the Portuguese geosciences community. This first phase was also accomplished with field tests. Interpretation of the transient sounding’s data curves were done by application of 1-D inversion methods already developed and published, as also with quasi 2-D and quasi 3-D inversion algorithms, where applicability was feasible. This was the second phase. The 2-D and 3-D approximation results are satisfactory and promising; although a higher spatial sounding’s density should certainly allow for better results. At phase three, these results have been compared against the available lithologic, hydrologic and hydrochemical data, in the context of Santiago’s island settings. The analyses of these merged data showed that two distinct origins for the observed inland groundwater salinity are possible; seashore shallow mixing with contemporary seawater and mixing with a deep and older salty layer from up flow groundwater. Relations between the electric resistivity and the salt water content distribution were found for the surveyed areas. To this environment condition, the electromagnetic transient method proved to be a reliable and powerful technique. The groundwater quality can be accessed beyond the few available watershed points, which have an uneven distribution.
Resumo:
Given $n$ independent replicates of a jointly distributed pair $(X,Y)\in {\cal R}^d \times {\cal R}$, we wish to select from a fixed sequence of model classes ${\cal F}_1, {\cal F}_2, \ldots$ a deterministic prediction rule $f: {\cal R}^d \to {\cal R}$ whose risk is small. We investigate the possibility of empirically assessingthe {\em complexity} of each model class, that is, the actual difficulty of the estimation problem within each class. The estimated complexities are in turn used to define an adaptive model selection procedure, which is based on complexity penalized empirical risk.The available data are divided into two parts. The first is used to form an empirical cover of each model class, and the second is used to select a candidate rule from each cover based on empirical risk. The covering radii are determined empirically to optimize a tight upper bound on the estimation error. An estimate is chosen from the list of candidates in order to minimize the sum of class complexity and empirical risk. A distinguishing feature of the approach is that the complexity of each model class is assessed empirically, based on the size of its empirical cover.Finite sample performance bounds are established for the estimates, and these bounds are applied to several non-parametric estimation problems. The estimates are shown to achieve a favorable tradeoff between approximation and estimation error, and to perform as well as if the distribution-dependent complexities of the model classes were known beforehand. In addition, it is shown that the estimate can be consistent,and even possess near optimal rates of convergence, when each model class has an infinite VC or pseudo dimension.For regression estimation with squared loss we modify our estimate to achieve a faster rate of convergence.
Resumo:
We lay out a small open economy version of the Calvo sticky price model, and show how the equilibrium dynamics can be reduced to simple representation in domestic inflation and the output gap. We use the resulting framework to analyze the macroeconomic implications of three alternative rule-based policy regimes for the small open economy: domestic inflation and CPI-based Taylor rules, and an exchange rate peg. We show that a key difference amongthese regimes lies in the relative amount of exchange rate volatility that they entail. We also discuss a special case for which domestic inflation targeting constitutes the optimal policy, and where a simple second order approximation to the utility of the representative consumer can be derived and used to evaluate the welfare losses associated with the suboptimal rules.
Resumo:
A family of scaling corrections aimed to improve the chi-square approximation of goodness-of-fit test statistics in small samples, large models, and nonnormal data was proposed in Satorra and Bentler (1994). For structural equations models, Satorra-Bentler's (SB) scaling corrections are available in standard computer software. Often, however, the interest is not on the overall fit of a model, but on a test of the restrictions that a null model say ${\cal M}_0$ implies on a less restricted one ${\cal M}_1$. If $T_0$ and $T_1$ denote the goodness-of-fit test statistics associated to ${\cal M}_0$ and ${\cal M}_1$, respectively, then typically the difference $T_d = T_0 - T_1$ is used as a chi-square test statistic with degrees of freedom equal to the difference on the number of independent parameters estimated under the models ${\cal M}_0$ and ${\cal M}_1$. As in the case of the goodness-of-fit test, it is of interest to scale the statistic $T_d$ in order to improve its chi-square approximation in realistic, i.e., nonasymptotic and nonnormal, applications. In a recent paper, Satorra (1999) shows that the difference between two Satorra-Bentler scaled test statistics for overall model fit does not yield the correct SB scaled difference test statistic. Satorra developed an expression that permits scaling the difference test statistic, but his formula has some practical limitations, since it requires heavy computations that are notavailable in standard computer software. The purpose of the present paper is to provide an easy way to compute the scaled difference chi-square statistic from the scaled goodness-of-fit test statistics of models ${\cal M}_0$ and ${\cal M}_1$. A Monte Carlo study is provided to illustrate the performance of the competing statistics.
Resumo:
We develop a general error analysis framework for the Monte Carlo simulationof densities for functionals in Wiener space. We also study variancereduction methods with the help of Malliavin derivatives. For this, wegive some general heuristic principles which are applied to diffusionprocesses. A comparison with kernel density estimates is made.
Resumo:
The paper defines concepts of real wealth and saving which take into account the intertemporal index number problem that results from changing interest rates. Unlike conventional measures of real wealth, which are based on the market value of assets and ignore the index number problem, the new measure correctly reflects the changes in the welfare of households over time. An empirically operational approximation to the theoretical measure is provided and applied to US data. A major empirical finding is that US real financial wealth increased strongly in the 1980s, much more than is revealed by the market value of assets.
Resumo:
RESUME Dès le printemps 2004, la construction d'une 2ème ligne de métro est entreprise dans la ville de Lausanne en Suisse. En reliant Ouchy, au bord du lac Léman (alt. 373 m) à Epalinges (alt. 711 m), le nouveau métro "M2" traversera dès 2008 l'agglomération lausannoise du Sud au Nord sur une distance de 6 km. Depuis l'avant-projet, en 1999, une grande quantité de données géologiques a été récolté et de nombreux forages exécutés sur le site. Ceci nous a donné une occasion unique d'entreprendre une étude de microgravimétrique urbaine de détail. Le mode de creusement du tunnel dépend fortement des matériaux à excaver et il est classiquement du domaine du géologue, avec ses connaissances de la géologie régionale et de la stratigraphie des forages, de fournir à l'ingénieur un modèle géologique. Ce modèle indiquera dans ce cas l'épaisseur des terrains meubles qui recouvrent le soubassement rocheux. La représentativité spatiale d'une information très localisée, comme celle d'un forage, est d'autant plus compliquée que le détail recherché est petit. C'est à ce moment là que la prospection géophysique, plus spécialement gravimétrique, peut apporter des informations complémentaires déterminantes pour régionaliser les données ponctuelles des forages. La microgravimétrie en milieu urbain implique de corriger avec soin les perturbations gravifiques sur la mesure de la pesanteur dues aux effets de la topographie, des bâtiments et des caves afin d'isoler l'effet gravifique dû exclusivement à l'épaisseur du remplissage des terrains meubles. Tenant compte de l'intensité des corrections topographiques en milieu urbain, nous avons donné une grande importance aux sous-sols, leurs effets gravifiques pouvant atteindre l'ordre du dixième de mGal. Nous avons donc intégré ces corrections celle de topographie et traité les effets des bâtiments de manière indépendante. Nous avons inclus dans le modèle numérique de terrain (MNT) la chaussée et les sous-sols afin de construire un modèle numérique de terrain urbain. Nous utiliserons un nouvel acronyme « MNTU »pour décrire ce modèle. Nous proposons d'établir des cartes de corrections topographiques préalables, basées sur les données à disposition fournies par le cadastre en faisant des hypothèses sur la profondeur des sous-sols et la hauteur des bâtiments. Les deux zones de test choisies sont caractéristiques des différents types d'urbanisation présente à Lausanne et se révèlent par conséquent très intéressantes pour élaborer une méthodologie globale de la microgravimétrie urbaine. Le but était d'évaluer l'épaisseur du remplissage morainique sur un fond rocheux molassique se situant à une profondeur variable de quelques mètres à une trentaine de mètres et d'en établir une coupe dans l'axe du futur tracé du métro. Les résultats des modélisations se sont révélés très convaincants en détectant des zones qui diffèrent sensiblement du modèle géologique d'avant projet. Nous avons également démontré que l'application de cette méthode géophysique, non destructive, est à même de limiter le nombre de sondages mécaniques lors de l'avant-projet et du projet définitif, ce qui peut limiter à la fois les coûts et le dérangement engendré par ces travaux de surface. L'adaptabilité de la technique gravimétrique permet d'intervenir dans toutes les différentes phases d'un projet de génie civil comme celui de la construction d'un métro en souterrain. KURZFASSUNG Seit dem Frühling 2004 ist in der Stadt Lausanne (Schweiz) die neue U-Bahn "M2" in Konstruktion. Diese soll auf 6 km Länge die Lausanner Agglomeration von Süd nach Nord durchqueren. Die dem Projekt zu Grunde liegende technische Planung sieht vor, daß die Bahnlinie hauptsächlich in der Molasse angesiedelt sein wird. Seit dem Vorentwurf (1999) ist eine große Anzahl geologischer Angaben gesammelt worden. Daraus ergab sich die einmalige Gelegenheit, die Informationen aus den damit verbundenen zahlreichen Bohrungen zu einer detaillierten mikrogravimetrischen Studie der Stadt Lausanne zu erweitern und zu vervollständigen. Das Ziel bestand darin, die Mächtigkeit der die Molasseüberdeckenden Moräneablagerung abzuschätzen, um eine entsprechendes geologisches Profile entlang der künftigen Bahnlinie zu erstellen. Weiterhin sollte gezeigt werden, daß die Anwendung dieser nicht-invasiven geophysikalischen Methode es ermöglicht, die Anzahl der benötigten Bohrungen sowohl in der Pilotphase wie auch im endgültigen Projekt zu reduzieren, was zu wesentlichen finanziellen Einsparungen in der Ausführung des Werkes beitragen würde. Die beiden in dieser Studie bearbeiteten Testzonen befinden sich im Nordteil und im Stadtzentrum von Lausanne und sind durch eine unterschiedliche Urbanisierung charakterisiert. Das anstehende Gestein liegt in verschiedenen Tiefen: von einigen Metern bis zu etwa dreißig Metern. Diese Zonen weisen alle Schwierigkeiten einer urbanen Bebauung mit hoher Verkehrsdichte auf und waren daher massgebend bei der Ausarbeitung einer globalen mikrogravimetrischen Methodologie für die Stadt Lausanne. Die so entwickelte Technik ermöglicht, die störenden Auswirkungen der Topographie, der Gebäude, der Keller und der Öffentlichen Infrastrukturen sorgfältig zu korrigieren, um so die ausschließlich auf die Mächtigkeit des Lockergesteins zurückzuführenden Effekte zu isolieren. In Bezug auf die Intensität der Auswirkungen der topographischen Korrekturen im Stadtgebiet wurde den Untergeschossen eine besonders grosse Bedeutung zugemessen da die entsprechenden Schwerkrafteffekte eine Grösse von rund einem Zehntel mGal erreichen können. Wir schlagen deshalb vor, vorläufige Karten der topographischen Korrekturen zu erstellen. Diese Korrekturen basieren auf den uns vom Katasterplan gelieferten Daten und einigen Hypothesen bezüglich der Tiefe der Untergeschosse und der Höhe der Gebäude. Die Verfügbarkeit einer derartigen Karte vor der eigentlichen gravimetrischen Messkampagne würde uns erlauben, die Position der Meßstationen besser zu wählen. Wir sahen zudem, daß ein entsprechenden a priori Filter benutzt werden kann, wenn die Form und die Intensität der Anomalie offensichtlich dem entsprechenden Gebäude zugeordnet werden können. Diese Strategie muß jedoch mit Vorsicht angewandt werden, denn falls weitere Anomalien dazukommen, können bedeutende Verschiebungen durch Übèrlagerungen der Schwerewirkung verschiedener Strukturen entstehen. Die Ergebnisse der Modellierung haben sich als sehr überzeugend erwiesen, da sie im Voraus unbekannte sensible Zonen korrekt identifiziert haben. Die Anwendbarkeit der in dieser Arbeit entwickelten gravimetrischen Technik ermöglicht es, während allen Phasen eines Grossbauprojekts, wie zum Beispiel bei der Konstruktion einer unterirdischen U-Bahn, einzugreifen. ABSTRACT Since Spring of 2004 a new metro line has been under construction in the city of Lausanne in Switzerland. The new line, the M2, will be 6 km long and will traverse the city from south to north. The civil engineering project determined that the line would be located primarily in the Molasse. Since the preparatory project in 1999, a great quantity of geological data has been collected, and the many drillings made on the site have proved to be a unique opportunity to undertake a study of urban microgravimetry. The goal was to evaluate the thickness of the morainic filling over the molassic bedrock, and to establish a section along the axis of the future line. It then had to be shown that the application of this nondestructive geophysical method could reduce the number of mechanical surveys required both for a preparatory and a definitive project, which would lead to real savings in the realization of a civil engineering project. The two test zones chosen, one in the northern part of the city and one in the city centre, are characterised by various types of urbanisation. Bedrock is at a depth varying from a few metres to about thirty metres. These zones well exemplify the various difficulties encountered in an urban environment and are therefore very interesting for the development of an overall methodology of urban microgravimetry. Microgravimetry in an urban environment requires careful corrections for gravific disturbances due to the effects of topography, buildings, cellars, and the infrastructure of distribution networks, in order to isolate the gravific effect due exclusively to the thickness of loose soil filling. Bearing in mind the intensity of the topographic corrections in an urban environment, we gave particular importance to basements. Their gravific effects can reach the order of one tenth of one meal, and can influence above all the precision of the Bouguer anomaly. We propose to establish preliminary topographic correction charts based on data provided to us by the land register, by making assumptions on the depths of basements and the heights of buildings. Availability of this chart previous to a gravimetry campaign would enable us to choose optimum measuring sites. We have also seen that an a priori filter can be used when the form and the intensity of the anomaly correspond visually to the corresponding building. This strategy must be used with caution because if other anomalies are to be associated, important shifts can be generated by the superposition of the effects of different structures. The results of the model have proved to be very convincing in detecting previously unknown sensitive zones. The adaptability of the gravimetry technique allows for application in all phases of a civil engineering project such as the construction of an underground metro line. RIASSUNTO Dalla primavera 2004 una nuova linea metropolitana é in costruzione nella città di Losanna in Svizzera. La nuova metropolitana "M2" traverserà per la lunghezza di 6 km il centro urbano di Losanna da sud a nord. II progetto d'ingegneria civile prevedeva un tracciato situato essenzialmente nel fondo roccioso arenaceo terziario (molassa). Dalla redazione del progetto preliminare, avvenuta nel 1999, una grande quantità di dati geologici sono stati raccolti e sono stati eseguiti numerosi sondaggi. Questo sì é presentato come un'occasione unica per mettere a punto uno studio microgravimetrico in ambiente urbano con lo scopo di valutare lo spessore dei terreni sciolti di origine glaciale che ricoprono il fondo roccioso di molassa e di mettere in evidenza come l'applicazione di questo metodo geofisico non distruttivo possa limitare il numero di sondaggi meccanici nella fase di progetto preliminare ed esecutivo con conseguente reale risparmio economico nella realizzazione di una tale opera. Le due zone di test sono situate una nella zona nord e la seconda nel centro storico di Losanna e sono caratterizzate da stili architettonici differenti. II fondo roccioso é situato ad una profondità variabile da qualche metro ad una trentina. Queste due zone sembrano ben rappresentare tutte le difficoltà di un ambiente urbano e ben si prestano per elaborare una metodologia globale per la microgravimetria in ambiente urbano. L'applicazione di questa tecnica nell'ambiente suddetto implica la correzione attenta delle perturbazioni sulla misura dell'accelerazione gravitazionale, causate dalla topografia, gli edifici, le cantine e le infrastrutture dei sottoservizi, per ben isolare il segnale esclusivamente causato dallo spessore dei terreni sciolti. Tenuto conto, dell'intensità delle correzioni topografiche, abbiamo dato grande importanza alle cantine, poiché il loro effetto sulle misure può raggiungere il decimo di mGal. Proponiamo quindi di redigere una carta delle correzioni topografiche preliminare all'acquisizione, facendo delle ipotesi sulla profondità delle cantine e sull'altezza degli edifici, sulla base delle planimetrie catastali. L'analisi di questa carta permetterà di scegliere le posizioni più adatte per le stazioni gravimetriche. Abbiamo anche osservato che un filtro a priori, qualora la forma e l'intensità dell'anomalia fosse facilmente riconducibile in maniera visuale ad un edificio, possa essere efficace. Tuttavia questa strategia deve essere utilizzata con precauzione, poiché può introdurre uno scarto, qualora più anomalie, dovute a differenti strutture, si sovrappongano. I risultati delle modellizzazioni si sono rivelati convincenti, evidenziando zone sensibili non conosciute preventivamente. L'adattabilità della tecnica gravimetrica ha mostrato di poter intervenire in differenti fasi di un progetto di ingegneria civile, quale è quella di un'opera in sotterraneo.
Resumo:
In a closed economy context there is common agreement on price inflation stabilization being one of the objects of monetary policy. Moving to an open economy context gives rise to the coexistence of two measures of inflation: domestic inflation (DI) and consumer price inflation (CPI). Which one of the two measures should be the target variable? This is the question addressed in this paper. In particular, I use a small open economy model to show that once sticky wages indexed to past CPI inflation are introduced, a complete inward looking monetary policy is no more optimal. I first, derive a loss function from a secondorder approximation of the utility function and then, I compute the fully optimalmonetary policy under commitment. Then, I use the optimal monetary policy as a benchmark to compare the performance of different monetary policy rules. The main result is that once a positive degree of indexation is introduced in the model the rule performing better (among the Taylor type rules considered) is the one targeting wage inflation and CPI inflation. Moreover this rule delivers results very close to the one obtained under the fully optimal monetary policy with commitment.
Resumo:
The approximants to regular continued fractions constitute `best approximations' to the numbers they converge to in two ways known as of the first and the second kind.This property of continued fractions provides a solution to Gosper's problem of the batting average: if the batting average of a baseball player is 0.334, what is the minimum number of times he has been at bat? In this paper, we tackle somehow the inverse question: given a rational number P/Q, what is the set of all numbers for which P/Q is a `best approximation' of one or the other kind? We prove that inboth cases these `Optimality Sets' are intervals and we give aprecise description of their endpoints.
Resumo:
Este artigo visa a apresentar um contributo para a reflexão acerca da integração de Cabo Verde na CEDEAO (Comunidade Económica dos Estados da África Ocidental) e sobre a aproximação de Cabo Verde às RUP (Regiões Ultraperiféricas) e, posteriormente, à União Europeia, tirando-se proveito da realidade de uma Nação que se gerou e se vai consolidando a partir da cooperação de povos e culturas oriundos dos dois espaços geográficos referidos. A discussão não se deve somente à posição geográfica privilegiada do arquipélago, mas, igualmente, liga-se a aspectos de ordem cultural, política, económica, comercial e de segurança, com especial destaque para a estruturação do Estado-Nação em Cabo Verde.
Resumo:
Este artigo procura compreender a integração regional do Arquipélago de Cabo Verde, na base do debate contemporâneo sobre a Nação. Resultado de cruzamento e convergência entre povos e culturas oriundos de dois espaços geográficos (África e Europa), a Nação cabo-verdiana afirma-se e consolida-se na margem dos debates perspetivados e delineados por diferentes gerações de intelectuais, políticos e académicos nacionais e estrangeiros. Estes procuram compreender e analisar a integração de Cabo Verde na Comunidade Económica dos Estados da África Ocidental (CEDEAO), a aproximação às Regiões Ultraperiféricas (RUP) e a Parceria Especial com a União Europeia (UE). A análise expressa no presente artigo, além de distinguir a posição geográfica privilegiada do arquipélago, enfatiza também aspetos de ordem cultural, política, económica, comercial e de segurança, relevantes para a construção da Nação em Cabo Verde.
Resumo:
The landslide of Rosiana is considered the largest slope movement amongst those known in historical times in Gran Canana, Canary Islands. It has been activated at least 4 times in the last century, and in the movement of 1956, when about 3.106 m3 of materials were involved, 250 people had to be evacuated and many buildings were destroyed. The present geological hazard has lead to specific studies of the phenomenon which, once characterised, can be used as a guide for the scientific and technical works that are to be made in this or similar areas. This paper wants to increase the knowledge about the unstable mass of Rosiana by using geophysical techniques based on the method of seismic by refraction. The geophysical measues have been interpreted with the aid of the available geomorphologic data, thus obtaining a first approximation to the geometry of the slope movements