936 resultados para Contingent Valuation


Relevância:

20.00% 20.00%

Publicador:

Resumo:

A growing body of longitudinal studies suggests that low self-esteem is a risk factor for depression. However, it is unclear whether other characteristics of self-esteem, besides its level, explain incremental or even greater variance in subsequent depression. We examined the prospective effects of self-esteem level, instability (i.e., the degree of variability in self-esteem across short periods), and contingency (i.e., the degree to which self-esteem fluctuates in response to self-relevant events) on depressive symptoms in 1 overarching model, using data from 2 longitudinal studies. In Study 1, 372 adults were assessed at 2 waves over 6 months, including 40 daily diary assessments at Wave 1. In Study 2, 235 young adults were assessed at 2 waves over 6 weeks, including about 6 daily diary assessments at each wave. Self-esteem contingency was measured by self-report and by a statistical index based on the diary data (capturing event-related fluctuations in self-esteem). In both studies self-esteem level, but not self-esteem contingency, predicted subsequent depressive symptoms. Self-esteem instability predicted subsequent depressive symptoms in Study 2 only, with a smaller effect size than self-esteem level. Also, level, instability, and contingency of self-esteem did not interact in the prediction of depressive symptoms. Moreover, the effect of self-esteem level held when controlling for neuroticism and for all other Big Five personality traits. Thus, the findings provide converging evidence for a vulnerability effect of self-esteem level, tentative evidence for a smaller vulnerability effect of self-esteem instability, and no evidence for a vulnerability effect of self-esteem contingency.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

External circumstances and internal bodily states often change and require organisms to flexibly adapt valuation processes to select the optimal action in a given context. Here, we investigate the neurobiology of context-dependent valuation in 22 human subjects using functional magnetic resonance imaging. Subjects made binary choices between visual stimuli with three attributes (shape, color, and pattern) that were associated with monetary values. Context changes required subjects to deviate from the default shape valuation and to integrate a second attribute in order to comply with the goal to maximize rewards. Critically, this binary choice task did not involve any conflict between opposing monetary, temporal, or social preferences. We tested the hypothesis that interactions between regions of dorsolateral and ventromedial prefrontal cortex (dlPFC; vmPFC) implicated in self-control choices would also underlie the more general function of context-dependent valuation. Consistent with this idea, we found that the degree to which stimulus attributes were reflected in vmPFC activity varied as a function of context. In addition, activity in dlPFC increased when context changes required a reweighting of stimulus attribute values. Moreover, the strength of the functional connectivity between dlPFC and vmPFC was associated with the degree of context-specific attribute valuation in vmPFC at the time of choice. Our findings suggest that functional interactions between dlPFC and vmPFC are a key aspect of context-dependent valuation and that the role of this network during choices that require self-control to adjudicate between competing outcome preferences is a specific application of this more general neural mechanism.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Web surveys are becoming increasingly popular in survey research. Compared with face-to-face, telephone and mail surveys, web surveys may contain a different and new source of measurement error and bias: the type of device that respondents use to answer the survey questions. To the best of our knowledge, this is the first study that tests whether the use of mobile devices affects survey characteristics and stated preferences in a web-based choice experiment. The web survey was carried out in Germany with 3,400 respondents, of which 12 per cent used a mobile device (i.e. tablet or smartphone), and comprised a stated choice experiment on externalities of renewable energy production using wind, solar and biomass. Our main finding is that survey characteristics such as interview length and acquiescence tendency are affected by the device used. In contrast to what might be expected, we find that, compared with respondents using desktop computers and laptops, mobile device users spent more time to answer the survey and are less likely to be prone to acquiescence bias. In the choice experiment, mobile device users tended to be more consistent in their stated choices, and there are differences in willingness to pay between both subsamples.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

L’article suivant rend compte des résultats d’un atelier tenu à l’occasion du congrès 2015 de la SEVAL. L’atelier portait sur l’indépendance des évaluateurs et des évaluatrices du point de vue de leur relation avec les commanditaires de l’évaluation. Sur la base d’un exemple de mandat concret présenté par deux intervenants – l’évaluatrice ainsi que le conseiller d’État mandant de l’évaluation en question – les différents garde-fous contractuels, communicationnels et procéduraux aptes à garantir une pleine indépendance de l’évaluation ont été discutés. Il en est ressorti que quatre dimensions étaient cruciales pour le succès de la démarche : les conditions-cadres de l’évaluation, la récolte des données, les relations avec les évalués, ainsi que la phase de restitution des résultats.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

In this paper, we extend the debate concerning Credit Default Swap valuation to include time varying correlation and co-variances. Traditional multi-variate techniques treat the correlations between covariates as constant over time; however, this view is not supported by the data. Secondly, since financial data does not follow a normal distribution because of its heavy tails, modeling the data using a Generalized Linear model (GLM) incorporating copulas emerge as a more robust technique over traditional approaches. This paper also includes an empirical analysis of the regime switching dynamics of credit risk in the presence of liquidity by following the general practice of assuming that credit and market risk follow a Markov process. The study was based on Credit Default Swap data obtained from Bloomberg that spanned the period January 1st 2004 to August 08th 2006. The empirical examination of the regime switching tendencies provided quantitative support to the anecdotal view that liquidity decreases as credit quality deteriorates. The analysis also examined the joint probability distribution of the credit risk determinants across credit quality through the use of a copula function which disaggregates the behavior embedded in the marginal gamma distributions, so as to isolate the level of dependence which is captured in the copula function. The results suggest that the time varying joint correlation matrix performed far superior as compared to the constant correlation matrix; the centerpiece of linear regression models.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This paper develops a reduced form three-factor model which includes a liquidity proxy of market conditions which is then used to provide implicit prices. The model prices are then compared with observed market prices of credit default swaps to determine if swap rates adequately reflect market risks. The findings of the analysis illustrate the importance of liquidity in the valuation process. Moreover, market liquidity, a measure of investors. willingness to commit resources in the credit default swap (CDS) market, was also found to improve the valuation of investors. autonomous credit risk. Thus a failure to include a liquidity proxy could underestimate the implied autonomous credit risk. Autonomous credit risk is defined as the fractional credit risk which does not vary with changes in market risk and liquidity conditions.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

I propose that the Last in, First out (LIFO) inventory valuation method needs to be reevaluated. I will evaluate the impact of the LIFO method on earnings of publically traded companies with a LIFO reserve over the past 10 years. I will begin my proposal with the history of how the LIFO method became an acceptable valuation method and discuss the significance of LIFO within the accounting profession Next I will provide a description of LIFO, the First in, First out (FIFO), and the weighted average inventory valuation methods and explore the differences among each. More specifically, I will explore the arguments for and against the use of the LIFO method and the potential shift towards financial standards that do not allow LIFO (a standard adopted and influenced by the International Financial Accounting Standards Board). Data will be collected from Compustat for publicly traded companies (with a LIFO Reserve) for the past 10 years. I will document which firms use LIFO, analyze trends relating to LIFO usage and LIFO reserves (the difference in the cost of inventory between using LIFO and FIFO), and evaluate the effect on earnings. The purpose of this research is to evaluate the accuracy of LIFO in portraying earnings and to see how much tax has gone uncollected over the years because of the use of LIFO. Moreover, I will provide an opinion as to whether U.S. GAAP should adopt a standard similar to IFRS and ban the LIFO method.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The purpose of this paper is to increase current empirical evidence on the relevance of real options for explaining firm investment decisions in oligopolistic markets. We study an actual investment case in the Spanish mobile telephony industry, the entrant in the market of a new operator, Yoigo. We analyze the option to abandon in order to show the relevance of the possibility of selling the company in an oligopolistic market where competitors are not allowed free entrance. The NPV (net present value) of the new entrant is calculated as a starting point. Then, based on the general approach proposed by Copeland and Antikarov (2001), a binomial tree is used to model managerial flexibility in discrete time periods, and value the option to abandon. The strike price of the option is calculated based on incremental EBITDA margins due to selling customers or merging with a competitor.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The present study analyzes residential models in coastal areas with large influxes of tourism, the sustainability of their planning and its repercussion on urban values. The project seeks to establish a methodology for territorial valuation through the analysis of externalities that have influenced urban growth and its impact on the formation of residential real estate values. This will make it possible to create a map for qualitative land valuation, resulting from a combination of environmental, landscape, social and productive valuations. This in turn will establish a reference value for each of the areas in question, as well as their spatial interrelations. These values become guidelines for the study of different territorial scenarios, which help improve the sustainable territorial planning process. This is a rating scale for urban planning. The results allow us to establish how the specific characteristics of the coast are valued and how they can be incorporated into sustainable development policies.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Après le séisme qui eut lieu en Haïti le 12 janvier 2010, dont l’épicentre a été localisé non loin de la capitale, Port-au-Prince (25 km en direction sud-est), d’une magnitude Mw 7.0 et à une profondeur de 13 km, le pays s’est retrouvé dans une situation catastrophique et d’extrême pauvreté, avec des graves carences en matière de santé, nutrition, éducation et logement. Les effets du tremblement de terre ont été dévastateurs pour la population : on compte plus de 300.000 morts, presque autant de blessés et 1,3 millions de sans-abri logès dans des campements « provisoires ». Quant aux conséquences matérielles, le séisme a totalement détruit près de 100.000 maisons et endommagé près de 200.000 (source : USGS). Ce tremblement de terre a été le plus fort enregistré dans la zone depuis 1770. De plus le séisme fut perceptible dans des pays voisins comme Cuba, la Jamaïque et la République Dominicaine, où il a provoqué l’alarme et des évacuations préventives. La reconstruction du pays reste un sujet prioritaire pour la coopération internationale. Le présent projet, SISMO-HAITÍ, a été développé dans le but d’apporter la connaissance et l’information nécessaires afin de faciliter la prise de mesures préventives face au risque sismique existant, afin d’éviter qu’un éventuel futur séisme ne déclenche une nouvelle catastrophe. Dans le cas d’Haïti, aucune institution n’était chargée d’assurer une surveillance sismique, mais un contact direct a été établi avec l’Observatoire National de l’Environnement et de la Vulnérabilité (ONEV) en Haïti à travers son directeur Dwinel Belizaire Ing. M. Sc., qui est précisément celui qui a sollicité l’aide qui a motivé la présente proposition. Le but ultime de ce projet est l’étude des mesures d’atténuation du risque élevé qui demeure, contribuant ainsi au développement durable de la région. Dans cette optique, la menace sismique en Haïti a été évaluée, constituant la base sur laquelle on prétend élaborer des critères de conception parasismique pour la reconstruction du pays qui pourront être inclus dans la première version de la norme antisismique, ainsi que le risque sismique à Port-au-Prince, dont les résultats serviront de base pour élaborer les plans d’urgence face à ce risque naturel. Les objectifs spécifiques atteints sont : • Évaluation de l'aléa sismique en Haïti. On en obtient des cartes de différents paramètres de mouvement pour différentes probabilités de dépassement (ce qui suppose connaître la probabilité associée aux mouvements dus à des tremblements futurs). • Évaluation de l'effet local à Port-au-Prince et élaboration d'une carte de microzonage de la ville. • Étude de la vulnérabilité sismique locale à Port-au-Prince. • Estimation du risque sismique à Port-au-Prince. • Mesures d'atténuation du risque et de conception parasismique. Ce rapport résume les activités et les résultats obtenus a cours de l'année 2011 lors de la mise en œuvre de ce projet. Le groupe de travail est une équipe multidisciplinaire composée de chercheurs de différents établissements universitaires et de recherche (Université Polytechnique de Madrid-UPM-, Conseil Supérieur de la Recherche Scientifique (CSIC) ; U. Complutense de Madrid-UCM-, U-UA-Alicante, Almeria-UAL-U., U. Autonome de Saint-Domingue, UASD et Université de Porto Rico Mayagüez--UPRM) experts dans les diverses matières impliquées dans le projet: géologie, sismologie, génie parasismique, architecture et gestion de l'information géographique. Tous les membres de cette équipe ont travaillé ensemble tout au long de l'année, en réalisant des réunions, des ateliers de travail, des vidéoconférences, en plus d'une visite à Port-au-Prince en Juillet 2011 afin de procéder à la première collecte de données.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Modern object oriented languages like C# and JAVA enable developers to build complex application in less time. These languages are based on selecting heap allocated pass-by-reference objects for user defined data structures. This simplifies programming by automatically managing memory allocation and deallocation in conjunction with automated garbage collection. This simplification of programming comes at the cost of performance. Using pass-by-reference objects instead of lighter weight pass-by value structs can have memory impact in some cases. These costs can be critical when these application runs on limited resource environments such as mobile devices and cloud computing systems. We explore the problem by using the simple and uniform memory model to improve the performance. In this work we address this problem by providing an automated and sounds static conversion analysis which identifies if a by reference type can be safely converted to a by value type where the conversion may result in performance improvements. This works focus on C# programs. Our approach is based on a combination of syntactic and semantic checks to identify classes that are safe to convert. We evaluate the effectiveness of our work in identifying convertible types and impact of this transformation. The result shows that the transformation of reference type to value type can have substantial performance impact in practice. In our case studies we optimize the performance in Barnes-Hut program which shows total memory allocation decreased by 93% and execution time also reduced by 15%.