958 resultados para real property
Resumo:
We evaluate the profitability of investments in residential property in Germany after unification with a focus on the comparison of East and West Germany. Calculations are carried out for (1) the after-tax return an investor might have expected at the beginning of the 1990s, and (2) the after-tax return that has been realized ten years after. We compare a set of statistical data for investments in fifty major cities by using complete financial budgeting. The results show that tax subsidies could not always protect investors from losing money, but they have boosted realized returns after tax considerably. Therefore, it was indeed the taxpayers, not the investors, who have borne the cost of reconstructing East Germany.
Resumo:
This paper examines the mean-reverting property of real exchange rates. Earlier studies have generally not been able to reject the null hypothesis of a unit-root in real exchange rates, especially for the post-Bretton Woods floating period. The results imply that long-run purchasing power parity does not hold. More recent studies, especially those using panel unit-root tests, have found more favorable results, however. But, Karlsson and Löthgren (2000) and others have recently pointed out several potential pitfalls of panel unit-root tests. Thus, the panel unit-root test results are suggestive, but they are far from conclusive. Moreover, consistent individual country time series evidence that supports long-run purchasing power parity continues to be scarce. In this paper, we test for long memory using Lo's (1991) modified rescaled range test, and the rescaled variance test of Giraitis, Kokoszka, Leipus, and Teyssière (2003). Our testing procedure provides a non-parametric alternative to the parametric tests commonly used in this literature. Our data set consists of monthly observations from April 1973 to April 2001 of the G-7 countries in the OECD. Our two tests find conflicting results when we use U.S. dollar real exchange rates. However, when non-U.S. dollar real exchange rates are used, we find only two cases out of fifteen where the null hypothesis of an unit-root with short-term dependence can be rejected in favor of the alternative hypothesis of long-term dependence using the modified rescaled range test, and only one case when using the rescaled variance test. Our results therefore provide a contrast to the recent favorable panel unit-root test results.
Resumo:
Thirty-six US states have already enacted some form of seller's property condition disclosure law. At a time when there is a movement in this direction nationally, this paper attempts to ascertain the factors that lead states to adopt disclosure law. Motivation for the study stems from the fact that not all states have yet adopted the law, and states that have enacted the law have done so in different years. The analytical structure employs hazard models, using a unique set of economic and institutional attributes for a panel of 50 US States spanning 21 years, from 1984 to 2004. The proportional hazard analysis of law adoption reveals that greater number of disciplinary actions tends to favor passage of the law. Greater broker supervision, implying generally higher awareness among real estate agents, seems to have a negative impact on the likelihood of a state adopting a property condition disclosure law.
Resumo:
At the time when at least two-thirds of the US states have already mandated some form of seller's property condition disclosure statement and there is a movement in this direction nationally, this paper examines the impact of seller's property condition disclosure law on the residential real estate values, the information asymmetry in housing transactions and shift of risk from buyers and brokers to the sellers, and attempts to ascertain the factors that lead to adoption of the disclosur law. The analytical structure employs parametric panel data models, semi-parametric propensity score matching models, and an event study framework using a unique set of economic and institutional attributes for a quarterly panel of 291 US Metropolitan Statistical Areas (MSAs) and 50 US States spanning 21 years from 1984 to 2004. Exploiting the MSA level variation in house prices, the study finds that the average seller may be able to fetch a higher price (about three to four percent) for the house if she furnishes a state-mandated seller's property condition disclosure statement to the buyer.
Resumo:
We examine the impact of seller's Property Condition Disclosure Law on the residential real estate values. A disclosure law may address the information asymmetry in housing transactions shifting of risk from buyers and brokers to the sellers and raising housing prices as a result. We combine propensity score techniques from the treatment effects literature with a traditional event study approach. We assemble a unique set of economic and institutional attributes for a quarterly panel of 291 US Metropolitan Statistical Areas (MSAs) and 50 US States spanning 21 years from 1984 to 2004 is used to exploit the MSA level variation in house prices. The study finds that the average seller may be able to fetch a higher price (about three to four percent) for the house if she furnishes a state-mandated seller.s property condition disclosure statement to the buyer. When we compare the results from parametric and semi-parametric event analyses, we find that the semi-parametric or the propensity score analysis generals moderately larger estimated effects of the law on housing prices.
Resumo:
This research is concerned with the development of distributed real-time systems, in which software is used for the control of concurrent physical processes. These distributed control systems are required to periodically coordinate the operation of several autonomous physical processes, with the property of an atomic action. The implementation of this coordination must be fault-tolerant if the integrity of the system is to be maintained in the presence of processor or communication failures. Commit protocols have been widely used to provide this type of atomicity and ensure consistency in distributed computer systems. The objective of this research is the development of a class of robust commit protocols, applicable to the coordination of distributed real-time control systems. Extended forms of the standard two phase commit protocol, that provides fault-tolerant and real-time behaviour, were developed. Petri nets are used for the design of the distributed controllers, and to embed the commit protocol models within these controller designs. This composition of controller and protocol model allows the analysis of the complete system in a unified manner. A common problem for Petri net based techniques is that of state space explosion, a modular approach to both the design and analysis would help cope with this problem. Although extensions to Petri nets that allow module construction exist, generally the modularisation is restricted to the specification, and analysis must be performed on the (flat) detailed net. The Petri net designs for the type of distributed systems considered in this research are both large and complex. The top down, bottom up and hybrid synthesis techniques that are used to model large systems in Petri nets are considered. A hybrid approach to Petri net design for a restricted class of communicating processes is developed. Designs produced using this hybrid approach are modular and allow re-use of verified modules. In order to use this form of modular analysis, it is necessary to project an equivalent but reduced behaviour on the modules used. These projections conceal events local to modules that are not essential for the purpose of analysis. To generate the external behaviour, each firing sequence of the subnet is replaced by an atomic transition internal to the module, and the firing of these transitions transforms the input and output markings of the module. Thus local events are concealed through the projection of the external behaviour of modules. This hybrid design approach preserves properties of interest, such as boundedness and liveness, while the systematic concealment of local events allows the management of state space. The approach presented in this research is particularly suited to distributed systems, as the underlying communication model is used as the basis for the interconnection of modules in the design procedure. This hybrid approach is applied to Petri net based design and analysis of distributed controllers for two industrial applications that incorporate the robust, real-time commit protocols developed. Temporal Petri nets, which combine Petri nets and temporal logic, are used to capture and verify causal and temporal aspects of the designs in a unified manner.
Resumo:
Modern distributed control systems comprise of a set of processors which are interconnected using a suitable communication network. For use in real-time control environments, such systems must be deterministic and generate specified responses within critical timing constraints. Also, they should be sufficiently robust to survive predictable events such as communication or processor faults. This thesis considers the problem of coordinating and synchronizing a distributed real-time control system under normal and abnormal conditions. Distributed control systems need to periodically coordinate the actions of several autonomous sites. Often the type of coordination required is the all or nothing property of an atomic action. Atomic commit protocols have been used to achieve this atomicity in distributed database systems which are not subject to deadlines. This thesis addresses the problem of applying time constraints to atomic commit protocols so that decisions can be made within a deadline. A modified protocol is proposed which is suitable for real-time applications. The thesis also addresses the problem of ensuring that atomicity is provided even if processor or communication failures occur. Previous work has considered the design of atomic commit protocols for use in non time critical distributed database systems. However, in a distributed real-time control system a fault must not allow stringent timing constraints to be violated. This thesis proposes commit protocols using synchronous communications which can be made resilient to a single processor or communication failure and still satisfy deadlines. Previous formal models used to design commit protocols have had adequate state coverability but have omitted timing properties. They also assumed that sites communicated asynchronously and omitted the communications from the model. Timed Petri nets are used in this thesis to specify and design the proposed protocols which are analysed for consistency and timeliness. Also the communication system is mcxielled within the Petri net specifications so that communication failures can be included in the analysis. Analysis of the Timed Petri net and the associated reachability tree is used to show the proposed protocols always terminate consistently and satisfy timing constraints. Finally the applications of this work are described. Two different types of applications are considered, real-time databases and real-time control systems. It is shown that it may be advantageous to use synchronous communications in distributed database systems, especially if predictable response times are required. Emphasis is given to the application of the developed commit protocols to real-time control systems. Using the same analysis techniques as those used for the design of the protocols it can be shown that the overall system performs as expected both functionally and temporally.
Resumo:
2000 Mathematics Subject Classification: 12D10
Resumo:
2000 Mathematics Subject Classification: Primary 46B20. Secondary 47A99, 46B42.
Resumo:
Pairwise comparison is a popular assessment method either for deriving criteria-weights or for evaluating alternatives according to a given criterion. In real-world applications consistency of the comparisons rarely happens: intransitivity can occur. The aim of the paper is to discuss the relationship between the consistency of the decision maker—described with the error-free property—and the consistency of the pairwise comparison matrix (PCM). The concept of error-free matrix is used to demonstrate that consistency of the PCM is not a sufficient condition of the error-free property of the decision maker. Informed and uninformed decision makers are defined. In the first stage of an assessment method a consistent or near-consistent matrix should be achieved: detecting, measuring and improving consistency are part of any procedure with both types of decision makers. In the second stage additional information are needed to reveal the decision maker’s real preferences. Interactive questioning procedures are recommended to reach that goal.
Resumo:
This is a dissertation about urban systems; within this broad subject I tackle three issues, one that focuses on an observed inter-city relationship and two that focus on an intra-city phenomenon. In Chapter II I adapt a model of random emergence of economic opportunities from the firm growth literature to the urban dynamics situation and present several predictions for urban system dynamics. One of these predictions is that the older the city the larger and more diversified it is going to be on average, which I proceed to verify empirically using two distinct datasets. In Chapter III I analyze the Residential Real Estate Bubble that took place in Miami-Dade County from 1999 to 2006. I adopt a Spatial-Economic model developed for the Paris Bubble episode of 1984–1993 and formulate an innovative test of the results in terms of speculative intensity on the basis of proxies of investor activity available in my dataset. My results support the idea that the best or more expensive areas are also where the greatest speculative activity takes place and where the rapid increase in prices begins. The most significant departure from previous studies that emerges in my results is the absence of a wider gap between high priced areas and low priced areas in the peak year. I develop a measure of dispersion in value among areas and contrast the Miami-Dade and Paris episodes. In Chapter IV I analyze the impact on tax equity of a Florida tax-limiting legislation known as Save Our Homes. I first compare homesteaded and non-homesteaded properties, and second, look within the subset of homesteaded properties. I find that non-homesteaded properties increase their share of taxes paid relative to homesteaded properties during an up market, but that this is reversed during a down market. For the subset of homesteaded properties I find that the impact on tax equity of SOH will depend on differential growth rates among higher and lower valued homes, but during times of rapid home price appreciation, in a scenario of no differential growth rates in property values, SOH increases progressivity relative to the prior system.
Resumo:
Investors and developers are often faced with the task of determining the worth or value of a real estate entity that presently exists or is proposed for development. This article explains the process for determining the value of a proposed project and, subsequently, the maximum investment dollars the project can cover, while at the same time producing a reasonable return for the investor. A proposed 300-room hotel serves as the real estate entity to be analyzed.
Resumo:
Data integration systems offer uniform access to a set of autonomous and heterogeneous data sources. One of the main challenges in data integration is reconciling semantic differences among data sources. Approaches that been used to solve this problem can be categorized as schema-based and attribute-based. Schema-based approaches use schema information to identify the semantic similarity in data; furthermore, they focus on reconciling types before reconciling attributes. In contrast, attribute-based approaches use statistical and structural information of attributes to identify the semantic similarity of data in different sources. This research examines an approach to semantic reconciliation based on integrating properties expressed at different levels of abstraction or granularity using the concept of property precedence. Property precedence reconciles the meaning of attributes by identifying similarities between attributes based on what these attributes represent in the real world. In order to use property precedence for semantic integration, we need to identify the precedence of attributes within and across data sources. The goal of this research is to develop and evaluate a method and algorithms that will identify precedence relations among attributes and build property precedence graph (PPG) that can be used to support integration.