48 resultados para in-domain data requirement
Resumo:
Several unit root tests in panel data have recently been proposed. The test developed by Harris and Tzavalis (1999 JoE) performs particularly well when the time dimension is moderate in relation to the cross-section dimension. However, in common with the traditional tests designed for the unidimensional case, it was found to perform poorly when there is a structural break in the time series under the alternative. Here we derive the asymptotic distribution of the test allowing for a shift in the mean, and assess the small sample performance. We apply this new test to show how the hypothesis of (perfect) hysteresis in Spanish unemployment is rejected in favour of the alternative of the natural unemployment rate, when the possibility of a change in the latter is considered.
Resumo:
The present study explores the statistical properties of a randomization test based on the random assignment of the intervention point in a two-phase (AB) single-case design. The focus is on randomization distributions constructed with the values of the test statistic for all possible random assignments and used to obtain p-values. The shape of those distributions is investigated for each specific data division defined by the moment in which the intervention is introduced. Another aim of the study consisted in testing the detection of inexistent effects (i.e., production of false alarms) in autocorrelated data series, in which the assumption of exchangeability between observations may be untenable. In this way, it was possible to compare nominal and empirical Type I error rates in order to obtain evidence on the statistical validity of the randomization test for each individual data division. The results suggest that when either of the two phases has considerably less measurement times, Type I errors may be too probable and, hence, the decision making process to be carried out by applied researchers may be jeopardized.
Resumo:
In recent years, new analytical tools have allowed researchers to extract historical information contained in molecular data, which has fundamentally transformed our understanding of processes ruling biological invasions. However, the use of these new analytical tools has been largely restricted to studies of terrestrial organisms despite the growing recognition that the sea contains ecosystems that are amongst the most heavily affected by biological invasions, and that marine invasion histories are often remarkably complex. Here, we studied the routes of invasion and colonisation histories of an invasive marine invertebrate Microcosmus squamiger (Ascidiacea) using microsatellite loci, mitochondrial DNA sequence data and 11 worldwide populations. Discriminant analysis of principal components, clustering methods and approximate Bayesian computation (ABC) methods showed that the most likely source of the introduced populations was a single admixture event that involved populations from two genetically differentiated ancestral regions - the western and eastern coasts of Australia. The ABC analyses revealed that colonisation of the introduced range of M. squamiger consisted of a series of non-independent introductions along the coastlines of Africa, North America and Europe. Furthermore, we inferred that the sequence of colonisation across continents was in line with historical taxonomic records - first the Mediterranean Sea and South Africa from an unsampled ancestral population, followed by sequential introductions in California and, more recently, the NE Atlantic Ocean. We revealed the most likely invasion history for world populations of M. squamiger, which is broadly characterized by the presence of multiple ancestral sources and non-independent introductions within the introduced range. The results presented here illustrate the complexity of marine invasion routes and identify a cause-effect relationship between human-mediated transport and the success of widespread marine non-indigenous species, which benefit from stepping-stone invasions and admixture processes involving different sources for the spread and expansion of their range.
Resumo:
The purpose of this paper is to study the determinants of equilibrium in the market for daily funds. We use the EONIA panel database which includes daily information on the lending rates applied by contributing commercial banks. The data clearly shows an increase in both the time series volatility and the cross section dispersion of rates towards the end of the reserve maintenance period. These increases are highly correlated. With respect to quantities, we find that the volume of trade as well as the use of the standing facilities are also larger at the end of the maintenance period. Our theoretical model shows how the operational framework of monetary policy causes a reduction in the elasticity of the supply of funds by banks throughout the reserve maintenance period. This reduction in the elasticity together with market segmentation and heterogeneity are able to generate distributions for the interest rates and quantities traded with the same properties as in the data.
Resumo:
In an earlier investigation (Burger et al., 2000) five sediment cores near the RodriguesTriple Junction in the Indian Ocean were studied applying classical statistical methods(fuzzy c-means clustering, linear mixing model, principal component analysis) for theextraction of endmembers and evaluating the spatial and temporal variation ofgeochemical signals. Three main factors of sedimentation were expected by the marinegeologists: a volcano-genetic, a hydro-hydrothermal and an ultra-basic factor. Thedisplay of fuzzy membership values and/or factor scores versus depth providedconsistent results for two factors only; the ultra-basic component could not beidentified. The reason for this may be that only traditional statistical methods wereapplied, i.e. the untransformed components were used and the cosine-theta coefficient assimilarity measure.During the last decade considerable progress in compositional data analysis was madeand many case studies were published using new tools for exploratory analysis of thesedata. Therefore it makes sense to check if the application of suitable data transformations,reduction of the D-part simplex to two or three factors and visualinterpretation of the factor scores would lead to a revision of earlier results and toanswers to open questions . In this paper we follow the lines of a paper of R. Tolosana-Delgado et al. (2005) starting with a problem-oriented interpretation of the biplotscattergram, extracting compositional factors, ilr-transformation of the components andvisualization of the factor scores in a spatial context: The compositional factors will beplotted versus depth (time) of the core samples in order to facilitate the identification ofthe expected sources of the sedimentary process.Kew words: compositional data analysis, biplot, deep sea sediments
Resumo:
The application of compositional data analysis through log ratio trans-formations corresponds to a multinomial logit model for the shares themselves.This model is characterized by the property of Independence of Irrelevant Alter-natives (IIA). IIA states that the odds ratio in this case the ratio of shares is invariant to the addition or deletion of outcomes to the problem. It is exactlythis invariance of the ratio that underlies the commonly used zero replacementprocedure in compositional data analysis. In this paper we investigate using thenested logit model that does not embody IIA and an associated zero replacementprocedure and compare its performance with that of the more usual approach ofusing the multinomial logit model. Our comparisons exploit a data set that com-bines voting data by electoral division with corresponding census data for eachdivision for the 2001 Federal election in Australia
Resumo:
R from http://www.r-project.org/ is ‘GNU S’ – a language and environment for statistical computingand graphics. The environment in which many classical and modern statistical techniques havebeen implemented, but many are supplied as packages. There are 8 standard packages and many moreare available through the cran family of Internet sites http://cran.r-project.org .We started to develop a library of functions in R to support the analysis of mixtures and our goal isa MixeR package for compositional data analysis that provides support foroperations on compositions: perturbation and power multiplication, subcomposition with or withoutresiduals, centering of the data, computing Aitchison’s, Euclidean, Bhattacharyya distances,compositional Kullback-Leibler divergence etc.graphical presentation of compositions in ternary diagrams and tetrahedrons with additional features:barycenter, geometric mean of the data set, the percentiles lines, marking and coloring ofsubsets of the data set, theirs geometric means, notation of individual data in the set . . .dealing with zeros and missing values in compositional data sets with R procedures for simpleand multiplicative replacement strategy,the time series analysis of compositional data.We’ll present the current status of MixeR development and illustrate its use on selected data sets
Resumo:
As stated in Aitchison (1986), a proper study of relative variation in a compositional data set should be based on logratios, and dealing with logratios excludes dealing with zeros. Nevertheless, it is clear that zero observations might be present in real data sets, either because the corresponding part is completelyabsent –essential zeros– or because it is below detection limit –rounded zeros. Because the second kind of zeros is usually understood as “a trace too small to measure”, it seems reasonable to replace them by a suitable small value, and this has been the traditional approach. As stated, e.g. by Tauber (1999) and byMartín-Fernández, Barceló-Vidal, and Pawlowsky-Glahn (2000), the principal problem in compositional data analysis is related to rounded zeros. One should be careful to use a replacement strategy that does not seriously distort the general structure of the data. In particular, the covariance structure of the involvedparts –and thus the metric properties– should be preserved, as otherwise further analysis on subpopulations could be misleading. Following this point of view, a non-parametric imputation method isintroduced in Martín-Fernández, Barceló-Vidal, and Pawlowsky-Glahn (2000). This method is analyzed in depth by Martín-Fernández, Barceló-Vidal, and Pawlowsky-Glahn (2003) where it is shown that thetheoretical drawbacks of the additive zero replacement method proposed in Aitchison (1986) can be overcome using a new multiplicative approach on the non-zero parts of a composition. The new approachhas reasonable properties from a compositional point of view. In particular, it is “natural” in the sense thatit recovers the “true” composition if replacement values are identical to the missing values, and it is coherent with the basic operations on the simplex. This coherence implies that the covariance structure of subcompositions with no zeros is preserved. As a generalization of the multiplicative replacement, in thesame paper a substitution method for missing values on compositional data sets is introduced
Resumo:
We study the relationship between openness and payment system development. In particular, we analyze how the existence of technology diffusion from a more developed country fosters a transformation of payment choice in a less developed country. We apply our analysis to Mexico. Economic growth in Mexico was not high enough to cause a transformation of payment choice observed in the data after 2001. We argue that the switch towards electronic payments can be attributed to openness and related payment technology spillovers from the US in the context of NAFTA.
Resumo:
Globalization involves several facility location problems that need to be handled at large scale. Location Allocation (LA) is a combinatorial problem in which the distance among points in the data space matter. Precisely, taking advantage of the distance property of the domain we exploit the capability of clustering techniques to partition the data space in order to convert an initial large LA problem into several simpler LA problems. Particularly, our motivation problem involves a huge geographical area that can be partitioned under overall conditions. We present different types of clustering techniques and then we perform a cluster analysis over our dataset in order to partition it. After that, we solve the LA problem applying simulated annealing algorithm to the clustered and non-clustered data in order to work out how profitable is the clustering and which of the presented methods is the most suitable
Resumo:
This paper reconsiders the empirical evidence on the asymmetricoutput effects of monetary policy. Asymmetric effects is a common feature ofmany theoretical models, and there are many different versions of suchasymmetries. We concentrate on the distinctions between positive andnegative money-supply changes, big and small changes in money-supply, andpossible combinations of the two asymmetries. Earlier research has foundempirical evidence in favor of the former of these in US data. Using M1 asthe monetary variable we find evidence in favor of neutrality of big shocksand non-neutrality of small shocks. The results may, however, be affected bystructual instability of M1 demand. Thus, we substitute M1 with the federalfunds rate. In these data we find that only small negative shocks affectreal aggregate activity. The results are interpreted in terms of menu-costmodels.
Resumo:
Much like cognitive abilities, emotional skills can have major effects on performance and economic outcomes. This paper studies the behavior of professionalsubjects involved in a dynamic competition in their own natural environment. Thesetting is a penalty shoot-out in soccer where two teams compete in a tournamentframework taking turns in a sequence of five penalty kicks each. As the kicking order is determined by the random outcome of a coin flip, the treatment and control groups are determined via explicit randomization. Therefore, absent any psychological effects, both teams should have the same probability of winning regardless of the kicking order. Yet, we find a systematic first-kicker advantage. Using data on 2,731 penalty kicks from 262 shoot-outs for a three decade period, we find that teams kicking first win the penalty shoot-out 60.5% of the time. A dynamic panel data analysis shows that the psychological mechanism underlying this result arises from the asymmetry in the partial score. As most kicks are scored, kicking first typically means having the opportunity to lead in the partial score, whereas kicking second typically means lagging in the score and having the opportunity to, at most, get even. Having a worse prospect than the opponent hinders subjects' performance.Further, we also find that professionals are self-aware of their own psychological effects. When a recent change in regulations gives winners of the coin toss the chance to choose the kicking order, they rationally react to it by systematically choosing to kick first. A survey of professional players reveals that when asked to explain why they prefer to kick first, they precisely identify the psychological mechanism for which we find empirical support in the data: they want to lead in the score inorder to put pressure on the opponent.
Resumo:
We study the effect of regional expenditure and revenue shocks on price differentials for47 US states and 9 EU countries. We identify shocks using sign restrictions on the dynamicsof deficits and output and construct two estimates for structural price differentials dynamics which optimally weight the information contained in the data for all units. Fiscal shocks explain between 14 and 23 percent of the variability of price differentials both in the US and in the EU. On average, expansionary fiscal disturbances produce positive price differential responses while distortionary balance budget shocks produce negative price differential responses. In a number of units, price differential responses to expansionary fiscal shocks are negative. Spillovers and labor supply effects partially explain this pattern while geographical, political, and economic indicators do not.
Resumo:
In this paper we study, having as theoretical reference the economic model of crime (Becker, 1968; Ehrlich, 1973), which are the socioeconomic and demographic determinants of crime in Spain paying attention on the role of provincial peculiarities. We estimate a crime equation using a panel dataset of Spanish provinces (NUTS3) for the period 1993 to 1999 employing the GMMsystem estimator. Empirical results suggest that lagged crime rate and clear-up rate are correlated to all typologies of crime rate considered. Property crimes are better explained by socioeconomic variables (GDP per capita, GDP growth rate and percentage of population with high school and university degree), while demographic factors reveal important and significant influences, in particular for crimes against the person. These results are obtained using an instrumental variable approach that takes advantage of the dynamic properties of our dataset to control for both measurement errors in crime data and joint endogeneity of the explanatory variables
Resumo:
In this paper we study, having as theoretical reference the economic model of crime (Becker, 1968; Ehrlich, 1973), which are the socioeconomic and demographic determinants of crime in Spain paying attention on the role of provincial peculiarities. We estimate a crime equation using a panel dataset of Spanish provinces (NUTS3) for the period 1993 to 1999 employing the GMMsystem estimator. Empirical results suggest that lagged crime rate and clear-up rate are correlated to all typologies of crime rate considered. Property crimes are better explained by socioeconomic variables (GDP per capita, GDP growth rate and percentage of population with high school and university degree), while demographic factors reveal important and significant influences, in particular for crimes against the person. These results are obtained using an instrumental variable approach that takes advantage of the dynamic properties of our dataset to control for both measurement errors in crime data and joint endogeneity of the explanatory variables