21 resultados para Euclidean distance model,

em Consorci de Serveis Universitaris de Catalunya (CSUC), Spain


Relevância:

90.00% 90.00%

Publicador:

Resumo:

Quantitative or algorithmic trading is the automatization of investments decisions obeying a fixed or dynamic sets of rules to determine trading orders. It has increasingly made its way up to 70% of the trading volume of one of the biggest financial markets such as the New York Stock Exchange (NYSE). However, there is not a signi cant amount of academic literature devoted to it due to the private nature of investment banks and hedge funds. This projects aims to review the literature and discuss the models available in a subject that publications are scarce and infrequently. We review the basic and fundamental mathematical concepts needed for modeling financial markets such as: stochastic processes, stochastic integration and basic models for prices and spreads dynamics necessary for building quantitative strategies. We also contrast these models with real market data with minutely sampling frequency from the Dow Jones Industrial Average (DJIA). Quantitative strategies try to exploit two types of behavior: trend following or mean reversion. The former is grouped in the so-called technical models and the later in the so-called pairs trading. Technical models have been discarded by financial theoreticians but we show that they can be properly cast into a well defined scientific predictor if the signal generated by them pass the test of being a Markov time. That is, we can tell if the signal has occurred or not by examining the information up to the current time; or more technically, if the event is F_t-measurable. On the other hand the concept of pairs trading or market neutral strategy is fairly simple. However it can be cast in a variety of mathematical models ranging from a method based on a simple euclidean distance, in a co-integration framework or involving stochastic differential equations such as the well-known Ornstein-Uhlenbeck mean reversal ODE and its variations. A model for forecasting any economic or financial magnitude could be properly defined with scientific rigor but it could also lack of any economical value and be considered useless from a practical point of view. This is why this project could not be complete without a backtesting of the mentioned strategies. Conducting a useful and realistic backtesting is by no means a trivial exercise since the \laws" that govern financial markets are constantly evolving in time. This is the reason because we make emphasis in the calibration process of the strategies' parameters to adapt the given market conditions. We find out that the parameters from technical models are more volatile than their counterpart form market neutral strategies and calibration must be done in a high-frequency sampling manner to constantly track the currently market situation. As a whole, the goal of this project is to provide an overview of a quantitative approach to investment reviewing basic strategies and illustrating them by means of a back-testing with real financial market data. The sources of the data used in this project are Bloomberg for intraday time series and Yahoo! for daily prices. All numeric computations and graphics used and shown in this project were implemented in MATLAB^R scratch from scratch as a part of this thesis. No other mathematical or statistical software was used.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

We construct a weighted Euclidean distance that approximates any distance or dissimilarity measure between individuals that is based on a rectangular cases-by-variables data matrix. In contrast to regular multidimensional scaling methods for dissimilarity data, the method leads to biplots of individuals and variables while preserving all the good properties of dimension-reduction methods that are based on the singular-value decomposition. The main benefits are the decomposition of variance into components along principal axes, which provide the numerical diagnostics known as contributions, and the estimation of nonnegative weights for each variable. The idea is inspired by the distance functions used in correspondence analysis and in principal component analysis of standardized data, where the normalizations inherent in the distances can be considered as differential weighting of the variables. In weighted Euclidean biplots we allow these weights to be unknown parameters, which are estimated from the data to maximize the fit to the chosen distances or dissimilarities. These weights are estimated using a majorization algorithm. Once this extra weight-estimation step is accomplished, the procedure follows the classical path in decomposing the matrix and displaying its rows and columns in biplots.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Report for the scientific sojourn at the University of Bern, Swiss, from Mars until June 2008. Writer identification consists in determining the writer of a piece of handwriting from a set of writers. Even though an important amount of compositions contains handwritten text in the music scores, the aim of the work is to use only music notation to determine the author. It’s been developed two approaches for writer identification in old handwritten music scores. The methods proposed extract features from every music line, and also features from a texture image of music symbols. First of all, the music sheet is first preprocessed for obtaining a binarized music score without the staff lines. The classification is performed using a k-NN classifier based on Euclidean distance. The proposed method has been tested on a database of old music scores from the 17th to 19th centuries, achieving encouraging identification rates.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Workgroup diversity can be conceptualized as variety, separation, or disparity. Thus, the proper operationalization of diversity depends on how a diversity dimension has been defined. Analytically, the minimal diversity must be obtained when there are no differences on an attribute among the members of a group, however maximal diversity has a different shape for each conceptualization of diversity. Previous work on diversity indexes indicated maximum values for variety (e.g., Blau"s index and Teachman"s index), separation (e.g., standard deviation and mean Euclidean distance), and disparity (e.g., coefficient of variation and the Gini coefficient of concentration), although these maximum values are not valid for all group characteristics (i.e., group size and group size parity) and attribute scales (i.e., number of categories). We demonstrate analytically appropriate upper boundaries for conditional diversity determined by some specific group characteristics, avoiding the bias related to absolute diversity. This will allow applied researchers to make better interpretations regarding the relationship between group diversity and group outcomes.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Variational steepest descent approximation schemes for the modified Patlak-Keller-Segel equation with a logarithmic interaction kernel in any dimension are considered. We prove the convergence of the suitably interpolated in time implicit Euler scheme, defined in terms of the Euclidean Wasserstein distance, associated to this equation for sub-critical masses. As a consequence, we recover the recent result about the global in time existence of weak-solutions to the modified Patlak-Keller-Segel equation for the logarithmic interaction kernel in any dimension in the sub-critical case. Moreover, we show how this method performs numerically in one dimension. In this particular case, this numerical scheme corresponds to a standard implicit Euler method for the pseudo-inverse of the cumulative distribution function. We demonstrate its capabilities to reproduce easily without the need of mesh-refinement the blow-up of solutions for super-critical masses.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

This paper examines competition in the standard one-dimensional Downsian model of two-candidate elections, but where one candidate (A) enjoys an advantage over the other candidate (D). Voters' preferences are Euclidean, but any voter will vote for candidate A over candidate D unless D is closer to her ideal point by some fixed distance \delta. The location of the median voter's ideal point is uncertain, and its distribution is commonly known by both candidates. The candidates simultaneously choose locations to maximize the probability of victory. Pure strategy equilibria often fails to exist in this model, except under special conditions about \delta and the distribution of the median ideal point. We solve for the essentially unique symmetric mixed equilibrium, show that candidate A adopts more moderate policies than candidate D, and obtain some comparative statics results about the probability of victory and the expected distance between the two candidates' policies.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

We analyze the two-dimensional parabolic-elliptic Patlak-Keller-Segel model in the whole Euclidean space R2. Under the hypotheses of integrable initial data with finite second moment and entropy, we first show local in time existence for any mass of "free-energy solutions", namely weak solutions with some free energy estimates. We also prove that the solution exists as long as the entropy is controlled from above. The main result of the paper is to show the global existence of free-energy solutions with initial data as before for the critical mass 8 Π/Χ. Actually, we prove that solutions blow-up as a delta dirac at the center of mass when t→∞ keeping constant their second moment at any time. Furthermore, all moments larger than 2 blow-up as t→∞ if initially bounded.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This paper analyzes both theoretically and empirically the relationship between distance and frequency of scheduled transportation services. We study the interaction between a monopoly firm providing high-speed scheduled service and personal trans- portation (i.e., car). Most interestingly, the carrier chooses to increase frequency of service on longer routes when competing with personal transportation because provid- ing a higher frequency (at extra cost) it can also charge higher fares that can boost its profits. However, when driving is not a relevant option, frequency of service de- creases for longer flights consistently with prior studies. An empirical application of our analysis to the European airline industry con?rms the predictions of our theoretical model.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This paper presents a theoretical and empirical analysis of the relationship be- tween frequency of scheduled transportation services and their substitutability with personal transportation (using distance as a proxy). We study the interaction between a monopoly firm providing a high-speed scheduled service and private transportation (i.e., car). Interestingly, the carrier chooses to increase the frequency of service on longer routes when competing with personal transportation because by providing higher frequency (at extra cost) it can also charge higher fares which can boost its profits. However, in line with the results of earlier studies, frequency decreases for longer flights when driving is not a viable option. An empirical application of our analysis to the European airline industry confirms the predictions of our theoretical model. Keywords: short-haul routes; long-haul routes; flight frequency; distance JEL Classification Numbers: L13; L2; L93

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This paper introduces local distance-based generalized linear models. These models extend (weighted) distance-based linear models firstly with the generalized linear model concept, then by localizing. Distances between individuals are the only predictor information needed to fit these models. Therefore they are applicable to mixed (qualitative and quantitative) explanatory variables or when the regressor is of functional type. Models can be fitted and analysed with the R package dbstats, which implements several distancebased prediction methods.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In this paper we present a novel structure from motion (SfM) approach able to infer 3D deformable models from uncalibrated stereo images. Using a stereo setup dramatically improves the 3D model estimation when the observed 3D shape is mostly deforming without undergoing strong rigid motion. Our approach first calibrates the stereo system automatically and then computes a single metric rigid structure for each frame. Afterwards, these 3D shapes are aligned to a reference view using a RANSAC method in order to compute the mean shape of the object and to select the subset of points on the object which have remained rigid throughout the sequence without deforming. The selected rigid points are then used to compute frame-wise shape registration and to extract the motion parameters robustly from frame to frame. Finally, all this information is used in a global optimization stage with bundle adjustment which allows to refine the frame-wise initial solution and also to recover the non-rigid 3D model. We show results on synthetic and real data that prove the performance of the proposed method even when there is no rigid motion in the original sequence

Relevância:

30.00% 30.00%

Publicador:

Resumo:

One of the criticisms leveled at the model of dispersed city found all over the world is its unarticulated, random, and undifferentiated nature. To check this idea in the Barcelona Metropolitan Region, we estimated the impact of the urban spatial structure (CBD, subcenters and transportation infrastructures) over the population density and commuting distance. The results are unfavorable to the hypothesis of the increasing destructuring of cities given that the explanatory capacity of both functions improves over time, both when other control variables are not included and when they are included.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Des de l’any 2000 es té constància de la presencia del llop a Catalunya. Des de llavors, com a mínim 14 llops diferents han entrat i sortit del territori català, encara que cap d’ells s’ha assentat de manera permanent. L’estudi analitza l’entorn català utilitzant GIS, creant un model d’adequació de l’hàbitat tenint en compte les següents variables: la distància a la carretera més propera, la biomassa disponible a la zona, l’altitud i el tipus i tant per cent de recobriment. El model es basa en la informació obtinguda mitjançant la consulta a experts tant del llop com del territori català, així com en una recerca bibliogràfica sobre l’adequació de l’hàbitat del llop. L’enquesta que es dirigí als experts té en compte els valors que cada variable pot prendre dins l’àrea d’estudi, estableix rangs dels valors de cada variable i pregunta als experts com cada rang pot afectar a l’adequació de l’hàbitat pel llop. Els resultats mostren com bona part de la zona Nord de Catalunya té unes condicions adequades perquè el llop pugui arribar a reproduir-s’hi. Es desenvolupa també una anàlisi dels possibles punts de conflicte humà-llop i una superposició dels espais protegits amb les zones més adequades per l’establiment del llop.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Comparison of donor-acceptor electronic couplings calculated within two-state and three-state models suggests that the two-state treatment can provide unreliable estimates of Vda because of neglecting the multistate effects. We show that in most cases accurate values of the electronic coupling in a π stack, where donor and acceptor are separated by a bridging unit, can be obtained as Ṽ da = (E2 - E1) μ12 Rda + (2 E3 - E1 - E2) 2 μ13 μ23 Rda2, where E1, E2, and E3 are adiabatic energies of the ground, charge-transfer, and bridge states, respectively, μij is the transition dipole moments between the states i and j, and Rda is the distance between the planes of donor and acceptor. In this expression based on the generalized Mulliken-Hush approach, the first term corresponds to the coupling derived within a two-state model, whereas the second term is the superexchange correction accounting for the bridge effect. The formula is extended to bridges consisting of several subunits. The influence of the donor-acceptor energy mismatch on the excess charge distribution, adiabatic dipole and transition moments, and electronic couplings is examined. A diagnostic is developed to determine whether the two-state approach can be applied. Based on numerical results, we showed that the superexchange correction considerably improves estimates of the donor-acceptor coupling derived within a two-state approach. In most cases when the two-state scheme fails, the formula gives reliable results which are in good agreement (within 5%) with the data of the three-state generalized Mulliken-Hush model

Relevância:

30.00% 30.00%

Publicador:

Resumo:

A new parametric minimum distance time-domain estimator for ARFIMA processes is introduced in this paper. The proposed estimator minimizes the sum of squared correlations of residuals obtained after filtering a series through ARFIMA parameters. The estimator iseasy to compute and is consistent and asymptotically normally distributed for fractionallyintegrated (FI) processes with an integration order d strictly greater than -0.75. Therefore, it can be applied to both stationary and non-stationary processes. Deterministic components are also allowed in the DGP. Furthermore, as a by-product, the estimation procedure provides an immediate check on the adequacy of the specified model. This is so because the criterion function, when evaluated at the estimated values, coincides with the Box-Pierce goodness of fit statistic. Empirical applications and Monte-Carlo simulations supporting the analytical results and showing the good performance of the estimator in finite samples are also provided.