975 resultados para common stochastic component
Resumo:
The computational architecture that enables the flexible coupling between otherwise independent eye and hand effector systems is not understood. By using a drift diffusion framework, in which variability of the reaction time (RT) distribution scales with mean RT, we tested the ability of a common stochastic accumulator to explain eye-hand coordination. Using a combination of behavior, computational modeling and electromyography, we show how a single stochastic accumulator to threshold, followed by noisy effector-dependent delays, explains eye-hand RT distributions and their correlation, while an alternate independent, interactive eye and hand accumulator model does not. Interestingly, the common accumulator model did not explain the RT distributions of the same subjects when they made eye and hand movements in isolation. Taken together, these data suggest that a dedicated circuit underlies coordinated eye-hand planning.
Resumo:
Many studies of reaching and pointing have shown significant spatial and temporal correlations between eye and hand movements. Nevertheless, it remains unclear whether these correlations are incidental, arising from common inputs (independent model); whether these correlations represent an interaction between otherwise independent eye and hand systems (interactive model); or whether these correlations arise from a single dedicated eye-hand system (common command model). Subjects were instructed to redirect gaze and pointing movements in a double-step task in an attempt to decouple eye-hand movements and causally distinguish between the three architectures. We used a drift-diffusion framework in the context of a race model, which has been previously used to explain redirect behavior for eye and hand movements separately, to predict the pattern of eye-hand decoupling. We found that the common command architecture could best explain the observed frequency of different eye and hand response patterns to the target step. A common stochastic accumulator for eye-hand coordination also predicts comparable variances, despite significant difference in the means of the eye and hand reaction time (RT) distributions, which we tested. Consistent with this prediction, we observed that the variances of the eye and hand RTs were similar, despite much larger hand RTs (similar to 90 ms). Moreover, changes in mean eye RTs, which also increased eye RT variance, produced a similar increase in mean and variance of the associated hand RT. Taken together, these data suggest that a dedicated circuit underlies coordinated eye-hand planning.
Resumo:
In this article, we illustrate experimentally an important consequence of the stochastic component in choice behaviour which has not been acknowledged so far. Namely, its potential to produce ‘regression to the mean’ (RTM) effects. We employ a novel approach to individual choice under risk, based on repeated multiple-lottery choices (i.e. choices among many lotteries), to show how the high degree of stochastic variability present in individual decisions can distort crucially certain results through RTM effects. We demonstrate the point in the context of a social comparison experiment.
Resumo:
This dissertation develops a strategic management accounting perspective of inventory routing. The thesis studies the drivers of cost efficiency gains by identifying the role of the underlying cost structure, demand, information sharing, forecasting accuracy, service levels, vehicle fleet, planning horizon and other strategic factors as well as the interaction effects among these factors with respect to performance outcomes. The task is to enhance the knowledge of the strategic situations that favor the implementation of inventory routing systems, understanding cause-and-effect relationships, linkages and gaining a holistic view of the value proposition of inventory routing. The thesis applies an exploratory case study design, which is based on normative quantitative empirical research using optimization, simulation and factor analysis. Data and results are drawn from a real world application to cash supply chains. The first research paper shows that performance gains require a common cost component and cannot be explained by simple linear or affine cost structures. Inventory management and distribution decisions become separable in the absence of a set-dependent cost structure, and neither economies of scope nor coordination problems are present in this case. The second research paper analyzes whether information sharing improves the overall forecasting accuracy. Analysis suggests that the potential for information sharing is limited to coordination of replenishments and that central information do not yield more accurate forecasts based on joint forecasting. The third research paper develops a novel formulation of the stochastic inventory routing model that accounts for minimal service levels and forecasting accuracy. The developed model allows studying the interaction of minimal service levels and forecasting accuracy with the underlying cost structure in inventory routing. Interestingly, results show that the factors minimal service level and forecasting accuracy are not statistically significant, and subsequently not relevant for the strategic decision problem to introduce inventory routing, or in other words, to effectively internalize inventory management and distribution decisions at the supplier. Consequently the main contribution of this thesis is the result that cost benefits of inventory routing are derived from the joint decision model that accounts for the underlying set-dependent cost structure rather than the level of information sharing. This result suggests that the value of information sharing of demand and inventory data is likely to be overstated in prior literature. In other words, cost benefits of inventory routing are primarily determined by the cost structure (i.e. level of fixed costs and transportation costs) rather than the level of information sharing, joint forecasting, forecasting accuracy or service levels.
Resumo:
The optimal control of problems that are constrained by partial differential equations with uncertainties and with uncertain controls is addressed. The Lagrangian that defines the problem is postulated in terms of stochastic functions, with the control function possibly decomposed into an unknown deterministic component and a known zero-mean stochastic component. The extra freedom provided by the stochastic dimension in defining cost functionals is explored, demonstrating the scope for controlling statistical aspects of the system response. One-shot stochastic finite element methods are used to find approximate solutions to control problems. It is shown that applying the stochastic collocation finite element method to the formulated problem leads to a coupling between stochastic collocation points when a deterministic optimal control is considered or when moments are included in the cost functional, thereby forgoing the primary advantage of the collocation method over the stochastic Galerkin method for the considered problem. The application of the presented methods is demonstrated through a number of numerical examples. The presented framework is sufficiently general to also consider a class of inverse problems, and numerical examples of this type are also presented. © 2011 Elsevier B.V.
Resumo:
The population of naive T cells in the periphery is best described by determining both its T cell receptor diversity, or number of clonotypes, and the sizes of its clonal subsets. In this paper, we make use of a previously introduced mathematical model of naive T cell homeostasis, to study the fate and potential of naive T cell clonotypes in the periphery. This is achieved by the introduction of several new stochastic descriptors for a given naive T cell clonotype, such as its maximum clonal size, the time to reach this maximum, the number of proliferation events required to reach this maximum, the rate of contraction of the clonotype during its way to extinction, as well as the time to a given number of proliferation events. Our results show that two fates can be identified for the dynamics of the clonotype: extinction in the short-term if the clonotype experiences too hostile a peripheral environment, or establishment in the periphery in the long-term. In this second case the probability mass function for the maximum clonal size is bimodal, with one mode near one and the other mode far away from it. Our model also indicates that the fate of a recent thymic emigrant (RTE) during its journey in the periphery has a clear stochastic component, where the probability of extinction cannot be neglected, even in a friendly but competitive environment. On the other hand, a greater deterministic behaviour can be expected in the potential size of the clonotype seeded by the RTE in the long-term, once it escapes extinction.
Resumo:
Forecasts of volatility and correlation are important inputs into many practical financial problems. Broadly speaking, there are two ways of generating forecasts of these variables. Firstly, time-series models apply a statistical weighting scheme to historical measurements of the variable of interest. The alternative methodology extracts forecasts from the market traded value of option contracts. An efficient options market should be able to produce superior forecasts as it utilises a larger information set of not only historical information but also the market equilibrium expectation of options market participants. While much research has been conducted into the relative merits of these approaches, this thesis extends the literature along several lines through three empirical studies. Firstly, it is demonstrated that there exist statistically significant benefits to taking the volatility risk premium into account for the implied volatility for the purposes of univariate volatility forecasting. Secondly, high-frequency option implied measures are shown to lead to superior forecasts of the intraday stochastic component of intraday volatility and that these then lead on to superior forecasts of intraday total volatility. Finally, the use of realised and option implied measures of equicorrelation are shown to dominate measures based on daily returns.
Resumo:
Brachyuran larvae are the most common zooplankton component of the Manora Channel, Karachi, Pakistan. The identification of these larvae would assist in the assessment of brachyuran species and provide information on percentage composition, occurrence, abundance and breeding habits of the brachyuran species in the Manora Channel area. However plankton caught larvae is not easily identified. An accurate identification of such material is only possible by the comparison with larvae reared under laboratory conditions and documented with illustrations. The identifications for this present study were based on the works of Gurney (1938); Atkins (1954); Chhapgar (1955); Raja Bai (1960); Hashmi (1969, 1970a, b); Baba and Miyata (1971); Kakati and Sankolli (1975); Rice (1975); Kakati (1977); Lim and Tan (1981); Yatsuzuka and Sakai (1984); Fielder, et al (1984); Amir (1989, M. Phil thesis unpublished); Ingle (1992); Siddiqui and Tirmizi (1992); Tirmizi et al (1993); Bano (1999, M. Phil thesis unpublished); Ghory and Siddiqui (2001); Ghory (2002, M. Phil thesis unpublished); Ghory and Siddiqui (2002).
Resumo:
The Niagara Grape and Wine Community (NGWC) is an industry that has undergone rapid change and expansion as a result of changes in governmental regulations and consumer preferences. As a result of these changes, the demands of the wine industry workforce have changed to reflect the need to implement new strategies and practices to remain viable and competitive. The influx of people into the community with little or no prior practical experience in grape growing (viticulture) or winemaking (oenology) has created a need for additional training and learning opportunities to meet workforce needs. This case study investigated the learning needs of the members of this community and how these needs are currently being met. The barriers to, and the opportunities for, members acquiring new knowledge and developing skills were also explored. Participants were those involved in all levels of the industry and sectors (viticulture, processing, and retail), and their views on needs and suggestions for programs of study were collected. Through cross analyses of sectors, areas of common and unique interest were identified as well as formats for delivery. A common fundamental component was identified by all sectors - any program must have a significant applied component or demonstration of proficiency and should utilize members as peer instructors, mentors, and collaborators to generate a larger shared collective of knowledge. Through the review of learning organizations, learning communities, communities of practices, and learning networks, the principles for the development of a Grape and Wine Learning Network to meet the learning needs of the NGWC outside of formal institutional or academic programs were developed. The roles and actions of members to make such a network successful are suggested.
Resumo:
El presente trabajo desarrolla un modelo macroeconómico de equilibrio general dinámico y estocástico (DSGE), con el fin de analizar los efectos macroeconómicos que se derivan de simular un choque positivo al componente estocástico de la productividad del sector minero-energético. Este hecho genera un aumento generalizado de los salarios en el sector formal y en el recaudo tributario, incrementando el consumo total de los miembros del hogar. Esto genera un incremento del precio de los bienes no transables relativo al precio de los bienes transables, disminuyendo la tasa de cambio real (apreciación) y provocando un desplazamiento de los recursos productivos, desde el sector transable (manufacturero) al no-transable, seguido de un aumento en el PIB y empleo formal de la economía. Esto hace que el sector formal agregado absorba trabajadores desde el sector informal a través del subsector formal no-transable, lo que disminuye el PIB informal. En consecuencia, el consumo neto de los miembros informales disminuye, lo que incentiva a que algunos miembros del hogar no se empleen en el sector informal y prefieran quedarse desempleados. Por lo tanto, el resultado final sobre el mercado laboral es una disminución de los trabajadores informales, de los cuales una parte se encuentra en el sector formal, y la parte restante está en condición de desempleo.
Resumo:
Neste trabalho abordamos a unitização como uma reinterpretação de cartel, partindo do modelo clássico de Green e Porter. A incerteza geológica é representada por um componente estocástico no custo marginal. Caracterizamos o contrato ótimo e, a partir da estática comparativa, avaliamos a eficiência e a viabilidade da cooperação. O preço e o grau da externalidade afetam positivamente o nível de eficiência do contrato ótimo. Mas enquanto preços elevados viabilizam os acordos, o grau de externalidade elevado pode conduzir a equilíbrios ineficientes ou mesmo inviabilizar a produção. O mesmo resultado ocorre com os custos fixos. Adicionalmente, quanto maior for o número de firmas envolvidas no acordo, menor será a chance de existir um contrato mais eficiente que a regra da captura.
Resumo:
Coordenação de Aperfeiçoamento de Pessoal de Nível Superior (CAPES)
Resumo:
Aims The effect Of anthropogenic landscape fragmentation on the genetic diversity and adaptive potential of plant populations is a major issue in conservation biology. However, little is known about the partitioning of genetic diversity in alpine species, which occur in naturally fragmented habitats. Here, we, investigate molecular patterns of three alpine plants (Epilobium fleischeri, Geum reptans and Campanula thyrsoides) across Switzerland and ask whether Spatial isolation has led to high levels of populations differentiation, increasing over distance, and a decrease of within-population variability. We further hypothesize that file contrasting potential for long-distance dispersal (LDD) of Seed in these Species will considerably influence and explain diversity partitioning. Methods For each study species, we Sampled 20-23 individuals from each of 20-32 populations across entire Switzerland. We applied Random Amplified Polymorphic Dimorphism markers to assess genetic diversity within (Nei's expected heterozygosity, H-e; percentage of polymorphic hands, P-P) and among (analysis of molecular variance, Phi(st)) populations and correlated population size and altitude with within-populalion diversity. Spatial patterns of genetic relatedness were investigated using Mantel tests and standardized major axis regression as well as unweighted pair group method with arithmetic mean cluster analyses and Monmonier's algorithm. To avoid known biases, We standardized the numbers of populations, individuals and markers using multiple random reductions. We modelled LDD with a high alpine wind data set using the terminal velocity and height of seed release as key parameters. Additionally, we assessed a number of important life-history traits and factors that potentially influence genetic diversity partitioning (e.g. breeding system, longevity and population size). Important findings For all three species, We found a significant isolation-by-distance relationship but only a moderately high differentiation among populations (Phi(st): 22.7, 48 and 16.8%, for E. fleischeri, G. reptans and C. thyrsoides, respectively). Within-population diversity (H-c: 0.19-0.21, P-p: 62-75%) was not reduced in comparison to known results from lowland species and even small populations with < 50 reproductive individuals contained high levels of genetic diversity. We further found no indication that a high long-distance seed dispersal potential enhances genetic connectivity among populations. Gene flow seems to have a strong stochastic component causing large dissimilarity between population pairs irrespective of the spatial distance. Our results suggest that other life-history traits, especially the breeding System, may play an important role in genetic diversity partitioning. We conclude that spatial isolation in the alpine environment has a strong influence on population relatedness but that a number of factors can considerably influence the strength of this relationship.
Resumo:
Este trabajo aborda el problema de modelizar sistemas din´amicos reales a partir del estudio de sus series temporales, usando una formulaci´on est´andar que pretende ser una abstracci´on universal de los sistemas din´amicos, independientemente de su naturaleza determinista, estoc´astica o h´ıbrida. Se parte de modelizaciones separadas de sistemas deterministas por un lado y estoc´asticos por otro, para converger finalmente en un modelo h´ıbrido que permite estudiar sistemas gen´ericos mixtos, esto es, que presentan una combinaci´on de comportamiento determinista y aleatorio. Este modelo consta de dos componentes, uno determinista consistente en una ecuaci´on en diferencias, obtenida a partir de un estudio de autocorrelaci´on, y otro estoc´astico que modeliza el error cometido por el primero. El componente estoc´astico es un generador universal de distribuciones de probabilidad, basado en un proceso compuesto de variables aleatorias, uniformemente distribuidas en un intervalo variable en el tiempo. Este generador universal es deducido en la tesis a partir de una nueva teor´ıa sobre la oferta y la demanda de un recurso gen´erico. El modelo resultante puede formularse conceptualmente como una entidad con tres elementos fundamentales: un motor generador de din´amica determinista, una fuente interna de ruido generadora de incertidumbre y una exposici´on al entorno que representa las interacciones del sistema real con el mundo exterior. En las aplicaciones estos tres elementos se ajustan en base al hist´orico de las series temporales del sistema din´amico. Una vez ajustados sus componentes, el modelo se comporta de una forma adaptativa tomando como inputs los nuevos valores de las series temporales del sistema y calculando predicciones sobre su comportamiento futuro. Cada predicci´on se presenta como un intervalo dentro del cual cualquier valor es equipro- bable, teniendo probabilidad nula cualquier valor externo al intervalo. De esta forma el modelo computa el comportamiento futuro y su nivel de incertidumbre en base al estado actual del sistema. Se ha aplicado el modelo en esta tesis a sistemas muy diferentes mostrando ser muy flexible para afrontar el estudio de campos de naturaleza dispar. El intercambio de tr´afico telef´onico entre operadores de telefon´ıa, la evoluci´on de mercados financieros y el flujo de informaci´on entre servidores de Internet son estudiados en profundidad en la tesis. Todos estos sistemas son modelizados de forma exitosa con un mismo lenguaje, a pesar de tratarse de sistemas f´ısicos totalmente distintos. El estudio de las redes de telefon´ıa muestra que los patrones de tr´afico telef´onico presentan una fuerte pseudo-periodicidad semanal contaminada con una gran cantidad de ruido, sobre todo en el caso de llamadas internacionales. El estudio de los mercados financieros muestra por su parte que la naturaleza fundamental de ´estos es aleatoria con un rango de comportamiento relativamente acotado. Una parte de la tesis se dedica a explicar algunas de las manifestaciones emp´ıricas m´as importantes en los mercados financieros como son los “fat tails”, “power laws” y “volatility clustering”. Por ´ultimo se demuestra que la comunicaci´on entre servidores de Internet tiene, al igual que los mercados financieros, una componente subyacente totalmente estoc´astica pero de comportamiento bastante “d´ocil”, siendo esta docilidad m´as acusada a medida que aumenta la distancia entre servidores. Dos aspectos son destacables en el modelo, su adaptabilidad y su universalidad. El primero es debido a que, una vez ajustados los par´ametros generales, el modelo se “alimenta” de los valores observables del sistema y es capaz de calcular con ellos comportamientos futuros. A pesar de tener unos par´ametros fijos, la variabilidad en los observables que sirven de input al modelo llevan a una gran riqueza de ouputs posibles. El segundo aspecto se debe a la formulaci´on gen´erica del modelo h´ıbrido y a que sus par´ametros se ajustan en base a manifestaciones externas del sistema en estudio, y no en base a sus caracter´ısticas f´ısicas. Estos factores hacen que el modelo pueda utilizarse en gran variedad de campos. Por ´ultimo, la tesis propone en su parte final otros campos donde se han obtenido ´exitos preliminares muy prometedores como son la modelizaci´on del riesgo financiero, los algoritmos de routing en redes de telecomunicaci´on y el cambio clim´atico. Abstract This work faces the problem of modeling dynamical systems based on the study of its time series, by using a standard language that aims to be an universal abstraction of dynamical systems, irrespective of their deterministic, stochastic or hybrid nature. Deterministic and stochastic models are developed separately to be merged subsequently into a hybrid model, which allows the study of generic systems, that is to say, those having both deterministic and random behavior. This model is a combination of two different components. One of them is deterministic and consisting in an equation in differences derived from an auto-correlation study and the other is stochastic and models the errors made by the deterministic one. The stochastic component is an universal generator of probability distributions based on a process consisting in random variables distributed uniformly within an interval varying in time. This universal generator is derived in the thesis from a new theory of offer and demand for a generic resource. The resulting model can be visualized as an entity with three fundamental elements: an engine generating deterministic dynamics, an internal source of noise generating uncertainty and an exposure to the environment which depicts the interactions between the real system and the external world. In the applications these three elements are adjusted to the history of the time series from the dynamical system. Once its components have been adjusted, the model behaves in an adaptive way by using the new time series values from the system as inputs and calculating predictions about its future behavior. Every prediction is provided as an interval, where any inner value is equally probable while all outer ones have null probability. So, the model computes the future behavior and its level of uncertainty based on the current state of the system. The model is applied to quite different systems in this thesis, showing to be very flexible when facing the study of fields with diverse nature. The exchange of traffic between telephony operators, the evolution of financial markets and the flow of information between servers on the Internet are deeply studied in this thesis. All these systems are successfully modeled by using the same “language”, in spite the fact that they are systems physically radically different. The study of telephony networks shows that the traffic patterns are strongly weekly pseudo-periodic but mixed with a great amount of noise, specially in the case of international calls. It is proved that the underlying nature of financial markets is random with a moderate range of variability. A part of this thesis is devoted to explain some of the most important empirical observations in financial markets, such as “fat tails”, “power laws” and “volatility clustering”. Finally it is proved that the communication between two servers on the Internet has, as in the case of financial markets, an underlaying random dynamics but with a narrow range of variability, being this lack of variability more marked as the distance between servers is increased. Two aspects of the model stand out as being the most important: its adaptability and its universality. The first one is due to the fact that once the general parameters have been adjusted , the model is “fed” on the observable manifestations of the system in order to calculate its future behavior. Despite the fact that the model has fixed parameters the variability in the observable manifestations of the system, which are used as inputs of the model, lead to a great variability in the possible outputs. The second aspect is due to the general “language” used in the formulation of the hybrid model and to the fact that its parameters are adjusted based on external manifestations of the system under study instead of its physical characteristics. These factors made the model suitable to be used in great variety of fields. Lastly, this thesis proposes other fields in which preliminary and promising results have been obtained, such as the modeling of financial risk, the development of routing algorithms for telecommunication networks and the assessment of climate change.
Resumo:
La investigación realizada en este trabajo de tesis se ha centrado en el estudio de la generación, anclaje y desenganche de paredes de dominio magnético en nanohilos de permalloy con defectos controlados. Las últimas tecnologías de nanofabricación han abierto importantes líneas de investigación centradas en el estudio del movimiento de paredes de dominio magnético, gracias a su potencial aplicación en memorias magnéticas del futuro. En el 2004, Stuart Parkin de IBM introdujo un concepto innovador, el dispositivo “Racetrack”, basado en un nanohilo ferromagnético donde los dominios de imanación representan los "bits" de información. La frontera entre dominios, ie pared magnética, se moverían en una situación ideal por medio de transferencia de espín de una corriente polarizada. Se anclan en determinadas posiciones gracias a pequeños defectos o constricciones de tamaño nanométrico fabricados por litografía electrónica. El éxito de esta idea se basa en la generación, anclaje y desenganche de las paredes de dominio de forma controlada y repetitiva, tanto para la lectura como para la escritura de los bits de información. Slonczewski en 1994 muestra que la corriente polarizada de espín puede transferir momento magnético a la imanación local y así mover paredes por transferencia de espín y no por el campo creado por la corriente. Desde entonces muchos grupos de investigación de todo el mundo trabajan en optimizar las condiciones de transferencia de espín para mover paredes de dominio. La fracción de electrones polarizados que viaja en un hilo ferromagnético es considerablemente pequeña, así hoy por hoy la corriente necesaria para mover una pared magnética por transferencia de espín es superior a 1 107 A/cm2. Una densidad de corriente tan elevada no sólo tiene como consecuencia una importante degradación del dispositivo sino también se observan importantes efectos relacionados con el calentamiento por efecto Joule inducido por la corriente. Otro de los problemas científico - tecnológicos a resolver es la diversidad de paredes de dominio magnético ancladas en el defecto. Los diferentes tipos de pared anclados en el defecto, su quiralidad o el campo o corriente necesarios para desenganchar la pared pueden variar dependiendo si el defecto posee dimensiones ligeramente diferentes o si la pared se ancla con un método distinto. Además, existe una componente estocástica presente tanto en la nucleación como en el proceso de anclaje y desenganche que por un lado puede ser debido a la naturaleza de la pared que viaja por el hilo a una determinada temperatura distinta de cero, así como a defectos inevitables en el proceso de fabricación. Esto constituye un gran inconveniente dado que según el tipo de pared es necesario aplicar distintos valores de corriente y/o campo para desenganchar la pared del defecto. Como se menciona anteriormente, para realizar de forma eficaz la lectura y escritura de los bits de información, es necesaria la inyección, anclaje y desenganche forma controlada y repetitiva. Esto implica generar, anclar y desenganchar las paredes de dominio siempre en las mismas condiciones, ie siempre a la misma corriente o campo aplicado. Por ello, en el primer capítulo de resultados de esta tesis estudiamos el anclaje y desenganche de paredes de dominio en defectos de seis formas distintas, cada uno, de dos profundidades diferentes. Hemos realizado un análisis estadístico en diferentes hilos, donde hemos estudiado la probabilidad de anclaje cada tipo de defecto y la dispersión en el valor de campo magnético aplicado necesario para desenganchar la pared. Luego, continuamos con el estudio de la nucleación de las paredes de dominio magnético con pulsos de corriente a través una linea adyacente al nanohilo. Estudiamos defectos de tres formas distintas e identificamos, en función del valor de campo magnético aplicado, los distintos tipos de paredes de dominio anclados en cada uno de ellos. Además, con la ayuda de este método de inyección que es rápido y eficaz, hemos sido capaces de generar y anclar un único tipo de pared minimizando el comportamiento estocástico de la pared mencionado anteriormente. En estas condiciones óptimas, hemos estudiado el desenganche de las paredes de dominio por medio de corriente polarizada en espín, donde hemos conseguido desenganchar la pared de forma controlada y repetitiva siempre para los mismos valores de corriente y campo magnético aplicados. Además, aplicando pulsos de corriente en distintas direcciones, estudiamos en base a su diferencia, la contribución térmica debido al efecto Joule. Los resultados obtenidos representan un importante avance hacia la explotación práctica de este tipo de dispositivos. ABSTRACT The research activity of this thesis was focused on the nucleation, pinning and depinning of magnetic domain walls (DWs) in notched permalloy nanowires. The access to nanofabrication techniques has boosted the number of applications based on magnetic domain walls (DWs) like memory devices. In 2004, Stuart Parkin at IBM, conceived an innovative concept, the “racetrack memory” based on a ferromagnetic nanowire were the magnetic domains constitute the “bits” of information. The frontier between those magnetic domains, ie magnetic domain wall, will move ideally assisted by a spin polarized current. DWs will pin at certain positions due to artificially created pinning sites or “notches” fabricated with ebeam lithography. The success of this idea relies on the careful and predictable control on DW nucleation and a defined pinning-depinning process in order to read and write the bits of information. Sloncsewski in 1994 shows that a spin polarized current can transfer magnetic moment to the local magnetization to move the DWs instead of the magnetic field created by the current. Since then many research groups worldwide have been working on optimizing the conditions for the current induced DW motion due to the spin transfer effect. The fraction of spin polarized electrons traveling through a ferromagnetic nanowire is considerably small, so nowadays the current density required to move a DW by STT exceeds 1 107 A/cm2. A high current density not only can produce a significant degradation of the device but also important effects related to Joule heating were also observed . There are other scientific and technological issues to solve regarding the diversity of DWs states pinned at the notch. The types of DWs pinned, their chirality or their characteristic depinning current or field, may change if the notch has slightly different dimensions, the stripe has different thickness or even if the DW is pinned by a different procedure. Additionally, there is a stochastic component in both the injection of the DW and in its pinning-depinning process, which may be partly intrinsic to the nature of the travelling DW at a non-zero temperature and partly due to the unavoidable defects introduced during the nano-fabrication process. This constitutes an important inconvenient because depending on the DW type different values of current of magnetic field need to be applied in order to depin a DW from the notch. As mentioned earlier, in order to write and read the bits of information accurately, a controlled reproducible and predictable pinning- depinning process is required. This implies to nucleate, pin and depin always at the same applied magnetic field or current. Therefore, in the first chapter of this thesis we studied the pinning and depinning of DW in six different notch shapes and two depths. An statistical analysis was conducted in order to determine which notch type performed best in terms of pinning probability and the dispersion measured in the magnetic field necessary to depin the magnetic DWs. Then, we continued studying the nucleation of DWs with nanosecond current pulses by an adjacent conductive stripe. We studied the conditions for DW injection that allow a selective pinning of the different types of DWs in Permalloy nanostripes with 3 different notch shapes. Furthermore, with this injection method, which has proven to be fast and reliable, we manage to nucleate only one type of DW avoiding its stochastic behavior mentioned earlier. Having achieved this optimized conditions we studied current induced depinning where we also achieved a controlled and reproducible depinning process at always the same applied current and magnetic field. Additionally, changing the pulse polarity we studied the joule heating contribution in a current induced depinning process. The results obtained represent an important step towards the practical exploitation of these devices.