962 resultados para Non-linear parameter estimation


Relevância:

100.00% 100.00%

Publicador:

Resumo:

In this talk we investigate the usage of spectrally shaped amplified spontaneous emission (ASE) in order to emulate highly dispersed wavelength division multiplexed (WDM) signals in an optical transmission system. Such a technique offers various simplifications to large scale WDM experiments. Not only does it offer a reduction in transmitter complexity, removing the need for multiple source lasers, it potentially reduces the test and measurement complexity by requiring only the centre channel of a WDM system to be measured in order to estimate WDM worst case performance. The use of ASE as a test and measurement tool is well established in optical communication systems and several measurement techniques will be discussed [1, 2]. One of the most prevalent uses of ASE is in the measurement of receiver sensitivity where ASE is introduced in order to degrade the optical signal to noise ratio (OSNR) and measure the resulting bit error rate (BER) at the receiver. From an analytical point of view noise has been used to emulate system performance, the Gaussian Noise model is used as an estimate of highly dispersed signals and has had consider- able interest [3]. The work to be presented here extends the use of ASE by using it as a metric to emulate highly dispersed WDM signals and in the process reduce WDM transmitter complexity and receiver measurement time in a lab environment. Results thus far have indicated [2] that such a transmitter configuration is consistent with an AWGN model for transmission, with modulation format complexity and nonlinearities playing a key role in estimating the performance of systems utilising the ASE channel emulation technique. We conclude this work by investigating techniques capable of characterising the nonlinear and damage limits of optical fibres and the resultant information capacity limits. REFERENCES McCarthy, M. E., N. Mac Suibhne, S. T. Le, P. Harper, and A. D. Ellis, “High spectral efficiency transmission emulation for non-linear transmission performance estimation for high order modulation formats," 2014 European Conference on IEEE Optical Communication (ECOC), 2014. 2. Ellis, A., N. Mac Suibhne, F. Gunning, and S. Sygletos, “Expressions for the nonlinear trans- mission performance of multi-mode optical fiber," Opt. Express, Vol. 21, 22834{22846, 2013. Vacondio, F., O. Rival, C. Simonneau, E. Grellier, A. Bononi, L. Lorcy, J. Antona, and S. Bigo, “On nonlinear distortions of highly dispersive optical coherent systems," Opt. Express, Vol. 20, 1022-1032, 2012.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

This research explores Bayesian updating as a tool for estimating parameters probabilistically by dynamic analysis of data sequences. Two distinct Bayesian updating methodologies are assessed. The first approach focuses on Bayesian updating of failure rates for primary events in fault trees. A Poisson Exponentially Moving Average (PEWMA) model is implemnented to carry out Bayesian updating of failure rates for individual primary events in the fault tree. To provide a basis for testing of the PEWMA model, a fault tree is developed based on the Texas City Refinery incident which occurred in 2005. A qualitative fault tree analysis is then carried out to obtain a logical expression for the top event. A dynamic Fault Tree analysis is carried out by evaluating the top event probability at each Bayesian updating step by Monte Carlo sampling from posterior failure rate distributions. It is demonstrated that PEWMA modeling is advantageous over conventional conjugate Poisson-Gamma updating techniques when failure data is collected over long time spans. The second approach focuses on Bayesian updating of parameters in non-linear forward models. Specifically, the technique is applied to the hydrocarbon material balance equation. In order to test the accuracy of the implemented Bayesian updating models, a synthetic data set is developed using the Eclipse reservoir simulator. Both structured grid and MCMC sampling based solution techniques are implemented and are shown to model the synthetic data set with good accuracy. Furthermore, a graphical analysis shows that the implemented MCMC model displays good convergence properties. A case study demonstrates that Likelihood variance affects the rate at which the posterior assimilates information from the measured data sequence. Error in the measured data significantly affects the accuracy of the posterior parameter distributions. Increasing the likelihood variance mitigates random measurement errors, but casuses the overall variance of the posterior to increase. Bayesian updating is shown to be advantageous over deterministic regression techniques as it allows for incorporation of prior belief and full modeling uncertainty over the parameter ranges. As such, the Bayesian approach to estimation of parameters in the material balance equation shows utility for incorporation into reservoir engineering workflows.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Les systèmes de communication optique avec des formats de modulation avancés sont actuellement l’un des sujets de recherche les plus importants dans le domaine de communication optique. Cette recherche est stimulée par les exigences pour des débits de transmission de donnée plus élevés. Dans cette thèse, on examinera les techniques efficaces pour la modulation avancée avec une détection cohérente, et multiplexage par répartition en fréquence orthogonale (OFDM) et multiples tonalités discrètes (DMT) pour la détection directe et la détection cohérente afin d’améliorer la performance de réseaux optiques. Dans la première partie, nous examinons la rétropropagation avec filtre numérique (DFBP) comme une simple technique d’atténuation de nonlinéarité d’amplificateur optique semiconducteur (SOA) dans le système de détection cohérente. Pour la première fois, nous démontrons expérimentalement l’efficacité de DFBP pour compenser les nonlinéarités générées par SOA dans un système de détection cohérente porteur unique 16-QAM. Nous comparons la performance de DFBP avec la méthode de Runge-Kutta quatrième ordre. Nous examinons la sensibilité de performance de DFBP par rapport à ses paramètres. Par la suite, nous proposons une nouvelle méthode d’estimation de paramètre pour DFBP. Finalement, nous démontrons la transmission de signaux de 16-QAM aux taux de 22 Gbaud sur 80km de fibre optique avec la technique d’estimation de paramètre proposée pour DFBP. Dans la deuxième partie, nous nous concentrons sur les techniques afin d’améliorer la performance des systèmes OFDM optiques en examinent OFDM optiques cohérente (CO-OFDM) ainsi que OFDM optiques détection directe (DDO-OFDM). Premièrement, nous proposons une combinaison de coupure et prédistorsion pour compenser les distorsions nonlinéaires d’émetteur de CO-OFDM. Nous utilisons une interpolation linéaire par morceaux (PLI) pour charactériser la nonlinéarité d’émetteur. Dans l’émetteur nous utilisons l’inverse de l’estimation de PLI pour compenser les nonlinéarités induites à l’émetteur de CO-OFDM. Deuxièmement, nous concevons des constellations irrégulières optimisées pour les systèmes DDO-OFDM courte distance en considérant deux modèles de bruit de canal. Nous démontrons expérimentalement 100Gb/s+ OFDM/DMT avec la détection directe en utilisant les constellations QAM optimisées. Dans la troisième partie, nous proposons une architecture réseaux optiques passifs (PON) avec DDO-OFDM pour la liaison descendante et CO-OFDM pour la liaison montante. Nous examinons deux scénarios pour l’allocations de fréquence et le format de modulation des signaux. Nous identifions la détérioration limitante principale du PON bidirectionnelle et offrons des solutions pour minimiser ses effets.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The idea of spacecraft formations, flying in tight configurations with maximum baselines of a few hundred meters in low-Earth orbits, has generated widespread interest over the last several years. Nevertheless, controlling the movement of spacecraft in formation poses difficulties, such as in-orbit high-computing demand and collision avoidance capabilities, which escalate as the number of units in the formation is increased and complicated nonlinear effects are imposed to the dynamics, together with uncertainty which may arise from the lack of knowledge of system parameters. These requirements have led to the need of reliable linear and nonlinear controllers in terms of relative and absolute dynamics. The objective of this thesis is, therefore, to introduce new control methods to allow spacecraft in formation, with circular/elliptical reference orbits, to efficiently execute safe autonomous manoeuvres. These controllers distinguish from the bulk of literature in that they merge guidance laws never applied before to spacecraft formation flying and collision avoidance capacities into a single control strategy. For this purpose, three control schemes are presented: linear optimal regulation, linear optimal estimation and adaptive nonlinear control. In general terms, the proposed control approaches command the dynamical performance of one or several followers with respect to a leader to asymptotically track a time-varying nominal trajectory (TVNT), while the threat of collision between the followers is reduced by repelling accelerations obtained from the collision avoidance scheme during the periods of closest proximity. Linear optimal regulation is achieved through a Riccati-based tracking controller. Within this control strategy, the controller provides guidance and tracking toward a desired TVNT, optimizing fuel consumption by Riccati procedure using a non-infinite cost function defined in terms of the desired TVNT, while repelling accelerations generated from the CAS will ensure evasive actions between the elements of the formation. The relative dynamics model, suitable for circular and eccentric low-Earth reference orbits, is based on the Tschauner and Hempel equations, and includes a control input and a nonlinear term corresponding to the CAS repelling accelerations. Linear optimal estimation is built on the forward-in-time separation principle. This controller encompasses two stages: regulation and estimation. The first stage requires the design of a full state feedback controller using the state vector reconstructed by means of the estimator. The second stage requires the design of an additional dynamical system, the estimator, to obtain the states which cannot be measured in order to approximately reconstruct the full state vector. Then, the separation principle states that an observer built for a known input can also be used to estimate the state of the system and to generate the control input. This allows the design of the observer and the feedback independently, by exploiting the advantages of linear quadratic regulator theory, in order to estimate the states of a dynamical system with model and sensor uncertainty. The relative dynamics is described with the linear system used in the previous controller, with a control input and nonlinearities entering via the repelling accelerations from the CAS during collision avoidance events. Moreover, sensor uncertainty is added to the control process by considering carrier-phase differential GPS (CDGPS) velocity measurement error. An adaptive control law capable of delivering superior closed-loop performance when compared to the certainty-equivalence (CE) adaptive controllers is finally presented. A novel noncertainty-equivalence controller based on the Immersion and Invariance paradigm for close-manoeuvring spacecraft formation flying in both circular and elliptical low-Earth reference orbits is introduced. The proposed control scheme achieves stabilization by immersing the plant dynamics into a target dynamical system (or manifold) that captures the desired dynamical behaviour. They key feature of this methodology is the addition of a new term to the classical certainty-equivalence control approach that, in conjunction with the parameter update law, is designed to achieve adaptive stabilization. This parameter has the ultimate task of shaping the manifold into which the adaptive system is immersed. The performance of the controller is proven stable via a Lyapunov-based analysis and Barbalat’s lemma. In order to evaluate the design of the controllers, test cases based on the physical and orbital features of the Prototype Research Instruments and Space Mission Technology Advancement (PRISMA) are implemented, extending the number of elements in the formation into scenarios with reconfigurations and on-orbit position switching in elliptical low-Earth reference orbits. An extensive analysis and comparison of the performance of the controllers in terms of total Δv and fuel consumption, with and without the effects of the CAS, is presented. These results show that the three proposed controllers allow the followers to asymptotically track the desired nominal trajectory and, additionally, those simulations including CAS show an effective decrease of collision risk during the performance of the manoeuvre.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

This thesis is a compilation of 6 papers that the author has written together with Alberto Lanconelli (chapters 3, 5 and 8) and Hyun-Jung Kim (ch 7). The logic thread that link all these chapters together is the interest to analyze and approximate the solutions of certain stochastic differential equations using the so called Wick product as the basic tool. In the first chapter we present arguably the most important achievement of this thesis; namely the generalization to multiple dimensions of a Wick-Wong-Zakai approximation theorem proposed by Hu and Oksendal. By exploiting the relationship between the Wick product and the Malliavin derivative we propose an original reduction method which allows us to approximate semi-linear systems of stochastic differential equations of the Itô type. Furthermore in chapter 4 we present a non-trivial extension of the aforementioned results to the case in which the system of stochastic differential equations are driven by a multi-dimensional fraction Brownian motion with Hurst parameter bigger than 1/2. In chapter 5 we employ our approach and present a “short time” approximation for the solution of the Zakai equation from non-linear filtering theory and provide an estimation of the speed of convergence. In chapters 6 and 7 we study some properties of the unique mild solution for the Stochastic Heat Equation driven by spatial white noise of the Wick-Skorohod type. In particular by means of our reduction method we obtain an alternative derivation of the Feynman-Kac representation for the solution, we find its optimal Hölder regularity in time and space and present a Feynman-Kac-type closed form for its spatial derivative. Chapter 8 treats a somewhat different topic; in particular we investigate some probabilistic aspects of the unique global strong solution of a two dimensional system of semi-linear stochastic differential equations describing a predator-prey model perturbed by Gaussian noise.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Fenômenos oscilatórios e ressonantes são explorados em vários cursos experimentais de física. Em geral os experimentos são interpretados no limite de pequenas oscilações e campos uniformes. Neste artigo descrevemos um experimento de baixo custo para o estudo da ressonância em campo magnético da agulha de uma bússola fora dos limites acima. Nesse caso, termos não lineares na equação diferencial são responsáveis por fenômenos interessantes de serem explorados em laboratórios didáticos.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

One of the electrical impedance tomography objectives is to estimate the electrical resistivity distribution in a domain based only on electrical potential measurements at its boundary generated by an imposed electrical current distribution into the boundary. One of the methods used in dynamic estimation is the Kalman filter. In biomedical applications, the random walk model is frequently used as evolution model and, under this conditions, poor tracking ability of the extended Kalman filter (EKF) is achieved. An analytically developed evolution model is not feasible at this moment. The paper investigates the identification of the evolution model in parallel to the EKF and updating the evolution model with certain periodicity. The evolution model transition matrix is identified using the history of the estimated resistivity distribution obtained by a sensitivity matrix based algorithm and a Newton-Raphson algorithm. To numerically identify the linear evolution model, the Ibrahim time-domain method is used. The investigation is performed by numerical simulations of a domain with time-varying resistivity and by experimental data collected from the boundary of a human chest during normal breathing. The obtained dynamic resistivity values lie within the expected values for the tissues of a human chest. The EKF results suggest that the tracking ability is significantly improved with this approach.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Quantum computers promise to increase greatly the efficiency of solving problems such as factoring large integers, combinatorial optimization and quantum physics simulation. One of the greatest challenges now is to implement the basic quantum-computational elements in a physical system and to demonstrate that they can be reliably and scalably controlled. One of the earliest proposals for quantum computation is based on implementing a quantum bit with two optical modes containing one photon. The proposal is appealing because of the ease with which photon interference can be observed. Until now, it suffered from the requirement for non-linear couplings between optical modes containing few photons. Here we show that efficient quantum computation is possible using only beam splitters, phase shifters, single photon sources and photo-detectors. Our methods exploit feedback from photo-detectors and are robust against errors from photon loss and detector inefficiency. The basic elements are accessible to experimental investigation with current technology.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Previous magnetic resonance imaging (MRI) studies described consistent age-related gray matter (GM) reductions in the fronto-parietal neocortex, insula and cerebellum in elderly subjects, but not as frequently in limbic/paralimbic structures. However, it is unclear whether such features are already present during earlier stages of adulthood, and if age-related GM changes may follow non-linear patterns at such age range. This voxel-based morphometry study investigated the relationship between GM volumes and age specifically during non-elderly life (18-50 years) in 89 healthy individuals (48 males and 41 females). Voxelwise analyses showed significant (p < 0.05, corrected) negative correlations in the right prefrontal cortex and left cerebellum, and positive correlations (indicating lack of GM loss) in the medial temporal region, cingulate gyrus, insula and temporal neocortex. Analyses using ROI masks showed that age-related dorsolateral prefrontal volume decrements followed non-linear patterns, and were less prominent in females compared to males at this age range. These findings further support for the notion of a heterogeneous and asynchronous pattern of age-related brain morphometric changes, with region-specific non-linear features. (C) 2009 Elsevier Inc. All rights reserved.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The catalytic properties of enzymes are usually evaluated by measuring and analyzing reaction rates. However, analyzing the complete time course can be advantageous because it contains additional information about the properties of the enzyme. Moreover, for systems that are not at steady state, the analysis of time courses is the preferred method. One of the major barriers to the wide application of time courses is that it may be computationally more difficult to extract information from these experiments. Here the basic approach to analyzing time courses is described, together with some examples of the essential computer code to implement these analyses. A general method that can be applied to both steady state and non-steady-state systems is recommended. (C) 2001 academic Press.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Quantum feedback can stabilize a two-level atom against decoherence (spontaneous emission), putting it into an arbitrary (specified) pure state. This requires perfect homodyne detection of the atomic emission, and instantaneous feedback. Inefficient detection was considered previously by two of us. Here we allow for a non-zero delay time tau in the feedback circuit. Because a two-level atom is a non-linear optical system, an analytical solution is not possible. However, quantum trajectories allow a simple numerical simulation of the resulting non-Markovian process. We find the effect of the time delay to be qualitatively similar to chat of inefficient detection. The solution of the non-Markovian quantum trajectory will not remain fixed, so that the time-averaged state will be mixed, not pure. In the case where one tries to stabilize the atom in the excited state, an approximate analytical solution to the quantum trajectory is possible. The result, that the purity (P = 2Tr[rho (2)] - 1) of the average state is given by P = 1 - 4y tau (where gamma is the spontaneous emission rate) is found to agree very well with the numerical results. (C) 2001 Elsevier Science B.V. All rights reserved.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Most finite element packages use the Newmark algorithm for time integration of structural dynamics. Various algorithms have been proposed to better optimize the high frequency dissipation of this algorithm. Hulbert and Chung proposed both implicit and explicit forms of the generalized alpha method. The algorithms optimize high frequency dissipation effectively, and despite recent work on algorithms that possess momentum conserving/energy dissipative properties in a non-linear context, the generalized alpha method remains an efficient way to solve many problems, especially with adaptive timestep control. However, the implicit and explicit algorithms use incompatible parameter sets and cannot be used together in a spatial partition, whereas this can be done for the Newmark algorithm, as Hughes and Liu demonstrated, and for the HHT-alpha algorithm developed from it. The present paper shows that the explicit generalized alpha method can be rewritten so that it becomes compatible with the implicit form. All four algorithmic parameters can be matched between the explicit and implicit forms. An element interface between implicit and explicit partitions can then be used, analogous to that devised by Hughes and Liu to extend the Newmark method. The stability of the explicit/implicit algorithm is examined in a linear context and found to exceed that of the explicit partition. The element partition is significantly less dissipative of intermediate frequencies than one using the HHT-alpha method. The explicit algorithm can also be rewritten so that the discrete equation of motion evaluates forces from displacements and velocities found at the predicted mid-point of a cycle. Copyright (C) 2003 John Wiley Sons, Ltd.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Understanding the performance of banks is of the utmost relevance, because of the impact of this sector on economic growth and financial stability. Of all the different assets that make up a bank portfolio, the residential mortgage loans constitute one of its main. Using the dynamic panel data method, we analyse the influence of residential mortgage loans on bank profitability and risk, using a sample of 555 banks in the European Union (EU-15), over the period from 1995 to 2008. We find that banks with larger weights of residential mortgage loans show lower credit risk in good times. This result explains why banks rush to lend on property during booms due to the positive effects it has on credit risk. The results show further that credit risk and profitability are lower during the upturn in the residential property price cycle. The results also reveal the existence of a non-linear relationship (U-shaped marginal effect), as a function of bank’s risk, between profitability and the residential mortgage loans exposure. For those banks that have high credit risk, a large exposure of residential mortgage loans is associated with higher risk-adjusted profitability, through lower risk. For banks with a moderate/low credit risk, the effects of higher residential mortgage loan exposure on its risk-adjusted profitability are also positive or marginally positive.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Understanding the performance of banks is of the u tmost importance due to the impact the sector may have on economic growth and financial stability. Residential mortgage loans constitute a large proportion of the portfolio of many banks and are one of the key assets in the determination of performance. Using a dynamic panel model , we analyse the impact of res idential mortgage loans on bank profitability and risk , based on a sample of 555 banks in the European Union ( EU - 15 ) , over the period from 1995 to 2008. We find that banks with larger weight s in residential mortgage loans display lower credit risk in good market conditions . This result may explain why banks rush to lend on property during b ooms due to the positive effect it has on credit risk . The results also show that credit risk and profitability are lower during the upturn in the residential property cy cle. Furthermore, t he results reveal the existence of a non - linear relationship ( U - shaped marginal effect), as a function of bank’s risk, between profitability and residential mortgage exposure . For those banks that have high er credit risk, a large exposur e to residential loans is associated with increased risk - adjusted profitability, through a reduction in risk. For banks with a moderate to low credit risk, the impact of higher exposure are also positive on risk - adjusted profitability.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

A geração de trajectórias de robôs em tempo real é uma tarefa muito complexa, não existindo ainda um algoritmo que a permita resolver de forma eficaz. De facto, há controladores eficientes para trajectórias previamente definidas, todavia, a adaptação a variações imprevisíveis, como sendo terrenos irregulares ou obstáculos, constitui ainda um problema em aberto na geração de trajectórias em tempo real de robôs. Neste trabalho apresentam-se modelos de geradores centrais de padrões de locomoção (CPGs), inspirados na biologia, que geram os ritmos locomotores num robô quadrúpede. Os CPGs são modelados matematicamente por sistemas acoplados de células (ou neurónios), sendo a dinâmica de cada célula dada por um sistema de equações diferenciais ordinárias não lineares. Assume-se que as trajectórias dos robôs são constituídas por esta parte rítmica e por uma parte discreta. A parte discreta pode ser embebida na parte rítmica, (a.1) como um offset ou (a.2) adicionada às expressões rítmicas, ou (b) pode ser calculada independentemente e adicionada exactamente antes do envio dos sinais para as articulações do robô. A parte discreta permite inserir no passo locomotor uma perturbação, que poderá estar associada à locomoção em terrenos irregulares ou à existência de obstáculos na trajectória do robô. Para se proceder á análise do sistema com parte discreta, será variado o parâmetro g. O parâmetro g, presente nas equações da parte discreta, representa o offset do sinal após a inclusão da parte discreta. Revê-se a teoria de bifurcação e simetria que permite a classificação das soluções periódicas produzidas pelos modelos de CPGs com passos locomotores quadrúpedes. Nas simulações numéricas, usam-se as equações de Morris-Lecar e o oscilador de Hopf como modelos da dinâmica interna de cada célula para a parte rítmica. A parte discreta é modelada por um sistema inspirado no modelo VITE. Medem-se a amplitude e a frequência de dois passos locomotores para variação do parâmetro g, no intervalo [-5;5]. Consideram-se duas formas distintas de incluir a parte discreta na parte rítmica: (a) como um (a.1) offset ou (a.2) somada nas expressões que modelam a parte rítmica, e (b) somada ao sinal da parte rítmica antes de ser enviado às articulações do robô. No caso (a.1), considerando o oscilador de Hopf como dinâmica interna das células, verifica-se que a amplitude e frequência se mantêm constantes para -50.2. A extensão do movimento varia de forma directamente proporcional à amplitude. No caso das equações de Morris-Lecar, quando a componente discreta é embebida (a.2), a amplitude e a frequência aumentam e depois diminuem para - 0.170.5 Pode concluir-se que: (1) a melhor forma de inserção da parte discreta que menos perturbação insere no robô é a inserção como offset; (2) a inserção da parte discreta parece ser independente do sistema de equações diferenciais ordinárias que modelam a dinâmica interna de cada célula. Como trabalho futuro, é importante prosseguir o estudo das diferentes formas de inserção da parte discreta na parte rítmica do movimento, para que se possa gerar uma locomoção quadrúpede, robusta, flexível, com objectivos, em terrenos irregulares, modelada por correcções discretas aos padrões rítmicos.