981 resultados para fixed path methods


Relevância:

30.00% 30.00%

Publicador:

Resumo:

In this paper, we present novel precoding methods for multiuser Rayleigh fading multiple-input-multiple-output (MIMO) systems when channel state information (CSI) is available at the transmitter (CSIT) but not at the receiver (CSIR). Such a scenario is relevant, for example, in time-division duplex (TDD) MIMO communications, where, due to channel reciprocity, CSIT can be directly acquired by sending a training sequence from the receiver to the transmitter(s). We propose three transmit precoding schemes that convert the fading MIMO channel into a fixed-gain additive white Gaussian noise (AWGN) channel while satisfying an average power constraint. We also extend one of the precoding schemes to the multiuser Rayleigh fading multiple-access channel (MAC), broadcast channel (BC), and interference channel (IC). The proposed schemes convert the fading MIMO channel into fixed-gain parallel AWGN channels in all three cases. Hence, they achieve an infinite diversity order, which is in sharp contrast to schemes based on perfect CSIR and no CSIT, which, at best, achieve a finite diversity order. Further, we show that a polynomial diversity order is retained, even in the presence of channel estimation errors at the transmitter. Monte Carlo simulations illustrate the bit error rate (BER) performance obtainable from the proposed precoding scheme compared with existing transmit precoding schemes.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

We present methods for fixed-lag smoothing using Sequential Importance sampling (SIS) on a discrete non-linear, non-Gaussian state space system with unknown parameters. Our particular application is in the field of digital communication systems. Each input data point is taken from a finite set of symbols. We represent transmission media as a fixed filter with a finite impulse response (FIR), hence a discrete state-space system is formed. Conventional Markov chain Monte Carlo (MCMC) techniques such as the Gibbs sampler are unsuitable for this task because they can only perform processing on a batch of data. Data arrives sequentially, so it would seem sensible to process it in this way. In addition, many communication systems are interactive, so there is a maximum level of latency that can be tolerated before a symbol is decoded. We will demonstrate this method by simulation and compare its performance to existing techniques.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Moving mesh methods (also called r-adaptive methods) are space-adaptive strategies used for the numerical simulation of time-dependent partial differential equations. These methods keep the total number of mesh points fixed during the simulation, but redistribute them over time to follow the areas where a higher mesh point density is required. There are a very limited number of moving mesh methods designed for solving field-theoretic partial differential equations, and the numerical analysis of the resulting schemes is challenging. In this thesis we present two ways to construct r-adaptive variational and multisymplectic integrators for (1+1)-dimensional Lagrangian field theories. The first method uses a variational discretization of the physical equations and the mesh equations are then coupled in a way typical of the existing r-adaptive schemes. The second method treats the mesh points as pseudo-particles and incorporates their dynamics directly into the variational principle. A user-specified adaptation strategy is then enforced through Lagrange multipliers as a constraint on the dynamics of both the physical field and the mesh points. We discuss the advantages and limitations of our methods. The proposed methods are readily applicable to (weakly) non-degenerate field theories---numerical results for the Sine-Gordon equation are presented.

In an attempt to extend our approach to degenerate field theories, in the last part of this thesis we construct higher-order variational integrators for a class of degenerate systems described by Lagrangians that are linear in velocities. We analyze the geometry underlying such systems and develop the appropriate theory for variational integration. Our main observation is that the evolution takes place on the primary constraint and the 'Hamiltonian' equations of motion can be formulated as an index 1 differential-algebraic system. We then proceed to construct variational Runge-Kutta methods and analyze their properties. The general properties of Runge-Kutta methods depend on the 'velocity' part of the Lagrangian. If the 'velocity' part is also linear in the position coordinate, then we show that non-partitioned variational Runge-Kutta methods are equivalent to integration of the corresponding first-order Euler-Lagrange equations, which have the form of a Poisson system with a constant structure matrix, and the classical properties of the Runge-Kutta method are retained. If the 'velocity' part is nonlinear in the position coordinate, we observe a reduction of the order of convergence, which is typical of numerical integration of DAEs. We also apply our methods to several models and present the results of our numerical experiments.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In the quest for a descriptive theory of decision-making, the rational actor model in economics imposes rather unrealistic expectations and abilities on human decision makers. The further we move from idealized scenarios, such as perfectly competitive markets, and ambitiously extend the reach of the theory to describe everyday decision making situations, the less sense these assumptions make. Behavioural economics has instead proposed models based on assumptions that are more psychologically realistic, with the aim of gaining more precision and descriptive power. Increased psychological realism, however, comes at the cost of a greater number of parameters and model complexity. Now there are a plethora of models, based on different assumptions, applicable in differing contextual settings, and selecting the right model to use tends to be an ad-hoc process. In this thesis, we develop optimal experimental design methods and evaluate different behavioral theories against evidence from lab and field experiments.

We look at evidence from controlled laboratory experiments. Subjects are presented with choices between monetary gambles or lotteries. Different decision-making theories evaluate the choices differently and would make distinct predictions about the subjects' choices. Theories whose predictions are inconsistent with the actual choices can be systematically eliminated. Behavioural theories can have multiple parameters requiring complex experimental designs with a very large number of possible choice tests. This imposes computational and economic constraints on using classical experimental design methods. We develop a methodology of adaptive tests: Bayesian Rapid Optimal Adaptive Designs (BROAD) that sequentially chooses the "most informative" test at each stage, and based on the response updates its posterior beliefs over the theories, which informs the next most informative test to run. BROAD utilizes the Equivalent Class Edge Cutting (EC2) criteria to select tests. We prove that the EC2 criteria is adaptively submodular, which allows us to prove theoretical guarantees against the Bayes-optimal testing sequence even in the presence of noisy responses. In simulated ground-truth experiments, we find that the EC2 criteria recovers the true hypotheses with significantly fewer tests than more widely used criteria such as Information Gain and Generalized Binary Search. We show, theoretically as well as experimentally, that surprisingly these popular criteria can perform poorly in the presence of noise, or subject errors. Furthermore, we use the adaptive submodular property of EC2 to implement an accelerated greedy version of BROAD which leads to orders of magnitude speedup over other methods.

We use BROAD to perform two experiments. First, we compare the main classes of theories for decision-making under risk, namely: expected value, prospect theory, constant relative risk aversion (CRRA) and moments models. Subjects are given an initial endowment, and sequentially presented choices between two lotteries, with the possibility of losses. The lotteries are selected using BROAD, and 57 subjects from Caltech and UCLA are incentivized by randomly realizing one of the lotteries chosen. Aggregate posterior probabilities over the theories show limited evidence in favour of CRRA and moments' models. Classifying the subjects into types showed that most subjects are described by prospect theory, followed by expected value. Adaptive experimental design raises the possibility that subjects could engage in strategic manipulation, i.e. subjects could mask their true preferences and choose differently in order to obtain more favourable tests in later rounds thereby increasing their payoffs. We pay close attention to this problem; strategic manipulation is ruled out since it is infeasible in practice, and also since we do not find any signatures of it in our data.

In the second experiment, we compare the main theories of time preference: exponential discounting, hyperbolic discounting, "present bias" models: quasi-hyperbolic (α, β) discounting and fixed cost discounting, and generalized-hyperbolic discounting. 40 subjects from UCLA were given choices between 2 options: a smaller but more immediate payoff versus a larger but later payoff. We found very limited evidence for present bias models and hyperbolic discounting, and most subjects were classified as generalized hyperbolic discounting types, followed by exponential discounting.

In these models the passage of time is linear. We instead consider a psychological model where the perception of time is subjective. We prove that when the biological (subjective) time is positively dependent, it gives rise to hyperbolic discounting and temporal choice inconsistency.

We also test the predictions of behavioral theories in the "wild". We pay attention to prospect theory, which emerged as the dominant theory in our lab experiments of risky choice. Loss aversion and reference dependence predicts that consumers will behave in a uniquely distinct way than the standard rational model predicts. Specifically, loss aversion predicts that when an item is being offered at a discount, the demand for it will be greater than that explained by its price elasticity. Even more importantly, when the item is no longer discounted, demand for its close substitute would increase excessively. We tested this prediction using a discrete choice model with loss-averse utility function on data from a large eCommerce retailer. Not only did we identify loss aversion, but we also found that the effect decreased with consumers' experience. We outline the policy implications that consumer loss aversion entails, and strategies for competitive pricing.

In future work, BROAD can be widely applicable for testing different behavioural models, e.g. in social preference and game theory, and in different contextual settings. Additional measurements beyond choice data, including biological measurements such as skin conductance, can be used to more rapidly eliminate hypothesis and speed up model comparison. Discrete choice models also provide a framework for testing behavioural models with field data, and encourage combined lab-field experiments.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In a paper published in 1961, L. Cesari [1] introduces a method which extends certain earlier existence theorems of Cesari and Hale ([2] to [6]) for perturbation problems to strictly nonlinear problems. Various authors ([1], [7] to [15]) have now applied this method to nonlinear ordinary and partial differential equations. The basic idea of the method is to use the contraction principle to reduce an infinite-dimensional fixed point problem to a finite-dimensional problem which may be attacked using the methods of fixed point indexes.

The following is my formulation of the Cesari fixed point method:

Let B be a Banach space and let S be a finite-dimensional linear subspace of B. Let P be a projection of B onto S and suppose Г≤B such that pГ is compact and such that for every x in PГ, P-1x∩Г is closed. Let W be a continuous mapping from Г into B. The Cesari method gives sufficient conditions for the existence of a fixed point of W in Г.

Let I denote the identity mapping in B. Clearly y = Wy for some y in Г if and only if both of the following conditions hold:

(i) Py = PWy.

(ii) y = (P + (I - P)W)y.

Definition. The Cesari fixed paint method applies to (Г, W, P) if and only if the following three conditions are satisfied:

(1) For each x in PГ, P + (I - P)W is a contraction from P-1x∩Г into itself. Let y(x) be that element (uniqueness follows from the contraction principle) of P-1x∩Г which satisfies the equation y(x) = Py(x) + (I-P)Wy(x).

(2) The function y just defined is continuous from PГ into B.

(3) There are no fixed points of PWy on the boundary of PГ, so that the (finite- dimensional) fixed point index i(PWy, int PГ) is defined.

Definition. If the Cesari fixed point method applies to (Г, W, P) then define i(Г, W, P) to be the index i(PWy, int PГ).

The three theorems of this thesis can now be easily stated.

Theorem 1 (Cesari). If i(Г, W, P) is defined and i(Г, W, P) ≠0, then there is a fixed point of W in Г.

Theorem 2. Let the Cesari fixed point method apply to both (Г, W, P1) and (Г, W, P2). Assume that P2P1=P1P2=P1 and assume that either of the following two conditions holds:

(1) For every b in B and every z in the range of P2, we have that ‖b=P2b‖ ≤ ‖b-z‖

(2)P2Г is convex.

Then i(Г, W, P1) = i(Г, W, P2).

Theorem 3. If Ω is a bounded open set and W is a compact operator defined on Ω so that the (infinite-dimensional) Leray-Schauder index iLS(W, Ω) is defined, and if the Cesari fixed point method applies to (Ω, W, P), then i(Ω, W, P) = iLS(W, Ω).

Theorems 2 and 3 are proved using mainly a homotopy theorem and a reduction theorem for the finite-dimensional and the Leray-Schauder indexes. These and other properties of indexes will be listed before the theorem in which they are used.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Static optical transmission is restudied by postulation of the optical path as the proper element in a three-dimensional Riemannian manifold (no torsion); this postulation can be applied to describe the light-medium interactive system. On the basis of the postulation, the behaviors of light transmitting through the medium with refractive index n are investigated, the investigation covering the realms of both geometrical optics and wave optics. The wave equation of light in static transmission is studied modally, the postulation being employed to derive the exact form of the optical field equation in a medium (in which the light is viewed as a single-component field). Correspondingly, the relationships concerning the conservation of optical fluid and the dynamic properties are given, and some simple applications of the theories mentioned are presented.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Estimation of individual egg production (realized fecundity) is a key step either to understand the stock and recruit relationship or to carry out fisheries-independent assessment of spawning stock biomass using egg production methods. Many fish are highly fecund and their ovaries may weigh over a kilogram; therefore the work time can be consuming and require large quantities of toxic fixative. Recently it has been shown for Atlantic cod (Gadus morhua) that image analysis can automate fecundity determination using a power equation that links follicles per gram ovary to the mean vitellogenic follicular diameter (the autodiametric method). In this article we demonstrate the precision of the autodiametric method applied to a range of species with different spawning strategies during maturation and spawning. A new method using a solid displacement pipette to remove quantitative fecundity samples (25, 50, 100, and 200 milligram [mg]) is evaluated, as are the underlying assumptions to effectively fix and subsample the ovary. Finally, we demonstrate the interpretation of dispersed formaldehyde-fixed ovarian samples (whole mounts) to assess the presence of atretic and postovulatory follicles to replace labor intensive histology. These results can be used to estimate down regulation (production of atretic follicles) of fecundity during maturation.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

During the VITAL cruise in the Bay of Biscay in summer 2002, two devices for measuring the length of swimming fish were tested: 1) a mechanical crown that emitted a pair of parallel laser beams and that was mounted on the main camera and 2) an underwater auto-focus video camera. The precision and accuracy of these devices were compared and the various sources of measurement errors were estimated by repeatedly measuring fixed and mobile objects and live fish. It was found that fish mobility is the main source of error for these devices because they require that the objects to be measured are perpendicular to the field of vision. The best performance was obtained with the laser method where a video-replay of laser spots (projected on fish bodies) carrying real-time size information was used. The auto-focus system performed poorly because of a delay in obtaining focus and because of some technical problems.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This paper is concerned with the modelling of strategic interactions between the human driver and the vehicle active front steering (AFS) controller in a path-following task where the two controllers hold different target paths. The work is aimed at extending the use of mathematical models in representing driver steering behaviour in complicated driving situations. Two game theoretic approaches, namely linear quadratic game and non-cooperative model predictive control (non-cooperative MPC), are used for developing the driver-AFS interactive steering control model. For each approach, the open-loop Nash steering control solution is derived; the influences of the path-following weights, preview and control horizons, driver time delay and arm neuromuscular system (NMS) dynamics are investigated, and the CPU time consumed is recorded. It is found that the two approaches give identical time histories as well as control gains, while the non-cooperative MPC method uses much less CPU time. Specifically, it is observed that the introduction of weight on the integral of vehicle lateral displacement error helps to eliminate the steady-state path-following error; the increase in preview horizon and NMS natural frequency and the decline in time delay and NMS damping ratio improve the path-following accuracy. © 2013 Copyright Taylor and Francis Group, LLC.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The paper is concerned with the identification of theoretical preview steering controllers using data obtained from five test subjects in a fixed-base driving simulator. An understanding of human steering control behaviour is relevant to the design of autonomous and semi-autonomous vehicle controls. The driving task involved steering a linear vehicle along a randomly curving path. The theoretical steering controllers identified from the data were based on optimal linear preview control. A direct-identification method was used, and the steering controllers were identified so that the predicted steering angle matched as closely as possible the measured steering angle of the test subjects. It was found that identification of the driver's time delay and noise is necessary to avoid bias in identification of the controller parameters. Most subjects' steering behaviour was predicted well by a theoretical controller based on the lateral/yaw dynamics of the vehicle. There was some evidence that an inexperienced driver's steering action was better represented by a controller based on a simpler model of the vehicle dynamics, perhaps reflecting incomplete learning by the driver. Copyright © 2014 Inderscience Enterprises Ltd.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

We present a fixed-grid finite element technique for fluid-structure interaction problems involving incompressible viscous flows and thin structures. The flow equations are discretised with isoparametric b-spline basis functions defined on a logically Cartesian grid. In addition, the previously proposed subdivision-stabilisation technique is used to ensure inf-sup stability. The beam equations are discretised with b-splines and the shell equations with subdivision basis functions, both leading to a rotation-free formulation. The interface conditions between the fluid and the structure are enforced with the Nitsche technique. The resulting coupled system of equations is solved with a Dirichlet-Robin partitioning scheme, and the fluid equations are solved with a pressure-correction method. Auxiliary techniques employed for improving numerical robustness include the level-set based implicit representation of the structure interface on the fluid grid, a cut-cell integration algorithm based on marching tetrahedra and the conservative data transfer between the fluid and structure discretisations. A number of verification and validation examples, primarily motivated by animal locomotion in air or water, demonstrate the robustness and efficiency of our approach. © 2013 John Wiley & Sons, Ltd.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Basis path testing is a very powerful structural testing criterion. The number of test paths equals to the cyclomatic complexity of program defined by McCabe. Traditional test generation methods select the paths either without consideration of the constraints of variables or interactively. In this note, an efficient method is presented to generate a set of feasible basis paths. The experiments show that this method can generate feasible basis paths for real-world C programs automatically in acceptable time.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Because of its high energy density direct current(dc)thermal plasmas are widely accepted as a processing medium which facilitates high processing rates high fluxes of radical species the potential for smaller jnstallations a wide choice of reactants and high quench rates[1].A broad range of industrial processing methods have been developed based on dc plasma technology. However,nonstationary features limited new applications of dc plasma in advanced processing, where reliability£¬reproducibility and precise controllability are required£. These challenges call for better understanding of the arc and jet behavior over a wide range of generating parameters and a comprehensive control of every aspect of lhe plasma processing.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Accurate measurement of network bandwidth is crucial for flexible Internet applications and protocols which actively manage and dynamically adapt to changing utilization of network resources. These applications must do so to perform tasks such as distributing and delivering high-bandwidth media, scheduling service requests and performing admission control. Extensive work has focused on two approaches to measuring bandwidth: measuring it hop-by-hop, and measuring it end-to-end along a path. Unfortunately, best-practice techniques for the former are inefficient and techniques for the latter are only able to observe bottlenecks visible at end-to-end scope. In this paper, we develop and simulate end-to-end probing methods which can measure bottleneck bandwidth along arbitrary, targeted subpaths of a path in the network, including subpaths shared by a set of flows. As another important contribution, we describe a number of practical applications which we foresee as standing to benefit from solutions to this problem, especially in emerging, flexible network architectures such as overlay networks, ad-hoc networks, peer-to-peer architectures and massively accessed content servers.