947 resultados para D-space
Resumo:
The quantum statistical mechanical propagator for a harmonic oscillator with a time-dependent force constant, m omega(2)(t), has been investigated in the past and was found to have only a formal solution in terms of the solutions of certain ordinary differential equations. Such path integrals are frequently encountered in semiclassical path integral evaluations and having exact analytical expressions for such path integrals is of great interest. In a previous work, we had obtained the exact propagator for motion in an arbitrary time-dependent harmonic potential in the overdamped limit of friction using phase space path integrals in the context of Levy flights - a result that can be easily extended to Brownian motion. In this paper, we make a connection between the overdamped Brownian motion and the imaginary time propagator of quantum mechanics and thereby get yet another way to evaluate the latter exactly. We find that explicit analytic solution for the quantum statistical mechanical propagator can be written when the time-dependent force constant has the form omega(2)(t) = lambda(2)(t) - d lambda(t)/dt where lambda(t) is any arbitrary function of t and use it to evaluate path integrals which have not been evaluated previously. We also employ this method to arrive at a formal solution of the propagator for both Levy flights and Brownian subjected to a time-dependent harmonic potential in the underdamped limit of friction. (C) 2015 Elsevier B.V. All rights reserved.
Resumo:
The set of all subspaces of F-q(n) is denoted by P-q(n). The subspace distance d(S)(X, Y) = dim(X) + dim(Y)-2dim(X boolean AND Y) defined on P-q(n) turns it into a natural coding space for error correction in random network coding. A subset of P-q(n) is called a code and the subspaces that belong to the code are called codewords. Motivated by classical coding theory, a linear coding structure can be imposed on a subset of P-q(n). Braun et al. conjectured that the largest cardinality of a linear code, that contains F-q(n), is 2(n). In this paper, we prove this conjecture and characterize the maximal linear codes that contain F-q(n).
Resumo:
We discuss here a semiconductors assembly comprising of titanium dioxide (TiO2) rods sensitized by cadmium sulfide (CdS) nanocrystals for potential applications in large area electronics on three dimensional (3-D) substrates. Vertically aligned TiO2 rods are grown on a substrate using a 150 degrees C process flow and then sensitized with CdS by SILAR method at room temperature. This structure forms an effective photoconductor as the photo-generated electrons are rapidly removed from the CdS via the TiO2 thereby permitting a hole rich CdS. Current-voltage characteristics are measured and models illustrate space charge limited photo-current as the mechanism of charge transport at moderate voltage bias. The stable assembly and high speed are achieved. The frequency response with a loading of 10 pF and 9 M Omega shows a half power frequency of 100 Hz. (C) 2015 The Electrochemical Society. All rights reserved.
Resumo:
Space time cube representation is an information visualization technique where spatiotemporal data points are mapped into a cube. Information visualization researchers have previously argued that space time cube representation is beneficial in revealing complex spatiotemporal patterns in a data set to users. The argument is based on the fact that both time and spatial information are displayed simultaneously to users, an effect difficult to achieve in other representations. However, to our knowledge the actual usefulness of space time cube representation in conveying complex spatiotemporal patterns to users has not been empirically validated. To fill this gap, we report on a between-subjects experiment comparing novice users' error rates and response times when answering a set of questions using either space time cube or a baseline 2D representation. For some simple questions, the error rates were lower when using the baseline representation. For complex questions where the participants needed an overall understanding of the spatiotemporal structure of the data set, the space time cube representation resulted in on average twice as fast response times with no difference in error rates compared to the baseline. These results provide an empirical foundation for the hypothesis that space time cube representation benefits users analyzing complex spatiotemporal patterns.
Resumo:
Space time cube representation is an information visualization technique where spatiotemporal data points are mapped into a cube. Fast and correct analysis of such information is important in for instance geospatial and social visualization applications. Information visualization researchers have previously argued that space time cube representation is beneficial in revealing complex spatiotemporal patterns in a dataset to users. The argument is based on the fact that both time and spatial information are displayed simultaneously to users, an effect difficult to achieve in other representations. However, to our knowledge the actual usefulness of space time cube representation in conveying complex spatiotemporal patterns to users has not been empirically validated. To fill this gap we report on a between-subjects experiment comparing novice users error rates and response times when answering a set of questions using either space time cube or a baseline 2D representation. For some simple questions the error rates were lower when using the baseline representation. For complex questions where the participants needed an overall understanding of the spatiotemporal structure of the dataset, the space time cube representation resulted in on average twice as fast response times with no difference in error rates compared to the baseline. These results provide an empirical foundation for the hypothesis that space time cube representation benefits users when analyzing complex spatiotemporal patterns.
Resumo:
Resumen: La producción del Dr. Pbro. H. D. Mandrioni metafóricamente podemos decir que configura un tetraedro, un espacio establecido entre cuatro planos: Teología (propuesto como carabase), Filosofía, Psicología y Arte. Expondremos su posición acerca de la Psicología, a partir de sus textos filosóficos, poéticos, y de sus clases. Recorreremos diferentes temáticas: psicología como ciencia, problemas en el abordaje de su método y objeto. Desde allí reflexionaremos acerca de lo psíquico, su estructuración, su desarrollo u ontogénesis del yo; el proyecto vocacional y las dificultades que éste encuentra en el mundo actual, mundo de la tecnociencia, culminando con la relación entre psicología y arte, dando especial importancia a la “novela” en la formación del psicólogo.
Resumo:
209 p. : graf.
“Deborah Numbers”, Coupling Multiple Space and Time Scales and Governing Damage Evolution to Failure
Resumo:
Two different spatial levels are involved concerning damage accumulation to eventual failure. nucleation and growth rates of microdamage nN* and V*. It is found that the trans-scale length ratio c*/L does not directly affect the process. Instead, two independent dimensionless numbers: the trans-scale one * * ( V*)including the * **5 * N c V including mesoscopic parameters only, play the key role in the process of damage accumulation to failure. The above implies that there are three time scales involved in the process: the macroscopic imposed time scale tim = /a and two meso-scopic time scales, nucleation and growth of damage, (* *4) N N t =1 n c and tV=c*/V*. Clearly, the dimensionless number De*=tV/tim refers to the ratio of microdamage growth time scale over the macroscopically imposed time scale. So, analogous to the definition of Deborah number as the ratio of relaxation time over external one in rheology. Let De be the imposed Deborah number while De represents the competition and coupling between the microdamage growth and the macroscopically imposed wave loading. In stress-wave induced tensile failure (spallation) De* < 1, this means that microdamage has enough time to grow during the macroscopic wave loading. Thus, the microdamage growth appears to be the predominate mechanism governing the failure. Moreover, the dimensionless number D* = tV/tN characterizes the ratio of two intrinsic mesoscopic time scales: growth over nucleation. Similarly let D be the “intrinsic Deborah number”. Both time scales are relevant to intrinsic relaxation rather than imposed one. Furthermore, the intrinsic Deborah number D* implies a certain characteristic damage. In particular, it is derived that D* is a proper indicator of macroscopic critical damage to damage localization, like D* ∼ (10–3~10–2) in spallation. More importantly, we found that this small intrinsic Deborah number D* indicates the energy partition of microdamage dissipation over bulk plastic work. This explains why spallation can not be formulated by macroscopic energy criterion and must be treated by multi-scale analysis.
Resumo:
The concept of seismogenic asperities and aseismic barriers has become a useful paradigm within which to understand the seismogenic behavior of major faults. Since asperities and barriers can be thought of as defining the potential rupture area of large megathrust earthquakes, it is thus important to identify their respective spatial extents, constrain their temporal longevity, and to develop a physical understanding for their behavior. Space geodesy is making critical contributions to the identification of slip asperities and barriers but progress in many geographical regions depends on improving the accuracy and precision of the basic measurements. This thesis begins with technical developments aimed at improving satellite radar interferometric measurements of ground deformation whereby we introduce an empirical correction algorithm for unwanted effects due to interferometric path delays that are due to spatially and temporally variable radar wave propagation speeds in the atmosphere. In chapter 2, I combine geodetic datasets with complementary spatio-temporal resolutions to improve our understanding of the spatial distribution of crustal deformation sources and their associated temporal evolution – here we use observations from Long Valley Caldera (California) as our test bed. In the third chapter I apply the tools developed in the first two chapters to analyze postseismic deformation associated with the 2010 Mw=8.8 Maule (Chile) earthquake. The result delimits patches where afterslip occurs, explores their relationship to coseismic rupture, quantifies frictional properties associated with inferred patches of afterslip, and discusses the relationship of asperities and barriers to long-term topography. The final chapter investigates interseismic deformation of the eastern Makran subduction zone by using satellite radar interferometry only, and demonstrates that with state-of-art techniques it is possible to quantify tectonic signals with small amplitude and long wavelength. Portions of the eastern Makran for which we estimate low fault coupling correspond to areas where bathymetric features on the downgoing plate are presently subducting, whereas the region of the 1945 M=8.1 earthquake appears to be more highly coupled.
Resumo:
This thesis presents a concept for ultra-lightweight deformable mirrors based on a thin substrate of optical surface quality coated with continuous active piezopolymer layers that provide modes of actuation and shape correction. This concept eliminates any kind of stiff backing structure for the mirror surface and exploits micro-fabrication technologies to provide a tight integration of the active materials into the mirror structure, to avoid actuator print-through effects. Proof-of-concept, 10-cm-diameter mirrors with a low areal density of about 0.5 kg/m² have been designed, built and tested to measure their shape-correction performance and verify the models used for design. The low cost manufacturing scheme uses replication techniques, and strives for minimizing residual stresses that deviate the optical figure from the master mandrel. It does not require precision tolerancing, is lightweight, and is therefore potentially scalable to larger diameters for use in large, modular space telescopes. Other potential applications for such a laminate could include ground-based mirrors for solar energy collection, adaptive optics for atmospheric turbulence, laser communications, and other shape control applications.
The immediate application for these mirrors is for the Autonomous Assembly and Reconfiguration of a Space Telescope (AAReST) mission, which is a university mission under development by Caltech, the University of Surrey, and JPL. The design concept, fabrication methodology, material behaviors and measurements, mirror modeling, mounting and control electronics design, shape control experiments, predictive performance analysis, and remaining challenges are presented herein. The experiments have validated numerical models of the mirror, and the mirror models have been used within a model of the telescope in order to predict the optical performance. A demonstration of this mirror concept, along with other new telescope technologies, is planned to take place during the AAReST mission.
Resumo:
The concept of a "projection function" in a finite-dimensional real or complex normed linear space H (the function PM which carries every element into the closest element of a given subspace M) is set forth and examined.
If dim M = dim H - 1, then PM is linear. If PN is linear for all k-dimensional subspaces N, where 1 ≤ k < dim M, then PM is linear.
The projective bound Q, defined to be the supremum of the operator norm of PM for all subspaces, is in the range 1 ≤ Q < 2, and these limits are the best possible. For norms with Q = 1, PM is always linear, and a characterization of those norms is given.
If H also has an inner product (defined independently of the norm), so that a dual norm can be defined, then when PM is linear its adjoint PMH is the projection on (kernel PM)⊥ by the dual norm. The projective bounds of a norm and its dual are equal.
The notion of a pseudo-inverse F+ of a linear transformation F is extended to non-Euclidean norms. The distance from F to the set of linear transformations G of lower rank (in the sense of the operator norm ∥F - G∥) is c/∥F+∥, where c = 1 if the range of F fills its space, and 1 ≤ c < Q otherwise. The norms on both domain and range spaces have Q = 1 if and only if (F+)+ = F for every F. This condition is also sufficient to prove that we have (F+)H = (FH)+, where the latter pseudo-inverse is taken using dual norms.
In all results, the real and complex cases are handled in a completely parallel fashion.
Resumo:
The low-thrust guidance problem is defined as the minimum terminal variance (MTV) control of a space vehicle subjected to random perturbations of its trajectory. To accomplish this control task, only bounded thrust level and thrust angle deviations are allowed, and these must be calculated based solely on the information gained from noisy, partial observations of the state. In order to establish the validity of various approximations, the problem is first investigated under the idealized conditions of perfect state information and negligible dynamic errors. To check each approximate model, an algorithm is developed to facilitate the computation of the open loop trajectories for the nonlinear bang-bang system. Using the results of this phase in conjunction with the Ornstein-Uhlenbeck process as a model for the random inputs to the system, the MTV guidance problem is reformulated as a stochastic, bang-bang, optimal control problem. Since a complete analytic solution seems to be unattainable, asymptotic solutions are developed by numerical methods. However, it is shown analytically that a Kalman filter in cascade with an appropriate nonlinear MTV controller is an optimal configuration. The resulting system is simulated using the Monte Carlo technique and is compared to other guidance schemes of current interest.