50 resultados para Velocity Measurements
em Consorci de Serveis Universitaris de Catalunya (CSUC), Spain
Resumo:
Recently a fingering morphology, resembling the hydrodynamic Saffman-Taylor instability, was identified in the quasi-two-dimensional electrodeposition of copper. We present here measurements of the dispersion relation of the growing front. The instability is accompanied by gravity-driven convection rolls at the electrodes, which are examined using particle image velocimetry. While at the anode the theory presented by Chazalviel et al. [J. Electroanal. Chem. 407, 61 (1996)] describes the convection roll, the flow field at the cathode is more complicated because of the growing deposit. In particular, the analysis of the orientation of the velocity vectors reveals some lag of the development of the convection roll compared to the finger envelope.
Resumo:
We present the dynamic velocity profiles of a Newtonian fluid (glycerol) and a viscoelastic Maxwell fluid (CPyCl-NaSal in water) driven by an oscillating pressure gradient in a vertical cylindrical pipe. The frequency range explored has been chosen to include the first three resonance peaks of the dynamic permeability of the viscoelastic-fluid¿pipe system. Three different optical measurement techniques have been employed. Laser Doppler anemometry has been used to measure the magnitude of the velocity at the center of the liquid column. Particle image velocimetry and optical deflectometry are used to determine the velocity profiles at the bulk of the liquid column and at the liquid-air interface respectively. The velocity measurements in the bulk are in good agreement with the theoretical predictions of a linear theory. The results, however, show dramatic differences in the dynamic behavior of Newtonian and viscoelastic fluids, and demonstrate the importance of resonance phenomena in viscoelastic fluid flows, biofluids in particular, in confined geometries.
Resumo:
The European Space Agency's Gaia mission will create the largest and most precise three dimensional chart of our galaxy (the Milky Way), by providing unprecedented position, parallax, proper motion, and radial velocity measurements for about one billion stars. The resulting catalogue will be made available to the scientific community and will be analyzed in many different ways, including the production of a variety of statistics. The latter will often entail the generation of multidimensional histograms and hypercubes as part of the precomputed statistics for each data release, or for scientific analysis involving either the final data products or the raw data coming from the satellite instruments. In this paper we present and analyze a generic framework that allows the hypercube generation to be easily done within a MapReduce infrastructure, providing all the advantages of the new Big Data analysis paradigmbut without dealing with any specific interface to the lower level distributed system implementation (Hadoop). Furthermore, we show how executing the framework for different data storage model configurations (i.e. row or column oriented) and compression techniques can considerably improve the response time of this type of workload for the currently available simulated data of the mission. In addition, we put forward the advantages and shortcomings of the deployment of the framework on a public cloud provider, benchmark against other popular solutions available (that are not always the best for such ad-hoc applications), and describe some user experiences with the framework, which was employed for a number of dedicated astronomical data analysis techniques workshops.
Resumo:
The European Space Agency's Gaia mission will create the largest and most precise three dimensional chart of our galaxy (the Milky Way), by providing unprecedented position, parallax, proper motion, and radial velocity measurements for about one billion stars. The resulting catalogue will be made available to the scientific community and will be analyzed in many different ways, including the production of a variety of statistics. The latter will often entail the generation of multidimensional histograms and hypercubes as part of the precomputed statistics for each data release, or for scientific analysis involving either the final data products or the raw data coming from the satellite instruments. In this paper we present and analyze a generic framework that allows the hypercube generation to be easily done within a MapReduce infrastructure, providing all the advantages of the new Big Data analysis paradigmbut without dealing with any specific interface to the lower level distributed system implementation (Hadoop). Furthermore, we show how executing the framework for different data storage model configurations (i.e. row or column oriented) and compression techniques can considerably improve the response time of this type of workload for the currently available simulated data of the mission. In addition, we put forward the advantages and shortcomings of the deployment of the framework on a public cloud provider, benchmark against other popular solutions available (that are not always the best for such ad-hoc applications), and describe some user experiences with the framework, which was employed for a number of dedicated astronomical data analysis techniques workshops.
Resumo:
Aortic stiffness is an independent predictor factor for cardiovascular risk. Different methods for determining pulse wave velocity (PWV) are used, among which the most common are mechanical methods such as SphygmoCor or Complior, which require specific devices and are limited by technical difficulty in obtaining measurements. Doppler guided by 2D ultrasound is a good alternative to these methods. We studied 40 patients (29 male, aged 21 to 82 years) comparing the Complior method with Doppler. Agreement of both devices was high (R = 0.91, 0.84-0.95, 95% CI). The reproducibility analysis revealed no intra-nor interobserver differences. Based on these results, we conclude that Doppler ultrasound is a reliable and reproducible alternative to other established methods for themeasurement of aortic PWV
Resumo:
Diffuse flow velocimetry (DFV) is introduced as a new, noninvasive, optical technique for measuring the velocity of diffuse hydrothermal flow. The technique uses images of a motionless, random medium (e.g.,rocks) obtained through the lens of a moving refraction index anomaly (e.g., a hot upwelling). The method works in two stages. First, the changes in apparent background deformation are calculated using particle image velocimetry (PIV). The deformation vectors are determined by a cross correlation of pixel intensities across consecutive images. Second, the 2-D velocity field is calculated by cross correlating the deformation vectors between consecutive PIV calculations. The accuracy of the method is tested with laboratory and numerical experiments of a laminar, axisymmetric plume in fluids with both constant and temperaturedependent viscosity. Results show that average RMS errors are ∼5%–7% and are most accurate in regions of pervasive apparent background deformation which is commonly encountered in regions of diffuse hydrothermal flow. The method is applied to a 25 s video sequence of diffuse flow from a small fracture captured during the Bathyluck’09 cruise to the Lucky Strike hydrothermal field (September 2009). The velocities of the ∼10°C–15°C effluent reach ∼5.5 cm/s, in strong agreement with previous measurements of diffuse flow. DFV is found to be most accurate for approximately 2‐D flows where background objects have a small spatial scale, such as sand or gravel
Resumo:
Velocity has been measured as a function of time for propagating crack tips as water is injected into solutions of end-capped associating polymers in a rectanguar Hele-Shaw cell. Measurements were performed for flows with different values of cell gap, channel width, polymer molecular weight, and polymer concentration. The condition for the onset of fracturelike behavior is well described by a Deborah number which uses the shear-thinning shear rate of the polymer solution as a characteristic frequency for network relaxation. At low molecular weight, the onset of fracturelike pattern evolution is accompanied by an abrupt jump in tip velocity, followed by a lower and approximately constant acceleration. At high molecular weight, the transition to fracturelike behavior involves passing through a regime that may be understood in terms of stick-slip dynamics. The crack-tip wanders from side to side and fluctuates (in both speed and velocity along the channel) with a characteristic frequency which depends linearly on the invading fluid injection rate.
Resumo:
We study fracturelike flow instabilities that arise when water is injected into a Hele-Shaw cell filled with aqueous solutions of associating polymers. We explore various polymer architectures, molecular weights, and solution concentrations. Simultaneous measurements of the finger tip velocity and of the pressure at the injection point allow us to describe the dynamics of the finger in terms of the finger mobility, which relates the velocity to the pressure gradient. The flow discontinuities, characterized by jumps in the finger tip velocity, which are observed in experiments with some of the polymer solutions, can be modeled by using a nonmonotonic dependence between a characteristic shear stress and the shear rate at the tip of the finger. A simple model, which is based on a viscosity function containing both a Newtonian and a non-Newtonian component, and which predicts nonmonotonic regions when the non-Newtonian component of the viscosity dominates, is shown to agree with the experimental data.
Resumo:
This paper proposes the use of an autonomous assistant mobile robot in order to monitor the environmental conditions of a large indoor area and develop an ambient intelligence application. The mobile robot uses single high performance embedded sensors in order to collect and geo-reference environmental information such as ambient temperature, air velocity and orientation and gas concentration. The data collected with the assistant mobile robot is analyzed in order to detect unusual measurements or discrepancies and develop focused corrective ambient actions. This paper shows an example of the measurements performed in a research facility which have enabled the detection and location of an uncomfortable temperature profile inside an office of the research facility. The ambient intelligent application has been developed by performing some localized ambient measurements that have been analyzed in order to propose some ambient actuations to correct the uncomfortable temperature profile.
Resumo:
This paper proposes the use of an autonomous assistant mobile robot in order to monitor the environmental conditions of a large indoor area and develop an ambient intelligence application. The mobile robot uses single high performance embedded sensors in order to collect and geo-reference environmental information such as ambient temperature, air velocity and orientation and gas concentration. The data collected with the assistant mobile robot is analyzed in order to detect unusual measurements or discrepancies and develop focused corrective ambient actions. This paper shows an example of the measurements performed in a research facility which have enabled the detection and location of an uncomfortable temperature profile inside an office of the research facility. The ambient intelligent application has been developed by performing some localized ambient measurements that have been analyzed in order to propose some ambient actuations to correct the uncomfortable temperature profile.
Resumo:
This paper presents a general equilibrium model of money demand where the velocity of money changes in response to endogenous fluctuations in the interest rate. The parameter space can be divided into two subsets: one where velocity is constant as in standard cash-in-advance models, and another one where velocity fluctuates as in Baumol (1952). The model provides an explanation of why, for a sample of 79 countries, the correlation between the velocity of money and the inflation rate appears to be low, unlike common wisdom would suggest. The reason is the diverse transaction technologies available in different economies.
Resumo:
TCP flows from applications such as the web or ftp are well supported by a Guaranteed Minimum Throughput Service (GMTS), which provides a minimum network throughput to the flow and, if possible, an extra throughput. We propose a scheme for a GMTS using Admission Control (AC) that is able to provide different minimum throughput to different users and that is suitable for "standard" TCP flows. Moreover, we consider a multidomain scenario where the scheme is used in one of the domains, and we propose some mechanisms for the interconnection with neighbor domains. The whole scheme uses a small set of packet classes in a core-stateless network where each class has a different discarding priority in queues assigned to it. The AC method involves only edge nodes and uses a special probing packet flow (marked as the highest discarding priority class) that is sent continuously from ingress to egress through a path. The available throughput in the path is obtained at the egress using measurements of flow aggregates, and then it is sent back to the ingress. At the ingress each flow is detected using an implicit way and then it is admission controlled. If it is accepted, it receives the GMTS and its packets are marked as the lowest discarding priority classes; otherwise, it receives a best-effort service. The scheme is evaluated through simulation in a simple "bottleneck" topology using different traffic loads consisting of "standard" TCP flows that carry files of varying sizes
Resumo:
This paper proposes MSISpIC, a probabilistic sonar scan matching algorithm for the localization of an autonomous underwater vehicle (AUV). The technique uses range scans gathered with a Mechanical Scanning Imaging Sonar (MSIS), the robot displacement estimated through dead-reckoning using a Doppler velocity log (DVL) and a motion reference unit (MRU). The proposed method is an extension of the pIC algorithm. An extended Kalman filter (EKF) is used to estimate the robot-path during the scan in order to reference all the range and bearing measurements as well as their uncertainty to a scan fixed frame before registering. The major contribution consists of experimentally proving that probabilistic sonar scan matching techniques have the potential to improve the DVL-based navigation. The algorithm has been tested on an AUV guided along a 600 m path within an abandoned marina underwater environment with satisfactory results
Resumo:
This paper presents the implementation details of a coded structured light system for rapid shape acquisition of unknown surfaces. Such techniques are based on the projection of patterns onto a measuring surface and grabbing images of every projection with a camera. Analyzing the pattern deformations that appear in the images, 3D information of the surface can be calculated. The implemented technique projects a unique pattern so that it can be used to measure moving surfaces. The structure of the pattern is a grid where the color of the slits are selected using a De Bruijn sequence. Moreover, since both axis of the pattern are coded, the cross points of the grid have two codewords (which permits to reconstruct them very precisely), while pixels belonging to horizontal and vertical slits have also a codeword. Different sets of colors are used for horizontal and vertical slits, so the resulting pattern is invariant to rotation. Therefore, the alignment constraint between camera and projector considered by a lot of authors is not necessary