791 resultados para Computing clouds
Resumo:
Floquet analysis is widely used for small-order systems (say, order M < 100) to find trim results of control inputs and periodic responses, and stability results of damping levels and frequencies, Presently, however, it is practical neither for design applications nor for comprehensive analysis models that lead to large systems (M > 100); the run time on a sequential computer is simply prohibitive, Accordingly, a massively parallel Floquet analysis is developed with emphasis on large systems, and it is implemented on two SIMD or single-instruction, multiple-data computers with 4096 and 8192 processors, The focus of this development is a parallel shooting method with damped Newton iteration to generate trim results; the Floquet transition matrix (FTM) comes out as a byproduct, The eigenvalues and eigenvectors of the FTM are computed by a parallel QR method, and thereby stability results are generated, For illustration, flap and flap-lag stability of isolated rotors are treated by the parallel analysis and by a corresponding sequential analysis with the conventional shooting and QR methods; linear quasisteady airfoil aerodynamics and a finite-state three-dimensional wake model are used, Computational reliability is quantified by the condition numbers of the Jacobian matrices in Newton iteration, the condition numbers of the eigenvalues and the residual errors of the eigenpairs, and reliability figures are comparable in both the parallel and sequential analyses, Compared to the sequential analysis, the parallel analysis reduces the run time of large systems dramatically, and the reduction increases with increasing system order; this finding offers considerable promise for design and comprehensive-analysis applications.
Resumo:
In this paper, we have developed a method to compute fractal dimension (FD) of discrete time signals, in the time domain, by modifying the box-counting method. The size of the box is dependent on the sampling frequency of the signal. The number of boxes required to completely cover the signal are obtained at multiple time resolutions. The time resolutions are made coarse by decimating the signal. The loglog plot of total number of boxes required to cover the curve versus size of the box used appears to be a straight line, whose slope is taken as an estimate of FD of the signal. The results are provided to demonstrate the performance of the proposed method using parametric fractal signals. The estimation accuracy of the method is compared with that of Katz, Sevcik, and Higuchi methods. In ddition, some properties of the FD are discussed.
Resumo:
Recent studies have shown that changes in solar radiation affect the hydrological cycle more strongly than equivalent CO(2) changes for the same change in global mean surface temperature. Thus, solar radiation management ``geoengineering'' proposals to completely offset global mean temperature increases by reducing the amount of absorbed sunlight might be expected to slow the global water cycle and reduce runoff over land. However, proposed countering of global warming by increasing the albedo of marine clouds would reduce surface solar radiation only over the oceans. Here, for an idealized scenario, we analyze the response of temperature and the hydrological cycle to increased reflection by clouds over the ocean using an atmospheric general circulation model coupled to a mixed layer ocean model. When cloud droplets are reduced in size over all oceans uniformly to offset the temperature increase from a doubling of atmospheric CO(2), the global-mean precipitation and evaporation decreases by about 1.3% but runoff over land increases by 7.5% primarily due to increases over tropical land. In the model, more reflective marine clouds cool the atmospheric column over ocean. The result is a sinking motion over oceans and upward motion over land. We attribute the increased runoff over land to this increased upward motion over land when marine clouds are made more reflective. Our results suggest that, in contrast to other proposals to increase planetary albedo, offsetting mean global warming by reducing marine cloud droplet size does not necessarily lead to a drying, on average, of the continents. However, we note that the changes in precipitation, evaporation and P-E are dominated by small but significant areas, and given the highly idealized nature of this study, a more thorough and broader assessment would be required for proposals of altering marine cloud properties on a large scale.
Resumo:
Given an undirected unweighted graph G = (V, E) and an integer k ≥ 1, we consider the problem of computing the edge connectivities of all those (s, t) vertex pairs, whose edge connectivity is at most k. We present an algorithm with expected running time Õ(m + nk3) for this problem, where |V| = n and |E| = m. Our output is a weighted tree T whose nodes are the sets V1, V2,..., V l of a partition of V, with the property that the edge connectivity in G between any two vertices s ε Vi and t ε Vj, for i ≠ j, is equal to the weight of the lightest edge on the path between Vi and Vj in T. Also, two vertices s and t belong to the same Vi for any i if and only if they have an edge connectivity greater than k. Currently, the best algorithm for this problem needs to compute all-pairs min-cuts in an O(nk) edge graph; this takes Õ(m + n5/2kmin{k1/2, n1/6}) time. Our algorithm is much faster for small values of k; in fact, it is faster whenever k is o(n5/6). Our algorithm yields the useful corollary that in Õ(m + nc3) time, where c is the size of the global min-cut, we can compute the edge connectivities of all those pairs of vertices whose edge connectivity is at most αc for some constant α. We also present an Õ(m + n) Monte Carlo algorithm for the approximate version of this problem. This algorithm is applicable to weighted graphs as well. Our algorithm, with some modifications, also solves another problem called the minimum T-cut problem. Given T ⊆ V of even cardinality, we present an Õ(m + nk3) algorithm to compute a minimum cut that splits T into two odd cardinality components, where k is the size of this cut.
Resumo:
Information forms the basis of modern technology. To meet the ever-increasing demand for information, means have to be devised for a more efficient and better-equipped technology to intelligibly process data. Advances in photonics have made their impact on each of the four key applications in information processing, i.e., acquisition, transmission, storage and processing of information. The inherent advantages of ultrahigh bandwidth, high speed and low-loss transmission has already established fiber-optics as the backbone of communication technology. However, the optics to electronics inter-conversion at the transmitter and receiver ends severely limits both the speed and bit rate of lightwave communication systems. As the trend towards still faster and higher capacity systems continues, it has become increasingly necessary to perform more and more signal-processing operations in the optical domain itself, i.e., with all-optical components and devices that possess a high bandwidth and can perform parallel processing functions to eliminate the electronic bottleneck.
Resumo:
Fragment Finder 2.0 is a web-based interactive computing server which can be used to retrieve structurally similar protein fragments from 25 and 90% nonredundant data sets. The computing server identifies structurally similar fragments using the protein backbone C alpha angles. In addition, the identified fragments can be superimposed using either of the two structural superposition programs, STAMP and PROFIT, provided in the server. The freely available Java plug-in Jmol has been interfaced with the server for the visualization of the query and superposed fragments. The server is the updated version of a previously developed search engine and employs an in-house-developed fast pattern matching algorithm. This server can be accessed freely over the World Wide Web through the URL http://cluster.physics.iisc.ernet.in/ff/.
Resumo:
Carbon nanotubes dispersed in polymer matrix have been aligned in the form of fibers and interconnects and cured electrically and by UV light. Conductivity and effective semiconductor tunneling against reverse to forward bias field have been designed to have differentiable current-voltage response of each of the fiber/channel. The current-voltage response is a function of the strain applied to the fibers along axial direction. Biaxial and shear strains are correlated by differentiating signals from the aligned fibers/channels. Using a small doping of magnetic nanoparticles in these composite fibers, magneto-resistance properties are realized which are strong enough to use the resulting magnetostriction as a state variable for signal processing and computing. Various basic analog signal processing tasks such as addition, convolution and filtering etc. can be performed. These preliminary study shows promising application of the concept in combined analog-digital computation in carbon nanotube based fibers. Various dynamic effects such as relaxation, electric field dependent nonlinearities and hysteresis on the output signals are studied using experimental data and analytical model.
Resumo:
This paper is a review prepared for the second Marseille Colloquium on the mechanics of turbulence, held in 2011, 50 years after the first. The review covers recent developments in our understanding of the large-scale dynamics of cumulus cloud flows and of the atmospheric boundary layer in the low-wind convective regime that is often encountered in the tropics. It has recently been shown that a variety of cumulus cloud forms and life cycles can be experimentally realized in the laboratory, with the transient diabatic plume taken as the flow model for a cumulus cloud. The plume is subjected to diabatic heating scaled to be dynamically similar to heat release from phase changes in clouds. The experiments are complemented by exact numerical solutions of the Navier-Stokes-Boussinesq equations for plumes with scaled off-source heating. The results show that the Taylor entrainment coefficient first increases with heating, reaches a positive maximum and then drops rapidly to zero or even negative values. This reduction in entrainment is a consequence of structural changes in the flow, smoothing out the convoluted boundaries in the non-diabatic plume, including the tongues engulfing the ambient flow. This is accompanied by a greater degree of mixedness in the core flow because of lower dilution by the ambient fluid. The cloud forms generated depend strongly on the history of the diabatic heating profile in the vertical direction. The striking effects of heating on the flow are attributable to the operation of the baroclinic torque due to the temperature field. The mean baroclinic torque is shown to peak around a quasi-cylindrical sheet situated midway between the axis of the flow and the edges. This torque is shear-enhancing and folds down the engulfment tongues. The increase in mixedness can be traced to an explosive growth in the enstrophy, triggered by a strong fluctuating baroclinic torque that acts as a source, especially at the higher wave numbers, thus enhancing the mixedness. In convective boundary layers field measurements show that, under conditions prevailing in the tropics, the eddy fluxes of momentum and energy do not follow the Monin-Obukhov similarity. Instead, the eddy momentum flux is found to be linear in the wind speed at low winds; and the eddy heat flux is, to a first approximation, governed by free convection laws, with wind acting as a small perturbation on a regime of free convection. A new boundary layer code, based on heat flux scaling rather than wall-stress scaling, shows promising improvements in predictive skills of a general circulation model.
Resumo:
The Reeb graph of a scalar function tracks the evolution of the topology of its level sets. This paper describes a fast algorithm to compute the Reeb graph of a piecewise-linear (PL) function defined over manifolds and non-manifolds. The key idea in the proposed approach is to maximally leverage the efficient contour tree algorithm to compute the Reeb graph. The algorithm proceeds by dividing the input into a set of subvolumes that have loop-free Reeb graphs using the join tree of the scalar function and computes the Reeb graph by combining the contour trees of all the subvolumes. Since the key ingredient of this method is a series of union-find operations, the algorithm is fast in practice. Experimental results demonstrate that it outperforms current generic algorithms by a factor of up to two orders of magnitude, and has a performance on par with algorithms that are catered to restricted classes of input. The algorithm also extends to handle large data that do not fit in memory.