991 resultados para Computational sciences
Resumo:
Computations have been carried out for simulating supersonic flow through a set of converging-diverging nozzles with their expanding jets forming a laser cavity and flow patterns through diffusers, past the cavity. A thorough numerical investigation with 3-D RANS code is carried out to capture the flow distribution which comprises of shock patterns and multiple supersonic jet interactions. The analysis of pressure recovery characteristics during the flow through the diffusers is an important parameter of the simulation and is critical for the performance of the laser device. The results of the computation have shown a close agreement with the experimentally measured parameters as well as other established results indicating that the flow analysis done is found to be satisfactory.
Resumo:
Long running multi-physics coupled parallel applications have gained prominence in recent years. The high computational requirements and long durations of simulations of these applications necessitate the use of multiple systems of a Grid for execution. In this paper, we have built an adaptive middleware framework for execution of long running multi-physics coupled applications across multiple batch systems of a Grid. Our framework, apart from coordinating the executions of the component jobs of an application on different batch systems, also automatically resubmits the jobs multiple times to the batch queues to continue and sustain long running executions. As the set of active batch systems available for execution changes, our framework performs migration and rescheduling of components using a robust rescheduling decision algorithm. We have used our framework for improving the application throughput of a foremost long running multi-component application for climate modeling, the Community Climate System Model (CCSM). Our real multi-site experiments with CCSM indicate that Grid executions can lead to improved application throughput for climate models.
Resumo:
The work reported here is concerned with a detailed thermochemical evaluation of the flaming mode behaviour of a gasifier based stove. Determination of the gas composition over the fuel bed, surface and gas temperatures in the gasification process constitute principal experimental features. A simple atomic balance for the gasification reaction combined with the gas composition from the experiments is used to determine the CH(4) equivalent of higher hydrocarbons and the gasification efficiency (eta g). The components of utilization efficiency, namely, gasification-combustion and heat transfer are explored. Reactive flow computational studies using the measured gas composition over the fuel bed are used to simulate the thermochemical flow field and heat transfer to the vessel; hither-to-ignored vessel size effects in the extraction of heat from the stove are established clearly. The overall flaming mode efficiency of the stove is 50-54%; the convective and radiative components of heat transfer are established to be 45-47 and 5-7% respectively. The efficiency estimates from reacting computational fluid dynamics (RCFD) compare well with experiments. (C) 2011 Elsevier Ltd. All rights reserved.
Resumo:
A new structured discretization of 2D space, named X-discretization, is proposed to solve bivariate population balance equations using the framework of minimal internal consistency of discretization of Chakraborty and Kumar [2007, A new framework for solution of multidimensional population balance equations. Chem. Eng. Sci. 62, 4112-4125] for breakup and aggregation of particles. The 2D space of particle constituents (internal attributes) is discretized into bins by using arbitrarily spaced constant composition radial lines and constant mass lines of slope -1. The quadrilaterals are triangulated by using straight lines pointing towards the mean composition line. The monotonicity of the new discretization makes is quite easy to implement, like a rectangular grid but with significantly reduced numerical dispersion. We use the new discretization of space to automate the expansion and contraction of the computational domain for the aggregation process, corresponding to the formation of larger particles and the disappearance of smaller particles by adding and removing the constant mass lines at the boundaries. The results show that the predictions of particle size distribution on fixed X-grid are in better agreement with the analytical solution than those obtained with the earlier techniques. The simulations carried out with expansion and/or contraction of the computational domain as population evolves show that the proposed strategy of evolving the computational domain with the aggregation process brings down the computational effort quite substantially; larger the extent of evolution, greater is the reduction in computational effort. (C) 2011 Elsevier Ltd. All rights reserved.
Resumo:
The DMS-FEM, which enables functional approximations with C(1) or still higher inter-element continuity within an FEM-based meshing of the domain, has recently been proposed by Sunilkumar and Roy [39,40]. Through numerical explorations on linear elasto-static problems, the method was found to have conspicuously superior convergence characteristics as well as higher numerical stability against locking. These observations motivate the present study, which aims at extending and exploring the DMS-FEM to (geometrically) nonlinear elasto-static problems of interest in solid mechanics and assessing its numerical performance vis-a-vis the FEM. In particular, the DMS-FEM is shown to vastly outperform the FEM (presently implemented through the commercial software ANSYS (R)) as the former requires fewer linearization and load steps to achieve convergence. In addition, in the context of nearly incompressible nonlinear systems prone to volumetric locking and with no special numerical artefacts (e.g. stabilized or mixed weak forms) employed to arrest locking, the DMS-FEM is shown to approach the incompressibility limit much more closely and with significantly fewer iterations than the FEM. The numerical findings are suggestive of the important role that higher order (uniform) continuity of the approximated field variables play in overcoming volumetric locking and the great promise that the method holds for a range of other numerically ill-conditioned problems of interest in computational structural mechanics. (C) 2011 Elsevier Ltd. All rights reserved.
Resumo:
Current scientific research is characterized by increasing specialization, accumulating knowledge at a high speed due to parallel advances in a multitude of sub-disciplines. Recent estimates suggest that human knowledge doubles every two to three years – and with the advances in information and communication technologies, this wide body of scientific knowledge is available to anyone, anywhere, anytime. This may also be referred to as ambient intelligence – an environment characterized by plentiful and available knowledge. The bottleneck in utilizing this knowledge for specific applications is not accessing but assimilating the information and transforming it to suit the needs for a specific application. The increasingly specialized areas of scientific research often have the common goal of converting data into insight allowing the identification of solutions to scientific problems. Due to this common goal, there are strong parallels between different areas of applications that can be exploited and used to cross-fertilize different disciplines. For example, the same fundamental statistical methods are used extensively in speech and language processing, in materials science applications, in visual processing and in biomedicine. Each sub-discipline has found its own specialized methodologies making these statistical methods successful to the given application. The unification of specialized areas is possible because many different problems can share strong analogies, making the theories developed for one problem applicable to other areas of research. It is the goal of this paper to demonstrate the utility of merging two disparate areas of applications to advance scientific research. The merging process requires cross-disciplinary collaboration to allow maximal exploitation of advances in one sub-discipline for that of another. We will demonstrate this general concept with the specific example of merging language technologies and computational biology.
Resumo:
In this paper we have developed methods to compute maps from differential equations. We take two examples. First is the case of the harmonic oscillator and the second is the case of Duffing's equation. First we convert these equations to a canonical form. This is slightly nontrivial for the Duffing's equation. Then we show a method to extend these differential equations. In the second case, symbolic algebra needs to be used. Once the extensions are accomplished, various maps are generated. The Poincare sections are seen as a special case of such generated maps. Other applications are also discussed.
Resumo:
In this paper, the diversity-multiplexing gain tradeoff (DMT) of single-source, single-sink (ss-ss), multihop relay networks having slow-fading links is studied. In particular, the two end-points of the DMT of ss-ss full-duplex networks are determined, by showing that the maximum achievable diversity gain is equal to the min-cut and that the maximum multiplexing gain is equal to the min-cut rank, the latter by using an operational connection to a deterministic network. Also included in the paper, are several results that aid in the computation of the DMT of networks operating under amplify-and-forward (AF) protocols. In particular, it is shown that the colored noise encountered in amplify-and-forward protocols can be treated as white for the purpose of DMT computation, lower bounds on the DMT of lower-triangular channel matrices are derived and the DMT of parallel MIMO channels is computed. All protocols appearing in the paper are explicit and rely only upon AF relaying. Half-duplex networks and explicit coding schemes are studied in a companion paper.
Resumo:
A variable resolution global spectral method is created on the sphere using High resolution Tropical Belt Transformation (HTBT). HTBT belongs to a class of map called reparametrisation maps. HTBT parametrisation of the sphere generates a clustering of points in the entire tropical belt; the density of the grid point distribution decreases smoothly in the domain outside the tropics. This variable resolution method creates finer resolution in the tropics and coarser resolution at the poles. The use of FFT procedure and Gaussian quadrature for the spectral computations retains the numerical efficiency available with the standard global spectral method. Accuracy of the method for meteorological computations are demonstrated by solving Helmholtz equation and non-divergent barotropic vorticity equation on the sphere. (C) 2011 Elsevier Inc. All rights reserved.
Resumo:
We have postulated a novel pathway that could assist in the nucleation of soot particles through covalent dimerization and oligomerizations of a variety of PAHs. DFT calculations were performed with the objective of obtaining the relative thermal stabilities and formation probabilities of oligomeric species that exploit the facile dimerization that is known to occur in linear oligoacenes. We propose that the presence of small stretches of linear oligoacence (tetracene or longer) in extended PAH, either embedded or tethered, would be adequate for enabling the formation of such dimeric and oligomeric adducts; these could then serve as nuclei for the growth of soot particles. Our studies also reveal the importance of p-stacking interactions between extended aromatic frameworks in governing the relative stabilities of the oligomeric species that are formed. (c) 2012 Wiley Periodicals, Inc.
Resumo:
The diffusion equation-based modeling of near infrared light propagation in tissue is achieved by using finite-element mesh for imaging real-tissue types, such as breast and brain. The finite-element mesh size (number of nodes) dictates the parameter space in the optical tomographic imaging. Most commonly used finite-element meshing algorithms do not provide the flexibility of distinct nodal spacing in different regions of imaging domain to take the sensitivity of the problem into consideration. This study aims to present a computationally efficient mesh simplification method that can be used as a preprocessing step to iterative image reconstruction, where the finite-element mesh is simplified by using an edge collapsing algorithm to reduce the parameter space at regions where the sensitivity of the problem is relatively low. It is shown, using simulations and experimental phantom data for simple meshes/domains, that a significant reduction in parameter space could be achieved without compromising on the reconstructed image quality. The maximum errors observed by using the simplified meshes were less than 0.27% in the forward problem and 5% for inverse problem.