29 resultados para Hardware Acceleration
Resumo:
The physical design of a VLSI circuit involves circuit partitioning as a subtask. Typically, it is necessary to partition a large electrical circuit into several smaller circuits such that the total cross-wiring is minimized. This problem is a variant of the more general graph partitioning problem, and it is known that there does not exist a polynomial time algorithm to obtain an optimal partition. The heuristic procedure proposed by Kernighan and Lin1,2 requires O(n2 log2n) time to obtain a near-optimal two-way partition of a circuit with n modules. In the VLSI context, due to the large problem size involved, this computational requirement is unacceptably high. This paper is concerned with the hardware acceleration of the Kernighan-Lin procedure on an SIMD architecture. The proposed parallel partitioning algorithm requires O(n) processors, and has a time complexity of O(n log2n). In the proposed scheme, the reduced array architecture is employed with due considerations towards cost effectiveness and VLSI realizability of the architecture.The authors are not aware of any earlier attempts to parallelize a circuit partitioning algorithm in general or the Kernighan-Lin algorithm in particular. The use of the reduced array architecture is novel and opens up the possibilities of using this computing structure for several other applications in electronic design automation.
Resumo:
Numerical Linear Algebra (NLA) kernels are at the heart of all computational problems. These kernels require hardware acceleration for increased throughput. NLA Solvers for dense and sparse matrices differ in the way the matrices are stored and operated upon although they exhibit similar computational properties. While ASIC solutions for NLA Solvers can deliver high performance, they are not scalable, and hence are not commercially viable. In this paper, we show how NLA kernels can be accelerated on REDEFINE, a scalable runtime reconfigurable hardware platform. Compared to a software implementation, Direct Solver (Modified Faddeev's algorithm) on REDEFINE shows a 29X improvement on an average and Iterative Solver (Conjugate Gradient algorithm) shows a 15-20% improvement. We further show that solution on REDEFINE is scalable over larger problem sizes without any notable degradation in performance.
Resumo:
Three new procedures - in the context of estimation of virial coefficients and summation of the partial virial series for hard discs and hard spheres - are proposed. They are based on the parametrised Euler transformation, a novel resummation, identity and the ε-convergence methods respectively. A comparison with other estimates (molecular dynamics, graph theory and empirical methods) reveals satisfactory agreement.
Resumo:
In this work an attempt has been made to evaluate the seismic hazard of South India (8.0 degrees N-20 degrees N; 72 degrees E-88 degrees E) based on the probabilistic seismic hazard analysis (PSHA). The earthquake data obtained from different sources were declustered to remove the dependent events. A total of 598 earthquakes of moment magnitude 4 and above were obtained from the study area after declustering, and were considered for further hazard analysis. The seismotectonic map of the study area was prepared by considering the faults, lineaments and the shear zones in the study area which are associated with earthquakes of magnitude 4 and above. For assessing theseismic hazard, the study area was divided into small grids of size 0.1 degrees x0.1 degrees, and the hazard parameters were calculated at the centre of each of these grid cells by considering all the seismic sources with in a radius of 300 km. Rock level peak horizontal acceleration (PHA) and spectral acceleration (SA) values at 1 corresponding to 10% and 2% probability of exceedance in 50 years have been calculated for all the grid points. The contour maps showing the spatial variation of these values are presented here. Uniform hazard response spectrum (UHRS) at rock level for 5% damping and 10% and 2% probability of exceedance in 50 years were also developed for all the grid points. The peak ground acceleration (PGA) at surface level was calculated for the entire South India for four different site classes. These values can be used to find the PGA values at any site in South India based on site class at that location. Thus, this method can be viewed as a simplified method to evaluate the PGA values at any site in the study area.
Resumo:
Three new procedures for the extrapolation of series coefficients from a given power series expansion are proposed. They are based on (i) a novel resummation identity, (ii) parametrised Euler transformation (pet) and (iii) a modifiedpet. Several examples taken from the Ising model series expansions, ferrimagnetic systems, etc., are illustrated. Apart from these applications, the higher order virial coefficients for hard spheres and hard discs have also been evaluated using the new techniques and these are compared with the estimates obtained by other methods. A satisfactory agreement is revealed between the two.
Resumo:
A simple graphical method is presented for velocity and acceleration analysis of complex mechanisms possessing low or high degree of complexity. The method is iterative in character and generally yields the solution within a few iterations. Several examples have been worked out to illustrate the method.
Resumo:
Effect of sonochemical irradiation on the conversion of 2-alkoxytetrahydrofurans to γ-butyro-1actores by Jones reagent, and its extension to the highly stereoselective synthesis of quercus lactone a, is reported.
Resumo:
Grid-connected systems when put to use at the site would experience scenarios like voltage sag, voltage swell, frequency deviations and unbalance which are common in the real world grid. When these systems are tested at laboratory, these scenarios do not exist and an almost stiff voltage source is what is usually seen. But, to qualify the grid-connected systems to operate at the site, it becomes essential to test them under the grid conditions mentioned earlier. The grid simulator is a hardware that can be programmed to generate some of the typical conditions experienced by the grid-connected systems at site. It is an inverter that is controlled to act like a voltage source in series with a grid impedance. The series grid impedance is emulated virtually within the inverter control rather than through physical components, thus avoiding the losses and the need for bulky reactive components. This paper describes the design of a grid simulator. Control implementation issues are highlighted in the experimental results.
Resumo:
A total synthesis of the bioactive tetracyclic natural product acremine G has been achieved in which a regio- and stereoselective biomimetic Diels-Alder reaction between two readily assembled building blocks, accelerated on a solid support (silica gel), forms the key step. (c) 2010 Elsevier Ltd. All rights reserved.
Resumo:
Massively parallel SIMD computing is applied to obtain an order of magnitude improvement in the executional speed of an important algorithm in VLSI design automation. The physical design of a VLSI circuit involves logic module placement as a subtask. The paper is concerned with accelerating the well known Min-cut placement technique for logic cell placement. The inherent parallelism of the Min-cut algorithm is identified, and it is shown that a parallel machine based on the efficient execution of the placement procedure.
Resumo:
Microwave irradiation, using a commercial microwave oven accelerates (in 10–15 min) the three-step ortho ester Claisen rearrangement of allyl and propynyl alcohols in dry DMF in open Erlenmeyer flasks.
Resumo:
The performance of a program will ultimately be limited by its serial (scalar) portion, as pointed out by Amdahl′s Law. Reported studies thus far of instruction-level parallelism have mixed data-parallel program portions with scalar program portions, often leading to contradictory and controversial results. We report an instruction-level behavioral characterization of scalar code containing minimal data-parallelism, extracted from highly vectorized programs of the PERFECT benchmark suite running on a Cray Y-MP system. We classify scalar basic blocks according to their instruction mix, characterize the data dependencies seen in each class, and, as a first step, measure the maximum intrablock instruction-level parallelism available. We observe skewed rather than balanced instruction distributions in scalar code and in individual basic block classes of scalar code; nonuniform distribution of parallelism across instruction classes; and, as expected, limited available intrablock parallelism. We identify frequently occurring data-dependence patterns and discuss new instructions to reduce latency. Toward effective scalar hardware, we study latency-pipelining trade-offs and restricted multiple instruction issue mechanisms.
Resumo:
The effect of fluid velocity fluctuations on the dynamics of the particles in a turbulent gas–solid suspension is analysed in the low-Reynolds-number and high Stokes number limits, where the particle relaxation time is long compared with the correlation time for the fluid velocity fluctuations, and the drag force on the particles due to the fluid can be expressed by the modified Stokes law. The direct numerical simulation procedure is used for solving the Navier–Stokes equations for the fluid, the particles are modelled as hard spheres which undergo elastic collisions and a one-way coupling algorithm is used where the force exerted by the fluid on the particles is incorporated, but not the reverse force exerted by the particles on the fluid. The particle mean and root-mean-square (RMS) fluctuating velocities, as well as the probability distribution function for the particle velocity fluctuations and the distribution of acceleration of the particles in the central region of the Couette (where the velocity profile is linear and the RMS velocities are nearly constant), are examined. It is found that the distribution of particle velocities is very different from a Gaussian, especially in the spanwise and wall-normal directions. However, the distribution of the acceleration fluctuation on the particles is found to be close to a Gaussian, though the distribution is highly anisotropic and there is a correlation between the fluctuations in the flow and gradient directions. The non-Gaussian nature of the particle velocity fluctuations is found to be due to inter-particle collisions induced by the large particle velocity fluctuations in the flow direction. It is also found that the acceleration distribution on the particles is in very good agreement with the distribution that is calculated from the velocity fluctuations in the fluid, using the Stokes drag law, indicating that there is very little correlation between the fluid velocity fluctuations and the particle velocity fluctuations in the presence of one-way coupling. All of these results indicate that the effect of the turbulent fluid velocity fluctuations can be accurately represented by an anisotropic Gaussian white noise.
Resumo:
A new technique named as model predictive spread acceleration guidance (MPSAG) is proposed in this paper. It combines nonlinear model predictive control and spread acceleration guidance philosophies. This technique is then used to design a nonlinear suboptimal guidance law for a constant speed missile against stationary target with impact angle constraint. MPSAG technique can be applied to a class of nonlinear problems, which leads to a closed form solution of the lateral acceleration (latax) history update. Guidance command assumed is the lateral acceleration (latax), applied normal to the velocity vector. The new guidance law is validated by considering the nonlinear kinematics with both lag-free as well as first order autopilot delay. The simulation results show that the proposed technique is quite promising to come up with a nonlinear guidance law that leads to both very small miss distance as well as the desired impact angle.