976 resultados para locality preserving projections


Relevância:

10.00% 10.00%

Publicador:

Resumo:

Polynomial chaos expansion (PCE) with Latin hypercube sampling (LHS) is employed for calculating the vibrational frequencies of an inviscid incompressible fluid partially filled in a rectangular tank with and without a baffle. Vibration frequencies of the coupled system are described through their projections on the PCE which uses orthogonal basis functions. PCE coefficients are evaluated using LHS. Convergence on the coefficient of variation is used to find the orthogonal polynomial basis function order which is employed in PCE. It is observed that the dispersion in the eigenvalues is more in the case of a rectangular tank with a baffle. The accuracy of the PCE method is verified with standard MCS results and is found to be more efficient.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

An assessment of the impact of projected climate change on forest ecosystems in India based on climate projections of the Regional Climate Model of the Hadley Centre (HadRM3) and the global dynamic vegetation model IBIS for A1B scenario is conducted for short-term (2021-2050) and long-term (2071-2100) periods. Based on the dynamic global vegetation modelling, vulnerable forested regions of India have been identified to assist in planning adaptation interventions. The assessment of climate impacts showed that at the national level, about 45% of the forested grids is projected to undergo change. Vulnerability assessment showed that such vulnerable forested grids are spread across India. However, their concentration is higher in the upper Himalayan stretches, parts of Central India, northern Western Ghats and the Eastern Ghats. In contrast, the northeastern forests, southern Western Ghats and the forested regions of eastern India are estimated to be the least vulnerable. Low tree density, low biodiversity status as well as higher levels of fragmentation, in addition to climate change, contribute to the vulnerability of these forests. The mountainous forests (sub-alpine and alpine forest, the Himalayan dry temperate forest and the Himalayan moist temperate forest) are susceptible to the adverse effects of climate change. This is because climate change is predicted to be larger for regions that have greater elevations.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

This paper is concerned with the dynamic analysis of flexible,non-linear multi-body beam systems. The focus is on problems where the strains within each elastic body (beam) remain small. Based on geometrically non-linear elasticity theory, the non-linear 3-D beam problem splits into either a linear or non-linear 2-D analysis of the beam cross-section and a non-linear 1-D analysis along the beam reference line. The splitting of the three-dimensional beam problem into two- and one-dimensional parts, called dimensional reduction,results in a tremendous savings of computational effort relative to the cost of three-dimensional finite element analysis,the only alternative for realistic beams. The analysis of beam-like structures made of laminated composite materials requires a much more complicated methodology. Hence, the analysis procedure based on Variational Asymptotic Method (VAM), a tool to carry out the dimensional reduction, is used here.The analysis methodology can be viewed as a 3-step procedure. First, the sectional properties of beams made of composite materials are determined either based on an asymptotic procedure that involves a 2-D finite element nonlinear analysis of the beam cross-section to capture trapeze effect or using strip-like beam analysis, starting from Classical Laminated Shell Theory (CLST). Second, the dynamic response of non-linear, flexible multi-body beam systems is simulated within the framework of energy-preserving and energy-decaying time integration schemes that provide unconditional stability for non-linear beam systems. Finally,local 3-D responses in the beams are recovered, based on the 1-D responses predicted in the second step. Numerical examples are presented and results from this analysis are compared with those available in the literature.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

This work intends to demonstrate the importance of geometrically nonlinear crosssectional analysis of certain composite beam-based four-bar mechanisms in predicting system dynamic characteristics. All component bars of the mechanism are made of fiber reinforced laminates and have thin rectangular cross-sections. They could, in general, be pre-twisted and/or possess initial curvature, either by design or by defect. They are linked to each other by means of revolute joints. We restrict ourselves to linear materials with small strains within each elastic body (beam). Each component of the mechanism is modeled as a beam based on geometrically nonlinear 3-D elasticity theory. The component problems are thus split into 2-D analyses of reference beam cross-sections and nonlinear 1-D analyses along the four beam reference curves. For thin rectangular cross-sections considered here, the 2-D cross-sectional nonlinearity is overwhelming. This can be perceived from the fact that such sections constitute a limiting case between thin-walled open and closed sections, thus inviting the nonlinear phenomena observed in both. The strong elastic couplings of anisotropic composite laminates complicate the model further. However, a powerful mathematical tool called the Variational Asymptotic Method (VAM) not only enables such a dimensional reduction, but also provides asymptotically correct analytical solutions to the nonlinear cross-sectional analysis. Such closed-form solutions are used here in conjunction with numerical techniques for the rest of the problem to predict multi-body dynamic responses, more quickly and accurately than would otherwise be possible. The analysis methodology can be viewed as a three-step procedure: First, the cross-sectional properties of each bar of the mechanism is determined analytically based on an asymptotic procedure, starting from Classical Laminated Shell Theory (CLST) and taking advantage of its thin strip geometry. Second, the dynamic response of the nonlinear, flexible fourbar mechanism is simulated by treating each bar as a 1-D beam, discretized using finite elements, and employing energy-preserving and -decaying time integration schemes for unconditional stability. Finally, local 3-D deformations and stresses in the entire system are recovered, based on the 1-D responses predicted in the previous step. With the model, tools and procedure in place, we shall attempt to identify and investigate a few problems where the cross-sectional nonlinearities are significant. This will be carried out by varying stacking sequences and material properties, and speculating on the dominating diagonal and coupling terms in the closed-form nonlinear beam stiffness matrix. Numerical examples will be presented and results from this analysis will be compared with those available in the literature, for linear cross-sectional analysis and isotropic materials as special cases.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Gottigere lake with a water spread area of about 14.98 ha is located in the Bellandur Lake catchment of the South Pennar River basin. In recent years, this lake catchment has been subjected to environmental stress mainly due to the rampant unplanned developmental activities in the catchment. The functional ability of the ecosystem is impaired due to structural changes in the ecosystem. This is evident from poor water quality, breeding of disease vectors, contamination of groundwater in the catchment, frequent flooding in the catchment due to topography alteration, decline in groundwater table, erosion in lake bed, etc. The development plans of the region (current as well as the proposed) ignore the integrated planning approaches considering all components of the ecosystem. Serious threats to the sustainability of the region due to lack of holistic approaches in aquatic resources management are land use changes (removal of vegetation cover, etc.), point and non-point sources of pollution impairing water quality, dumping of solid waste (building waste, etc.). Conservation of lake ecosystem is possible only when the physical and chemical integrity of its catchment is maintained. Alteration in the catchment either due to land use changes (leading to paved surface area from vegetation cover), alteration in topography, construction of roads in the immediate vicinity are detrimental to water yield in the catchment and hence, the sustenance of the lake. Open spaces in the form of lakes and parks aid as kidney and lung in an urban ecosystem, which maintain the health of the people residing in the locality. Identification of core buffer zones and conservation of buffer zones (500 to 1000 m from shore) is to be taken up on priority for conservation and sustainable management of Bangalore lakes. Bangalore is located over a ridge delineating four watersheds, viz. Hebbal, Koramangala, Challaghatta and Vrishabhavathi. Lakes and tanks are an integral part of natural drainage and help in retaining water during rainfall, which otherwise get drained off as flash floods. Each lake harvests rainwater from its catchment and surplus flows downstream spilling into the next lake in the chain. The topography of Bangalore has uniquely supported the creation of a large number of lakes. These lakes form chains, being a series of impoundments across streams. This emphasises the interconnectivity among Bangalore lakes, which has to be retained to prevent Bangalore from flooding or from water scarcity. The main source of replenishment of groundwater is the rainfall. The slope of the terrain allows most of the rainwater to flow as run-off. With the steep gradients available in the major valleys of Bangalore, the rainwater will flow out of the city within four to five hours. Only a small fraction of the rainwater infiltrates into the soil. The infiltration of water into the subsoil has declined with more and more buildings and paved road being constructed in the city. Thus the natural drainage of Bangalore is governed by flows from the central ridge to all lower contours and is connected with various tanks and ponds. There are no major rivers flowing in Bangalore and there is an urgent need to sustain these vital ecosystems through proper conservation and management measures. The proposed peripheral ring road connecting Hosur Road (NH 7) and Mysore Road (SH 17) at Gottigere lake falls within the buffer zone of the lake. This would alter the catchment integrity and hence water yield affecting flora, fauna and local people, and ultimately lead to the disappearance of Gottigere lake. Developmental activities in lake catchments, which has altered lake’s ecological integrity is in violation of the Indian Fisheries Act – 1857, the Indian Forest Act – 1927, Wildlife (Protection) Act – 1972, Water (Prevention and Control of Pollution) Act – 1974, Water (Prevention and Control of Pollution) Act – 1977, Forest (Conservation Act) – 1980, Environmental (Protection) Act – 1986, Wildlife (Protection) Amendment Act – 1991 and National Conservation Strategy and Policy Statement on Environment and Development – 1992. Considering 65% decline of waterbodies in Bangalore (during last three decades), decision makers should immediately take preventive measures to ensure that lake ecosystems are not affected. This report discusses the impacts due to the proposed infrastructure developmental activities in the vicinity of Gottigere tank.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

As the gap between processor and memory continues to grow Memory performance becomes a key performance bottleneck for many applications. Compilers therefore increasingly seek to modify an application’s data layout to improve cache locality and cache reuse. Whole program Structure Layout [WPSL] transformations can significantly increase the spatial locality of data and reduce the runtime of programs that use link-based data structures, by increasing the cache line utilization. However, in production compilers WPSL transformations do not realize the entire performance potential possible due to a number of factors. Structure layout decisions made on the basis of whole program aggregated affinity/hotness of structure fields, can be sub optimal for local code regions. WPSL is also restricted in applicability in production compilers for type unsafe languages like C/C++ due to the extensive legality checks and field sensitive pointer analysis required over the entire application. In order to overcome the issues associated with WPSL, we propose Region Based Structure Layout (RBSL) optimization framework, using selective data copying. We describe our RBSL framework, implemented in the production compiler for C/C++ on HP-UX IA-64. We show that acting in complement to the existing and mature WPSL transformation framework in our compiler, RBSL improves application performance in pointer intensive SPEC benchmarks ranging from 3% to 28% over WPSL

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Pixel based image fusion entails combining geometric details of a high-resolution Panchromatic (PAN) image and spectral information of a low-resolution Multispectral (MS) image to produce images with highest spatial content while preserving the spectral information. This work reviews and implements six fusion techniques – À Trous algorithm based wavelet transform (ATW), Mulitresolution Analysis based Intensity Modulation, Gram Schmidt fusion, CN Spectral, Luminance Chrominance and High pass fusion (HPF) on IKONOS imagery having 1 m PAN and 4 m MS channels. Comparative performance analysis of techniques by various methods reveals that ATW followed by HPF perform best among all the techniques.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

We propose a randomized algorithm for large scale SVM learning which solves the problem by iterating over random subsets of the data. Crucial to the algorithm for scalability is the size of the subsets chosen. In the context of text classification we show that, by using ideas from random projections, a sample size of O(log n) can be used to obtain a solution which is close to the optimal with a high probability. Experiments done on synthetic and real life data sets demonstrate that the algorithm scales up SVM learners, without loss in accuracy. 1

Relevância:

10.00% 10.00%

Publicador:

Resumo:

In general the objective of accurately encoding the input data and the objective of extracting good features to facilitate classification are not consistent with each other. As a result, good encoding methods may not be effective mechanisms for classification. In this paper, an earlier proposed unsupervised feature extraction mechanism for pattern classification has been extended to obtain an invertible map. The method of bimodal projection-based features was inspired by the general class of methods called projection pursuit. The principle of projection pursuit concentrates on projections that discriminate between clusters and not faithful representations. The basic feature map obtained by the method of bimodal projections has been extended to overcome this. The extended feature map is an embedding of the input space in the feature space. As a result, the inverse map exists and hence the representation of the input space in the feature space is exact. This map can be naturally expressed as a feedforward neural network.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

In this paper, we address the reconstruction problem from laterally truncated helical cone-beam projections. The reconstruction problem from lateral truncation, though similar to that of interior radon problem, is slightly different from it as well as the local (lambda) tomography and pseudo-local tomography in the sense that we aim to reconstruct the entire object being scanned from a region-of-interest (ROI) scan data. The method proposed in this paper is a projection data completion approach followed by the use of any standard accurate FBP type reconstruction algorithm. In particular, we explore a windowed linear prediction (WLP) approach for data completion and compare the quality of reconstruction with the linear prediction (LP) technique proposed earlier.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The inherent temporal locality in memory accesses is filtered out by the L1 cache. As a consequence, an L2 cache with LRU replacement incurs significantly higher misses than the optimal replacement policy (OPT). We propose to narrow this gap through a novel replacement strategy that mimics the replacement decisions of OPT. The L2 cache is logically divided into two components, a Shepherd Cache (SC) with a simple FIFO replacement and a Main Cache (MC) with an emulation of optimal replacement. The SC plays the dual role of caching lines and guiding the replacement decisions in MC. Our pro- posed organization can cover 40% of the gap between OPT and LRU for a 2MB cache resulting in 7% overall speedup. Comparison with the dynamic insertion policy, a victim buffer, a V-Way cache and an LRU based fully associative cache demonstrates that our scheme performs better than all these strategies.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

We present two efficient discrete parameter simulation optimization (DPSO) algorithms for the long-run average cost objective. One of these algorithms uses the smoothed functional approximation (SFA) procedure, while the other is based on simultaneous perturbation stochastic approximation (SPSA). The use of SFA for DPSO had not been proposed previously in the literature. Further, both algorithms adopt an interesting technique of random projections that we present here for the first time. We give a proof of convergence of our algorithms. Next, we present detailed numerical experiments on a problem of admission control with dependent service times. We consider two different settings involving parameter sets that have moderate and large sizes, respectively. On the first setting, we also show performance comparisons with the well-studied optimal computing budget allocation (OCBA) algorithm and also the equal allocation algorithm. Note to Practitioners-Even though SPSA and SFA have been devised in the literature for continuous optimization problems, our results indicate that they can be powerful techniques even when they are adapted to discrete optimization settings. OCBA is widely recognized as one of the most powerful methods for discrete optimization when the parameter sets are of small or moderate size. On a setting involving a parameter set of size 100, we observe that when the computing budget is small, both SPSA and OCBA show similar performance and are better in comparison to SFA, however, as the computing budget is increased, SPSA and SFA show better performance than OCBA. Both our algorithms also show good performance when the parameter set has a size of 10(8). SFA is seen to show the best overall performance. Unlike most other DPSO algorithms in the literature, an advantage with our algorithms is that they are easily implementable regardless of the size of the parameter sets and show good performance in both scenarios.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

When hosting XML information on relational backends, a mapping has to be established between the schemas of the information source and the target storage repositories. A rich body of recent literature exists for mapping isolated components of XML Schema to their relational counterparts, especially with regard to table configurations. In this paper, we present the Elixir system for designing industrial-strength mappings for real-world applications. Specifically, it produces an information-preserving holistic mapping that transforms the complete XML world-view (XML schema with constraints, XML documents XQuery queries including triggers and views) into a full-scale relational mapping (table definitions, integrity constraints, indices, triggers and views) that is tuned to the application workload. A key design feature of Elixir is that it performs all its mapping-related optimizations in the XML source space, rather than in the relational target space. Further, unlike the XML mapping tools of commercial database systems, which rely heavily on user inputs, Elixir takes a principled cost-based approach to automatically find an efficient relational mapping. A prototype of Elixir is operational and we quantitatively demonstrate its functionality and efficacy on a variety of real-life XML schemas.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

With the introduction of 2D flat-panel X-ray detectors, 3D image reconstruction using helical cone-beam tomography is fast replacing the conventional 2D reconstruction techniques. In 3D image reconstruction, the source orbit or scanning geometry should satisfy the data sufficiency or completeness condition for exact reconstruction. The helical scan geometry satisfies this condition and hence can give exact reconstruction. The theoretically exact helical cone-beam reconstruction algorithm proposed by Katsevich is a breakthrough and has attracted interest in the 3D reconstruction using helical cone-beam Computed Tomography.In many practical situations, the available projection data is incomplete. One such case is where the detector plane does not completely cover the full extent of the object being imaged in lateral direction resulting in truncated projections. This result in artifacts that mask small features near to the periphery of the ROI when reconstructed using the convolution back projection (CBP) method assuming that the projection data is complete. A number of techniques exist which deal with completion of missing data followed by the CBP reconstruction. In 2D, linear prediction (LP)extrapolation has been shown to be efficient for data completion, involving minimal assumptions on the nature of the data, producing smooth extensions of the missing projection data.In this paper, we propose to extend the LP approach for extrapolating helical cone beam truncated data. The projection on the multi row flat panel detectors has missing columns towards either ends in the lateral direction in truncated data situation. The available data from each detector row is modeled using a linear predictor. The available data is extrapolated and this completed projection data is backprojected using the Katsevich algorithm. Simulation results show the efficacy of the proposed method.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Many downscaling techniques have been developed in the past few years for projection of station-scale hydrological variables from large-scale atmospheric variables simulated by general circulation models (GCMs) to assess the hydrological impacts of climate change. This article compares the performances of three downscaling methods, viz. conditional random field (CRF), K-nearest neighbour (KNN) and support vector machine (SVM) methods in downscaling precipitation in the Punjab region of India, belonging to the monsoon regime. The CRF model is a recently developed method for downscaling hydrological variables in a probabilistic framework, while the SVM model is a popular machine learning tool useful in terms of its ability to generalize and capture nonlinear relationships between predictors and predictand. The KNN model is an analogue-type method that queries days similar to a given feature vector from the training data and classifies future days by random sampling from a weighted set of K closest training examples. The models are applied for downscaling monsoon (June to September) daily precipitation at six locations in Punjab. Model performances with respect to reproduction of various statistics such as dry and wet spell length distributions, daily rainfall distribution, and intersite correlations are examined. It is found that the CRF and KNN models perform slightly better than the SVM model in reproducing most daily rainfall statistics. These models are then used to project future precipitation at the six locations. Output from the Canadian global climate model (CGCM3) GCM for three scenarios, viz. A1B, A2, and B1 is used for projection of future precipitation. The projections show a change in probability density functions of daily rainfall amount and changes in the wet and dry spell distributions of daily precipitation. Copyright (C) 2011 John Wiley & Sons, Ltd.