79 resultados para Sampling schemes
em Indian Institute of Science - Bangalore - Índia
Resumo:
We propose data acquisition from continuous-time signals belonging to the class of real-valued trigonometric polynomials using an event-triggered sampling paradigm. The sampling schemes proposed are: level crossing (LC), close to extrema LC, and extrema sampling. Analysis of robustness of these schemes to jitter, and bandpass additive gaussian noise is presented. In general these sampling schemes will result in non-uniformly spaced sample instants. We address the issue of signal reconstruction from the acquired data-set by imposing structure of sparsity on the signal model to circumvent the problem of gap and density constraints. The recovery performance is contrasted amongst the various schemes and with random sampling scheme. In the proposed approach, both sampling and reconstruction are non-linear operations, and in contrast to random sampling methodologies proposed in compressive sensing these techniques may be implemented in practice with low-power circuitry.
Resumo:
The accelerated rate of increase in atmospheric CO2 concentration in recent years has revived the idea of stabilizing the global climate through geoengineering schemes. Majority of the proposed geoengineering schemes will attempt to reduce the amount of solar radiation absorbed by our planet. Climate modelling studies of these so called 'sunshade geoengineering schemes' show that global warming from increasing concentrations of CO2 can be mitigated by intentionally manipulating the amount of sunlight absorbed by the climate system. These studies also suggest that the residual changes could be large on regional scales, so that climate change may not be mitigated on a local basis. More recent modelling studies have shown that these schemes could lead to a slow-down in the global hydrological cycle. Other problems such as changes in the terrestrial carbon cycle and ocean acidification remain unsolved by sunshade geoengineering schemes. In this article, I review the proposed geoengineering schemes, results from climate models and discuss why geoengineering is not the best option to deal with climate change.
Resumo:
We present a technique for an all-digital on-chip delay measurement system to measure the skews in a clock distribution network. It uses the principle of sub-sampling. Measurements from a prototype fabricated in a 65 nm industrial process, indicate the ability to measure delays with a resolution of 0.5ps and a DNL of 1.2 ps.
Resumo:
The short duration of the Doppler signal and noise content in it necessitate a validation scheme to be incorporated in the electronic processor used for frequency measurement, There are several different validation schemes that can be employed in period timing devices. A detailed study of the influence of these validation schemes on the measured frequency has been reported here. These studies were carried out by using a combination of a fast A/D converter and computer. Doppler bursts obtained from an air flow were digitised and stored on magnetic discs. Suitable computer programs were then used to simulate the performance of period timing devices with different validation schemes and the frequency of the stored bursts were evaluated. It is found that best results are obtained when the validation scheme enables frequency measurement to be made over a large number of cycles within the burst.
Resumo:
It is well known that the notions of normal forms and acyclicity capture many practical desirable properties for database schemes. The basic schema design problem is to develop design methodologies that strive toward these ideals. The usual approach is to first normalize the database scheme as far as possible. If the resulting scheme is cyclic, then one tries to transform it into an acyclic scheme. In this paper, we argue in favor of carrying out these two phases of design concurrently. In order to do this efficiently, we need to be able to incrementally analyze the acyclicity status of a database scheme as it is being designed. To this end, we propose the formalism of "binary decompositions". Using this, we characterize design sequences that exactly generate theta-acyclic schemes, for theta = agr,beta. We then show how our results can be put to use in database design. Finally, we also show that our formalism above can be effectively used as a proof tool in dependency theory. We demonstrate its power by showing that it leads to a significant simplification of the proofs of some previous results connecting sets of multivalued dependencies and acyclic join dependencies.
Resumo:
Database schemes can be viewed as hypergraphs with individual relation schemes corresponding to the edges of a hypergraph. Under this setting, a new class of "acyclic" database schemes was recently introduced and was shown to have a claim to a number of desirable properties. However, unlike the case of ordinary undirected graphs, there are several unequivalent notions of acyclicity of hypergraphs. Of special interest among these are agr-, beta-, and gamma-, degrees of acyclicity, each characterizing an equivalence class of desirable properties for database schemes, represented as hypergraphs. In this paper, two complementary approaches to designing beta-acyclic database schemes have been presented. For the first part, a new notion called "independent cycle" is introduced. Based on this, a criterion for beta-acyclicity is developed and is shown equivalent to the existing definitions of beta-acyclicity. From this and the concept of the dual of a hypergraph, an efficient algorithm for testing beta-acyclicity is developed. As for the second part, a procedure is evolved for top-down generation of beta-acyclic schemes and its correctness is established. Finally, extensions and applications of ideas are described.
Resumo:
A simple instrument that can provide a sequence of timed pulses for first initiating a transient process and then enabling sampling and recording periodically during the course of the transient event is described. The time delay between the first of these sampling pulses and the start of the transient event is adjustable. This sequence generator has additional features that make it ideal for use in acquiring the waveforms when a storage oscilloscope is used as the recording device. For avoiding the clutter caused by many waveforms being recorded at the same place on an oscilloscope screen such features as displacements of waveforms in the X and Y directions and trace blanking at places where the waveform is not required, have been incorporated. This sequence generator has been employed to acquire a sequence of Raman scattered radiation signals from an adiabatically expanding saturated vapour probed by a flashlamp-pumped dye laser.
Resumo:
Among the iterative schemes for computing the Moore — Penrose inverse of a woll-conditioned matrix, only those which have an order of convergence three or two are computationally efficient. A Fortran programme for these schemes is provided.
Resumo:
Because of limited sensor and communication ranges, designing efficient mechanisms for cooperative tasks is difficult. In this article, several negotiation schemes for multiple agents performing a cooperative task are presented. The negotiation schemes provide suboptimal solutions, but have attractive features of fast decision-making, and scalability to large number of agents without increasing the complexity of the algorithm. A software agent architecture of the decision-making process is also presented. The effect of the magnitude of information flow during the negotiation process is studied by using different models of the negotiation scheme. The performance of the various negotiation schemes, using different information structures, is studied based on the uncertainty reduction achieved for a specified number of search steps. The negotiation schemes perform comparable to that of optimal strategy in terms of uncertainty reduction and also require very low computational time, similar to 7 per cent to that of optimal strategy. Finally, analysis on computational and communication requirement for the negotiation schemes is carried out.
Resumo:
We propose a self-regularized pseudo-time marching strategy for ill-posed, nonlinear inverse problems involving recovery of system parameters given partial and noisy measurements of system response. While various regularized Newton methods are popularly employed to solve these problems, resulting solutions are known to sensitively depend upon the noise intensity in the data and on regularization parameters, an optimal choice for which remains a tricky issue. Through limited numerical experiments on a couple of parameter re-construction problems, one involving the identification of a truss bridge and the other related to imaging soft-tissue organs for early detection of cancer, we demonstrate the superior features of the pseudo-time marching schemes.
Resumo:
One of the most important applications of adaptive systems is in noise cancellation using adaptive filters. Ln this paper, we propose adaptive noise cancellation schemes for the enhancement of EEG signals in the presence of EOG artifacts. The effect of two reference inputs is studied on simulated as well as recorded EEG signals and it is found that one reference input is enough to get sufficient minimization of EOG artifacts. This has been verified through correlation analysis also. We use signal to noise ratio and linear prediction spectra, along with time plots, for comparing the performance of the proposed schemes for minimizing EOG artifacts from contaminated EEG signals. Results show that the proposed schemes are very effective (especially the one which employs Newton's method) in minimizing the EOG artifacts from contaminated EEG signals.
Resumo:
In this paper, we propose a new security metric for measuring resilience of a symmetric key distribution scheme in wireless sensor network. A polynomial-based and a novel complete connectivity schemes are proposed and an analytical comparison, in terms of security and connectivity, between the schemes is shown. Motivated by the schemes, we derive general expressions for security and connectivity. A number of conclusions are made using these general expressions.
Resumo:
Understanding of the shape and size of different features of the human body from scanned data is necessary for automated design and evaluation of product ergonomics. In this paper, a computational framework is presented for automatic detection and recognition of important facial feature regions, from scanned head and shoulder polyhedral models. A noise tolerant methodology is proposed using discrete curvature computations, band-pass filtering, and morphological operations for isolation of the primary feature regions of the face, namely, the eyes, nose, and mouth. Spatial disposition of the critical points of these isolated feature regions is analyzed for the recognition of these critical points as the standard landmarks associated with the primary facial features. A number of clinically identified landmarks lie on the facial midline. An efficient algorithm for detection and processing of the midline, using a point sampling technique, is also presented. The results obtained using data of more than 20 subjects are verified through visualization and physical measurements. A color based and triangle skewness based schemes for isolation of geometrically nonprominent features and ear region are also presented. [DOI: 10.1115/1.3330420]
Resumo:
We evaluated trained listener-based acoustic sampling as a reliable and non-invasive method for rapid assessment of ensiferan species diversity in tropical evergreen forests. This was done by evaluating the reliability of identification of species and numbers of calling individuals using psychoacoustic experiments in the laboratory and by comparing psychoacoustic sampling in the field with ambient noise recordings made at the same time. The reliability of correct species identification by the trained listener was 100% for 16 out of 20 species tested in the laboratory. The reliability of identifying the numbers of individuals correctly was 100% for 13 out of 20 species. The human listener performed slightly better than the instrument in detecting low frequency and broadband calls in the field, whereas the recorder detected high frequency calls with greater probability. To address the problem of pseudoreplication during spot sampling in the field, we monitored the movement of calling individuals using focal animal sampling. The average distance moved by calling individuals for 17 out of 20 species was less than 1.5 m in half an hour. We suggest that trained listener-based sampling is preferable for crickets and low frequency katydids, whereas broadband recorders are preferable for katydid species with high frequency calls for accurate estimation of ensiferan species richness and relative abundance in an area.
Resumo:
The Orthogonal Frequency Division Multiplexing (OFDM) is a form of Multi-Carrier Modulation where the data stream is transmitted over a number of carriers which are orthogonal to each other i.e. the carrier spacing is selected such that each carrier is located at the zeroes of all other carriers in the spectral domain. This paper proposes a new novel sampling offset estimation algorithm for an OFDM system in order to receive the OFDM data symbols error-free over the noisy channel at the receiver and to achieve fine timing synchronization between the transmitter and the receiver. The performance of this algorithm has been studied in AWGN, ADSL and SUI channels successfully.