936 resultados para static random access memory
Resumo:
In this paper, we study the Foschini Miljanic algorithm, which was originally proposed in a static channel environment. We investigate the algorithm in a random channel environment, study its convergence properties and apply the Gerschgorin theorem to derive sufficient conditions for the convergence of the algorithm. We apply the Foschini and Miljanic algorithm to cellular networks and derive sufficient conditions for the convergence of the algorithm in distribution and validate the results with simulations. In cellular networks, the conditions which ensure convergence in distribution can be easily verified.
Resumo:
We calculate analytically the average number of fixed points in the Hopfield model of associative memory when a random antisymmetric part is added to the otherwise symmetric synaptic matrix. Addition of the antisymmetric part causes an exponential decrease in the total number of fixed points. If the relative strength of the antisymmetric component is small, then its presence does not cause any substantial degradation of the quality of retrieval when the memory loading level is low. We also present results of numerical simulations which provide qualitative (as well as quantitative for some aspects) confirmation of the predictions of the analytic study. Our numerical results suggest that the analytic calculation of the average number of fixed points yields the correct value for the typical number of fixed points.
Resumo:
In this paper we propose a new method of data handling for web servers. We call this method Network Aware Buffering and Caching (NABC for short). NABC facilitates reduction of data copies in web server's data sending path, by doing three things: (1) Layout the data in main memory in a way that protocol processing can be done without data copies (2) Keep a unified cache of data in kernel and ensure safe access to it by various processes and kernel and (3) Pass only the necessary meta data between processes so that bulk data handling time spent during IPC can be reduced. We realize NABC by implementing a set of system calls and an user library. The end product of the implementation is a set of APIs specifically designed for use by the web servers. We port an in house web server called SWEET, to NABC APIs and evaluate performance using a range of workloads both simulated and real. The results show a very impressive gain of 12% to 21% in throughput for static file serving and 1.6 to 4 times gain in throughput for lightweight dynamic content serving for a server using NABC APIs over the one using UNIX APIs.
Resumo:
In this paper we address the problem of transmission of correlated sources over a fading multiple access channel (MAC). We provide sufficient conditions for transmission with given distortions. Next these conditions are specialized to a Gaussian MAC (GMAC). Transmission schemes for discrete and Gaussian sources over a fading GMAC are considered. Various power allocation strategies are also compared. Keywords: Fading MAC, Power allocation, Random TDMA, Amplify and Forward, Correlated sources.
Resumo:
Today's SoCs are complex designs with multiple embedded processors, memory subsystems, and application specific peripherals. The memory architecture of embedded SoCs strongly influences the power and performance of the entire system. Further, the memory subsystem constitutes a major part (typically up to 70%) of the silicon area for the current day SoC. In this article, we address the on-chip memory architecture exploration for DSP processors which are organized as multiple memory banks, where banks can be single/dual ported with non-uniform bank sizes. In this paper we propose two different methods for physical memory architecture exploration and identify the strengths and applicability of these methods in a systematic way. Both methods address the memory architecture exploration for a given target application by considering the application's data access characteristics and generates a set of Pareto-optimal design points that are interesting from a power, performance and VLSI area perspective. To the best of our knowledge, this is the first comprehensive work on memory space exploration at physical memory level that integrates data layout and memory exploration to address the system objectives from both hardware design and application software development perspective. Further we propose an automatic framework that explores the design space identifying 100's of Pareto-optimal design points within a few hours of running on a standard desktop configuration.
Resumo:
Structural dynamics of dendritic spines is one of the key correlative measures of synaptic plasticity for encoding short-term and long-term memory. Optical studies of structural changes in brain tissue using confocal microscopy face difficulties of scattering. This results in low signal-to-noise ratio and thus limiting the imaging depth to few tens of microns. Multiphoton microscopy (MpM) overcomes this limitation by using low-energy photons to cause localized excitation and achieve high resolution in all three dimensions. Multiple low-energy photons with longer wavelengths minimize scattering and allow access to deeper brain regions at several hundred microns. In this article, we provide a basic understanding of the physical phenomena that give MpM an edge over conventional microscopy. Further, we highlight a few of the key studies in the field of learning and memory which would not have been possible without the advent of MpM.
Resumo:
It is essential to accurately estimate the working set size (WSS) of an application for various optimizations such as to partition cache among virtual machines or reduce leakage power dissipated in an over-allocated cache by switching it OFF. However, the state-of-the-art heuristics such as average memory access latency (AMAL) or cache miss ratio (CMR) are poorly correlated to the WSS of an application due to 1) over-sized caches and 2) their dispersed nature. Past studies focus on estimating WSS of an application executing on a uniprocessor platform. Estimating the same for a chip multiprocessor (CMP) with a large dispersed cache is challenging due to the presence of concurrently executing threads/processes. Hence, we propose a scalable, highly accurate method to estimate WSS of an application. We call this method ``tagged WSS (TWSS)'' estimation method. We demonstrate the use of TWSS to switch-OFF the over-allocated cache ways in Static and Dynamic NonUniform Cache Architectures (SNUCA, DNUCA) on a tiled CMP. In our implementation of adaptable way SNUCA and DNUCA caches, decision of altering associativity is taken by each L2 controller. Hence, this approach scales better with the number of cores present on a CMP. It gives overall (geometric mean) 26% and 19% higher energy-delay product savings compared to AMAL and CMR heuristics on SNUCA, respectively.
Resumo:
Programming for parallel architectures that do not have a shared address space is extremely difficult due to the need for explicit communication between memories of different compute devices. A heterogeneous system with CPUs and multiple GPUs, or a distributed-memory cluster are examples of such systems. Past works that try to automate data movement for distributed-memory architectures can lead to excessive redundant communication. In this paper, we propose an automatic data movement scheme that minimizes the volume of communication between compute devices in heterogeneous and distributed-memory systems. We show that by partitioning data dependences in a particular non-trivial way, one can generate data movement code that results in the minimum volume for a vast majority of cases. The techniques are applicable to any sequence of affine loop nests and works on top of any choice of loop transformations, parallelization, and computation placement. The data movement code generated minimizes the volume of communication for a particular configuration of these. We use a combination of powerful static analyses relying on the polyhedral compiler framework and lightweight runtime routines they generate, to build a source-to-source transformation tool that automatically generates communication code. We demonstrate that the tool is scalable and leads to substantial gains in efficiency. On a heterogeneous system, the communication volume is reduced by a factor of 11X to 83X over state-of-the-art, translating into a mean execution time speedup of 1.53X. On a distributed-memory cluster, our scheme reduces the communication volume by a factor of 1.4X to 63.5X over state-of-the-art, resulting in a mean speedup of 1.55X. In addition, our scheme yields a mean speedup of 2.19X over hand-optimized UPC codes.
Resumo:
A new representation of spatio-temporal random processes is proposed in this work. In practical applications, such processes are used to model velocity fields, temperature distributions, response of vibrating systems, to name a few. Finding an efficient representation for any random process leads to encapsulation of information which makes it more convenient for a practical implementations, for instance, in a computational mechanics problem. For a single-parameter process such as spatial or temporal process, the eigenvalue decomposition of the covariance matrix leads to the well-known Karhunen-Loeve (KL) decomposition. However, for multiparameter processes such as a spatio-temporal process, the covariance function itself can be defined in multiple ways. Here the process is assumed to be measured at a finite set of spatial locations and a finite number of time instants. Then the spatial covariance matrix at different time instants are considered to define the covariance of the process. This set of square, symmetric, positive semi-definite matrices is then represented as a third-order tensor. A suitable decomposition of this tensor can identify the dominant components of the process, and these components are then used to define a closed-form representation of the process. The procedure is analogous to the KL decomposition for a single-parameter process, however, the decompositions and interpretations vary significantly. The tensor decompositions are successfully applied on (i) a heat conduction problem, (ii) a vibration problem, and (iii) a covariance function taken from the literature that was fitted to model a measured wind velocity data. It is observed that the proposed representation provides an efficient approximation to some processes. Furthermore, a comparison with KL decomposition showed that the proposed method is computationally cheaper than the KL, both in terms of computer memory and execution time.
Resumo:
Fractal dimension based damage detection method is investigated for a composite plate with random material properties. Composite material shows spatially varying random material properties because of complex manufacturing processes. Matrix cracks are considered as damage in the composite plate. Such cracks are often seen as the initial damage mechanism in composites under fatigue loading and also occur due to low velocity impact. Static deflection of the cantilevered composite plate with uniform loading is calculated using the finite element method. Damage detection is carried out based on sliding window fractal dimension operator using the static deflection. Two dimensional homogeneous Gaussian random field is generated using Karhunen-Loeve (KL) expansion to represent the spatial variation of composite material property. The robustness of fractal dimension based damage detection method is demonstrated considering the composite material properties as a two dimensional random field.
Resumo:
Fractal dimension based damage detection method is studied for a composite structure with random material properties. A composite plate with localized matrix crack is considered. Matrix cracks are often seen as the initial damage mechanism in composites. Fractal dimension based method is applied to the static deformation curve of the structure to detect localized damage. Static deflection of a cantilevered composite plate under uniform loading is calculated using the finite element method. Composite material shows spatially varying random material properties because of complex manufacturing processes. Spatial variation of material property is represented as a two dimensional homogeneous Gaussian random field. Karhunen-Loeve (KL) expansion is used to generate a random field. The robustness of fractal dimension based damage detection methods is studied considering the composite plate with spatial variation in material properties.
Resumo:
10 p.
Resumo:
[En]The present study aimed at investigating the existence of long memory properties in ten developed stock markets across the globe. When return series exhibit long memory, the series realizations are not independent over time and past returns can help predict future returns, thus violating the market efficiency hypothesis. It poses a serious challenge to the supporters of random walk behavior of the stock returns indicating a potentially predictable component in the series dynamics. We computed Hurst-Mandelbrot’s Classical R/S statistic, Lo’s statistic and semi parametric GPH statistic using spectral regression. The findings suggest existence of long memory in volatility and random walk for logarithmic return series in general for all the selected stock market indices. Findings are in line with the stylized facts of financial time series.
Resumo:
This thesis is an investigation into the nature of data analysis and computer software systems which support this activity.
The first chapter develops the notion of data analysis as an experimental science which has two major components: data-gathering and theory-building. The basic role of language in determining the meaningfulness of theory is stressed, and the informativeness of a language and data base pair is studied. The static and dynamic aspects of data analysis are then considered from this conceptual vantage point. The second chapter surveys the available types of computer systems which may be useful for data analysis. Particular attention is paid to the questions raised in the first chapter about the language restrictions imposed by the computer system and its dynamic properties.
The third chapter discusses the REL data analysis system, which was designed to satisfy the needs of the data analyzer in an operational relational data system. The major limitation on the use of such systems is the amount of access to data stored on a relatively slow secondary memory. This problem of the paging of data is investigated and two classes of data structure representations are found, each of which has desirable paging characteristics for certain types of queries. One representation is used by most of the generalized data base management systems in existence today, but the other is clearly preferred in the data analysis environment, as conceptualized in Chapter I.
This data representation has strong implications for a fundamental process of data analysis -- the quantification of variables. Since quantification is one of the few means of summarizing and abstracting, data analysis systems are under strong pressure to facilitate the process. Two implementations of quantification are studied: one analagous to the form of the lower predicate calculus and another more closely attuned to the data representation. A comparison of these indicates that the use of the "label class" method results in orders of magnitude improvement over the lower predicate calculus technique.
Resumo:
Neurodevelopmental disruptions caused by obstetric complications play a role in the etiology of several phenotypes associated with neuropsychiatric diseases and cognitive dysfunctions. Importantly, it has been noticed that epigenetic processes occurring early in life may mediate these associations. Here, DNA methylation signatures at IGF2 (insulin-like growth factor 2) and IGF2BP1-3 (IGF2-binding proteins 1-3) were examined in a sample consisting of 34 adult monozygotic (MZ) twins informative for obstetric complications and cognitive performance. Multivariate linear regression analysis of twin data was implemented to test for associations between methylation levels and both birth weight (BW) and adult working memory (WM) performance. Familial and unique environmental factors underlying these potential relationships were evaluated. A link was detected between DNA methylation levels of two CpG sites in the IGF2BP1 gene and both BW and adult WM performance. The BW-IGF2BP1 methylation association seemed due to non-shared environmental factors influencing BW, whereas the WM-IGF2BP1 methylation relationship seemed mediated by both genes and environment. Our data is in agreement with previous evidence indicating that DNA methylation status may be related to prenatal stress and later neurocognitive phenotypes. While former reports independently detected associations between DNA methylation and either BW or WM, current results suggest that these relationships are not confounded by each other.