990 resultados para Significance driven computation
Resumo:
In this paper we present a design methodology for algorithm/architecture co-design of a voltage-scalable, process variation aware motion estimator based on significance driven computation. The fundamental premise of our approach lies in the fact that all computations are not equally significant in shaping the output response of video systems. We use a statistical technique to intelligently identify these significant/not-so-significant computations at the algorithmic level and subsequently change the underlying architecture such that the significant computations are computed in an error free manner under voltage over-scaling. Furthermore, our design includes an adaptive quality compensation (AQC) block which "tunes" the algorithm and architecture depending on the magnitude of voltage over-scaling and severity of process variations. Simulation results show average power savings of similar to 33% for the proposed architecture when compared to conventional implementation in the 90 nm CMOS technology. The maximum output quality loss in terms of Peak Signal to Noise Ratio (PSNR) was similar to 1 dB without incurring any throughput penalty.
Resumo:
In this paper, we propose a design paradigm for energy efficient and variation-aware operation of next-generation multicore heterogeneous platforms. The main idea behind the proposed approach lies on the observation that not all operations are equally important in shaping the output quality of various applications and of the overall system. Based on such an observation, we suggest that all levels of the software design stack, including the programming model, compiler, operating system (OS) and run-time system should identify the critical tasks and ensure correct operation of such tasks by assigning them to dynamically adjusted reliable cores/units. Specifically, based on error rates and operating conditions identified by a sense-and-adapt (SeA) unit, the OS selects and sets the right mode of operation of the overall system. The run-time system identifies the critical/less-critical tasks based on special directives and schedules them to the appropriate units that are dynamically adjusted for highly-accurate/approximate operation by tuning their voltage/frequency. Units that execute less significant operations can operate at voltages less than what is required for correct operation and consume less power, if required, since such tasks do not need to be always exact as opposed to the critical ones. Such scheme can lead to energy efficient and reliable operation, while reducing the design cost and overheads of conventional circuit/micro-architecture level techniques.
Resumo:
Dynamic Voltage and Frequency Scaling (DVFS) exhibits fundamental limitations as a method to reduce energy consumption in computing systems. In the HPC domain, where performance is of highest priority and codes are heavily optimized to minimize idle time, DVFS has limited opportunity to achieve substantial energy savings. This paper explores if operating processors Near the transistor Threshold Volt- age (NTV) is a better alternative to DVFS for break- ing the power wall in HPC. NTV presents challenges, since it compromises both performance and reliability to reduce power consumption. We present a first of its kind study of a significance-driven execution paradigm that selectively uses NTV and algorithmic error tolerance to reduce energy consumption in performance- constrained HPC environments. Using an iterative algorithm as a use case, we present an adaptive execution scheme that switches between near-threshold execution on many cores and above-threshold execution on one core, as the computational significance of iterations in the algorithm evolves over time. Using this scheme on state-of-the-art hardware, we demonstrate energy savings ranging between 35% to 67%, while compromising neither correctness nor performance.
Resumo:
Approximate execution is a viable technique for energy-con\-strained environments, provided that applications have the mechanisms to produce outputs of the highest possible quality within the given energy budget.
We introduce a framework for energy-constrained execution with controlled and graceful quality loss. A simple programming model allows users to express the relative importance of computations for the quality of the end result, as well as minimum quality requirements. The significance-aware runtime system uses an application-specific analytical energy model to identify the degree of concurrency and approximation that maximizes quality while meeting user-specified energy constraints. Evaluation on a dual-socket 8-core server shows that the proposed
framework predicts the optimal configuration with high accuracy, enabling energy-constrained executions that result in significantly higher quality compared to loop perforation, a compiler approximation technique.
Resumo:
Computers employing some degree of data flow organisation are now well established as providing a possible vehicle for concurrent computation. Although data-driven computation frees the architecture from the constraints of the single program counter, processor and global memory, inherent in the classic von Neumann computer, there can still be problems with the unconstrained generation of fresh result tokens if a pure data flow approach is adopted. The advantages of allowing serial processing for those parts of a program which are inherently serial, and of permitting a demand-driven, as well as data-driven, mode of operation are identified and described. The MUSE machine described here is a structured architecture supporting both serial and parallel processing which allows the abstract structure of a program to be mapped onto the machine in a logical way.
Resumo:
The scalability of CMOS technology has driven computation into a diverse range of applications across the power consumption, performance and size spectra. Communication is a necessary adjunct to computation, and whether this is to push data from node-to-node in a high-performance computing cluster or from the receiver of wireless link to a neural stimulator in a biomedical implant, interconnect can take up a significant portion of the overall system power budget. Although a single interconnect methodology cannot address such a broad range of systems efficiently, there are a number of key design concepts that enable good interconnect design in the age of highly-scaled CMOS: an emphasis on highly-digital approaches to solving ‘analog’ problems, hardware sharing between links as well as between different functions (such as equalization and synchronization) in the same link, and adaptive hardware that changes its operating parameters to mitigate not only variation in the fabrication of the link, but also link conditions that change over time. These concepts are demonstrated through the use of two design examples, at the extremes of the power and performance spectra.
A novel all-digital clock and data recovery technique for high-performance, high density interconnect has been developed. Two independently adjustable clock phases are generated from a delay line calibrated to 2 UI. One clock phase is placed in the middle of the eye to recover the data, while the other is swept across the delay line. The samples produced by the two clocks are compared to generate eye information, which is used to determine the best phase for data recovery. The functions of the two clocks are swapped after the data phase is updated; this ping-pong action allows an infinite delay range without the use of a PLL or DLL. The scheme's generalized sampling and retiming architecture is used in a sharing technique that saves power and area in high-density interconnect. The eye information generated is also useful for tuning an adaptive equalizer, circumventing the need for dedicated adaptation hardware.
On the other side of the performance/power spectra, a capacitive proximity interconnect has been developed to support 3D integration of biomedical implants. In order to integrate more functionality while staying within size limits, implant electronics can be embedded onto a foldable parylene (‘origami’) substrate. Many of the ICs in an origami implant will be placed face-to-face with each other, so wireless proximity interconnect can be used to increase communication density while decreasing implant size, as well as facilitate a modular approach to implant design, where pre-fabricated parylene-and-IC modules are assembled together on-demand to make custom implants. Such an interconnect needs to be able to sense and adapt to changes in alignment. The proposed array uses a TDC-like structure to realize both communication and alignment sensing within the same set of plates, increasing communication density and eliminating the need to infer link quality from a separate alignment block. In order to distinguish the communication plates from the nearby ground plane, a stimulus is applied to the transmitter plate, which is rectified at the receiver to bias a delay generation block. This delay is in turn converted into a digital word using a TDC, providing alignment information.
Resumo:
Inherently error-resilient applications in areas such as signal processing, machine learning and data analytics provide opportunities for relaxing reliability requirements, and thereby reducing the overhead incurred by conventional error correction schemes. In this paper, we exploit the tolerable imprecision of such applications by designing an energy-efficient fault-mitigation scheme for unreliable data memories to meet target yield. The proposed approach uses a bit-shuffling mechanism to isolate faults into bit locations with lower significance. This skews the bit-error distribution towards the low order bits, substantially limiting the output error magnitude. By controlling the granularity of the shuffling, the proposed technique enables trading-off quality for power, area, and timing overhead. Compared to error-correction codes, this can reduce the overhead by as much as 83% in read power, 77% in read access time, and 89% in area, when applied to various data mining applications in 28nm process technology.
Resumo:
ICT (Information and Communication Technology) creates numerous opportunities for teachers to re-think their pedagogies. In subjects like mathematics which draws upon abstract concepts, ICT creates such an opportunity. Instead of a mimetic pedagogical approach, suitably designed activities with ICT can enable learners to engage more proactively with their learning. In this quasi-experimental designed study, ICT was used in teaching mathematics to a group of first year high school students (N=25) in Australia. The control group was taught predominantly through traditional pedagogies (N=22). Most of the variables that had previously impacted on the design of such studies were suitably controlled in this yearlong investigation. Quantitative and qualitative results showed that students who were taught by ICT driven pedagogies benefitted from the experience. Pre and post-test means showed that there was a difference between the treatment and control groups. Of greater significance was that the students (in the treatment group) believed that the technology enabled them to engage more with their learning.
Resumo:
The increasing demand for mobile video has attracted much attention from both industry and researchers. To satisfy users and to facilitate the usage of mobile video, providing optimal quality to the users is necessary. As a result, quality of experience (QoE) becomes an important focus in measuring the overall quality perceived by the end-users, from the aspects of both objective system performance and subjective experience. However, due to the complexity of user experience and diversity of resources (such as videos, networks and mobile devices), it is still challenging to develop QoE models for mobile video that can represent how user-perceived value varies with changing conditions. Previous QoE modelling research has two main limitations: aspects influencing QoE are insufficiently considered; and acceptability as the user value is seldom studied. Focusing on the QoE modelling issues, two aims are defined in this thesis: (i) investigating the key influencing factors of mobile video QoE; and (ii) establishing QoE prediction models based on the relationships between user acceptability and the influencing factors, in order to help provide optimal mobile video quality. To achieve the first goal, a comprehensive user study was conducted. It investigated the main impacts on user acceptance: video encoding parameters such as quantization parameter, spatial resolution, frame rate, and encoding bitrate; video content type; mobile device display resolution; and user profiles including gender, preference for video content, and prior viewing experience. Results from both quantitative and qualitative analysis revealed the significance of these factors, as well as how and why they influenced user acceptance of mobile video quality. Based on the results of the user study, statistical techniques were used to generate a set of QoE models that predict the subjective acceptability of mobile video quality by using a group of the measurable influencing factors, including encoding parameters and bitrate, content type, and mobile device display resolution. Applying the proposed QoE models into a mobile video delivery system, optimal decisions can be made for determining proper video coding parameters and for delivering most suitable quality to users. This would lead to consistent user experience on different mobile video content and efficient resource allocation. The findings in this research enhance the understanding of user experience in the field of mobile video, which will benefit mobile video design and research. This thesis presents a way of modelling QoE by emphasising user acceptability of mobile video quality, which provides a strong connection between technical parameters and user-desired quality. Managing QoE based on acceptability promises the potential for adapting to the resource limitations and achieving an optimal QoE in the provision of mobile video content.
Computation of ECG signal features using MCMC modelling in software and FPGA reconfigurable hardware
Resumo:
Computational optimisation of clinically important electrocardiogram signal features, within a single heart beat, using a Markov-chain Monte Carlo (MCMC) method is undertaken. A detailed, efficient data-driven software implementation of an MCMC algorithm has been shown. Initially software parallelisation is explored and has been shown that despite the large amount of model parameter inter-dependency that parallelisation is possible. Also, an initial reconfigurable hardware approach is explored for future applicability to real-time computation on a portable ECG device, under continuous extended use.
Resumo:
Designed for undergraduate and postgraduate students, academic researchers and industrial practitioners, this book provides comprehensive case studies on numerical computing of industrial processes and step-by-step procedures for conducting industrial computing. It assumes minimal knowledge in numerical computing and computer programming, making it easy to read, understand and follow. Topics discussed include fundamentals of industrial computing, finite difference methods, the Wavelet-Collocation Method, the Wavelet-Galerkin Method, High Resolution Methods, and comparative studies of various methods. These are discussed using examples of carefully selected models from real processes of industrial significance. The step-by-step procedures in all these case studies can be easily applied to other industrial processes without a need for major changes and thus provide readers with useful frameworks for the applications of engineering computing in fundamental research problems and practical development scenarios.
Resumo:
Wound healing and tumour growth involve collective cell spreading, which is driven by individual motility and proliferation events within a population of cells. Mathematical models are often used to interpret experimental data and to estimate the parameters so that predictions can be made. Existing methods for parameter estimation typically assume that these parameters are constants and often ignore any uncertainty in the estimated values. We use approximate Bayesian computation (ABC) to estimate the cell diffusivity, D, and the cell proliferation rate, λ, from a discrete model of collective cell spreading, and we quantify the uncertainty associated with these estimates using Bayesian inference. We use a detailed experimental data set describing the collective cell spreading of 3T3 fibroblast cells. The ABC analysis is conducted for different combinations of initial cell densities and experimental times in two separate scenarios: (i) where collective cell spreading is driven by cell motility alone, and (ii) where collective cell spreading is driven by combined cell motility and cell proliferation. We find that D can be estimated precisely, with a small coefficient of variation (CV) of 2–6%. Our results indicate that D appears to depend on the experimental time, which is a feature that has been previously overlooked. Assuming that the values of D are the same in both experimental scenarios, we use the information about D from the first experimental scenario to obtain reasonably precise estimates of λ, with a CV between 4 and 12%. Our estimates of D and λ are consistent with previously reported values; however, our method is based on a straightforward measurement of the position of the leading edge whereas previous approaches have involved expensive cell counting techniques. Additional insights gained using a fully Bayesian approach justify the computational cost, especially since it allows us to accommodate information from different experiments in a principled way.
Resumo:
The studies presented in this thesis contribute to the understanding of evolutionary ecology of three major viruses threatening cultivated sweetpotato (Ipomoea batatas Lam) in East Africa: Sweet potato feathery mottle virus (SPFMV; genus Potyvirus; Potyviridae), Sweet potato chlorotic stunt virus (SPCSV; genus Crinivirus; Closteroviridae) and Sweet potato mild mottle virus (SPMMV; genus Ipomovirus; Potyviridae). The viruses were serologically detected and the positive results confirmed by RT-PCR and sequencing. SPFMV was detected in 24 wild plant species of family Convolvulacea (genera Ipomoea, Lepistemon and Hewittia), of which 19 species were new natural hosts for SPFMV. SPMMV and SPCSV were detected in wild plants belonging to 21 and 12 species (genera Ipomoea, Lepistemon and Hewittia), respectively, all of which were previously unknown to be natural hosts of these viruses. SPFMV was the most abundant virus being detected in 17% of the plants, while SPMMV and SPCSV were detected in 9.8% and 5.4% of the assessed plants, respectively. Wild plants in Uganda were infected with the East African (EA), common (C), and the ordinary (O) strains, or co-infected with the EA and the C strain of SPFMV. The viruses and virus-like diseases were more frequent in the eastern agro-ecological zone than the western and central zones, which contrasted with known incidences of these viruses in sweetpotato crops, except for northern zone where incidences were lowest in wild plants as in sweetpotato. The NIb/CP junction in SPMMV was determined experimentally which facilitated CP-based phylogenetic and evolutionary analyses of SPMMV. Isolates of all the three viruses from wild plants were genetically similar to those found in cultivated sweetpotatoes in East Africa. There was no evidence of host-driven population genetic structures suggesting frequent transmission of these viruses between their wild and cultivated hosts. The p22 RNA silencing suppressor-encoding sequence was absent in a few SPCSV isolates, but regardless of this, SPCSV isolates incited sweet potato virus disease (SPVD) in sweetpotato plants co-infected with SPFMV, indicating that p22 is redundant for synergism between SCSV and SPFMV. Molecular evolutionary analysis revealed that isolates of strain EA of SPFMV that is largely restricted geographically in East Africa experience frequent recombination in comparison to isolates of strain C that is globally distributed. Moreover, non-homologous recombination events between strains EA and C were rare, despite frequent co-infections of these strains in wild plants, suggesting purifying selection against non-homologous recombinants between these strains or that such recombinants are mostly not infectious. Recombination was detected also in the 5 - and 3 -proximal regions of the SPMMV genome providing the first evidence of recombination in genus Ipomovirus, but no recombination events were detected in the characterized genomic regions of SPCSV. Strong purifying selection was implicated on evolution of majority of amino acids of the proteins encoded by the analyzed genomic regions of SPFMV, SPMMV and SPCSV. However, positive selection was predicted on 17 amino acids distributed over the whole the coat protein (CP) in the globally distributed strain C, as compared to only 4 amino acids in the multifunctional CP N-terminus (CP-NT) of strain EA largely restricted geographically to East Africa. A few amino acid sites in the N-terminus of SPMMV P1, the p7 protein and RNA silencing suppressor proteins p22 and RNase3 of SPCSV were also submitted to positive selection. Positively selected amino acids may constitute ligand-binding domains that determine interactions with plant host and/or insect vector factors. The P1 proteinase of SPMMV (genus Ipomovirus) seems to respond to needs of adaptation, which was not observed with the helper component proteinase (HC-Pro) of SPMMV, although the HC-Pro is responsible for many important molecular interactions in genus Potyvirus. Because the centre of origin of cultivated sweetpotato is in the Americas from where the crop was dispersed to other continents in recent history (except for the Australasia and South Pacific region), it would be expected that identical viruses and their strains occur worldwide, presuming virus dispersal with the host. Apparently, this seems not to be the case with SPMMV, the strain EA of SPFMV and the strain EA of SPCSV that are largely geographically confined in East Africa where they are predominant and occur both in natural and agro-ecosystems. The geographical distribution of plant viruses is constrained more by virus-vector relations than by virus-host interactions, which in accordance of the wide range of natural host species and the geographical confinement to East Africa suggest that these viruses existed in East African wild plants before the introduction of sweetpotato. Subsequently, these studies provide compelling evidence that East Africa constitutes a cradle of SPFMV strain EA, SPCSV strain EA, and SPMMV. Therefore, sweet potato virus disease (SPVD) in East Africa may be one of the examples of damaging virus diseases resulting from exchange of viruses between introduced crops and indigenous wild plant species. Keywords: Convolvulaceae, East Africa, epidemiology, evolution, genetic variability, Ipomoea, recombination, SPCSV, SPFMV, SPMMV, selection pressure, sweetpotato, wild plant species Author s Address: Arthur K. Tugume, Department of Agricultural Sciences, Faculty of Agriculture and Forestry, University of Helsinki, Latokartanonkaari 7, P.O Box 27, FIN-00014, Helsinki, Finland. Email: tugume.arthur@helsinki.fi Author s Present Address: Arthur K. Tugume, Department of Botany, Faculty of Science, Makerere University, P.O. Box 7062, Kampala, Uganda. Email: aktugume@botany.mak.ac.ug, tugumeka@yahoo.com
Resumo:
We consider the problem of detecting statistically significant sequential patterns in multineuronal spike trains. These patterns are characterized by ordered sequences of spikes from different neurons with specific delays between spikes. We have previously proposed a data-mining scheme to efficiently discover such patterns, which occur often enough in the data. Here we propose a method to determine the statistical significance of such repeating patterns. The novelty of our approach is that we use a compound null hypothesis that not only includes models of independent neurons but also models where neurons have weak dependencies. The strength of interaction among the neurons is represented in terms of certain pair-wise conditional probabilities. We specify our null hypothesis by putting an upper bound on all such conditional probabilities. We construct a probabilistic model that captures the counting process and use this to derive a test of significance for rejecting such a compound null hypothesis. The structure of our null hypothesis also allows us to rank-order different significant patterns. We illustrate the effectiveness of our approach using spike trains generated with a simulator.
Resumo:
Computations have been carried out for simulating supersonic flow through a set of converging-diverging nozzles with their expanding jets forming a laser cavity and flow patterns through diffusers, past the cavity. A thorough numerical investigation with 3-D RANS code is carried out to capture the flow distribution which comprises of shock patterns and multiple supersonic jet interactions. The analysis of pressure recovery characteristics during the flow through the diffusers is an important parameter of the simulation and is critical for the performance of the laser device. The results of the computation have shown a close agreement with the experimentally measured parameters as well as other established results indicating that the flow analysis done is found to be satisfactory.