17 resultados para ALS data-set
em Cambridge University Engineering Department Publications Database
Resumo:
DNA microarrays provide such a huge amount of data that unsupervised methods are required to reduce the dimension of the data set and to extract meaningful biological information. This work shows that Independent Component Analysis (ICA) is a promising approach for the analysis of genome-wide transcriptomic data. The paper first presents an overview of the most popular algorithms to perform ICA. These algorithms are then applied on a microarray breast-cancer data set. Some issues about the application of ICA and the evaluation of biological relevance of the results are discussed. This study indicates that ICA significantly outperforms Principal Component Analysis (PCA).
Resumo:
Space time cube representation is an information visualization technique where spatiotemporal data points are mapped into a cube. Information visualization researchers have previously argued that space time cube representation is beneficial in revealing complex spatiotemporal patterns in a data set to users. The argument is based on the fact that both time and spatial information are displayed simultaneously to users, an effect difficult to achieve in other representations. However, to our knowledge the actual usefulness of space time cube representation in conveying complex spatiotemporal patterns to users has not been empirically validated. To fill this gap, we report on a between-subjects experiment comparing novice users' error rates and response times when answering a set of questions using either space time cube or a baseline 2D representation. For some simple questions, the error rates were lower when using the baseline representation. For complex questions where the participants needed an overall understanding of the spatiotemporal structure of the data set, the space time cube representation resulted in on average twice as fast response times with no difference in error rates compared to the baseline. These results provide an empirical foundation for the hypothesis that space time cube representation benefits users analyzing complex spatiotemporal patterns.
Resumo:
A direct comparison between time resolved PLIF measurements of OH and two dimensional slices from a full three dimensional DNS data set of turbulent premixed flame kernels in lean methane/air mixture was presented. The local flame structure and the degree of flame wrinkling were examined in response to differing turbulence intensities and turbulent Reynolds numbers. Simulations were performed using the SEGA DNS code, which is based on the solution of the compressible Navier Stokes, species, and energy equations for a lean hydrocarbon mixture. For the OH PLIF measurements, a cluster of four Nd:YAG laser was fired sequentially at high repetition rates and used to pump a dye laser. The frequency doubled laser beam was formed into a sheet of 40 mm height using a cylindrical telescope. The combination of PLIF and DNS has been demonstrated as a powerful tool for flame analysis. This research will form the basis for the development of sub-grid-scale (SGS) models for LES of lean-premixed combustion systems such as gas turbines. This is an abstract of a paper presented at the 30th International Symposium on Combustion (Chicago, IL 7/25-30/2004).
Resumo:
Using fluorescence microscopy with single molecule sensitivity it is now possible to follow the movement of individual fluorophore tagged molecules such as proteins and lipids in the cell membrane with nanometer precision. These experiments are important as they allow many key biological processes on the cell membrane and in the cell, such as transcription, translation and DNA replication, to be studied at new levels of detail. Computerized microscopes generate sequences of images (in the order of tens to hundreds) of the molecules diffusing and one of the challenges is to track these molecules to obtain reliable statistics such as speed distributions, diffusion patterns, intracellular positioning, etc. The data set is challenging because the molecules are tagged with a single or small number of fluorophores, which makes it difficult to distinguish them from the background, the fluorophore bleaches irreversibly over time, the number of tagged molecules are unknown and there is occasional loss of signal from the tagged molecules. All these factors make accurate tracking over long trajectories difficult. Also the experiments are technically difficulty to conduct and thus there is a pressing need to develop better algorithms to extract the maximum information from the data. For this purpose we propose a Bayesian approach and apply our technique to synthetic and a real experimental data set.
Resumo:
In spite of over two decades of intense research, illumination and pose invariance remain prohibitively challenging aspects of face recognition for most practical applications. The objective of this work is to recognize faces using video sequences both for training and recognition input, in a realistic, unconstrained setup in which lighting, pose and user motion pattern have a wide variability and face images are of low resolution. In particular there are three areas of novelty: (i) we show how a photometric model of image formation can be combined with a statistical model of generic face appearance variation, learnt offline, to generalize in the presence of extreme illumination changes; (ii) we use the smoothness of geodesically local appearance manifold structure and a robust same-identity likelihood to achieve invariance to unseen head poses; and (iii) we introduce an accurate video sequence "reillumination" algorithm to achieve robustness to face motion patterns in video. We describe a fully automatic recognition system based on the proposed method and an extensive evaluation on 171 individuals and over 1300 video sequences with extreme illumination, pose and head motion variation. On this challenging data set our system consistently demonstrated a nearly perfect recognition rate (over 99.7%), significantly outperforming state-of-the-art commercial software and methods from the literature. © Springer-Verlag Berlin Heidelberg 2006.
Resumo:
We present a novel method to perform an accurate registration of 3-D nonrigid bodies by using phase-shift properties of the dual-tree complex wavelet transform (DT-CWT). Since the phases of DT-\BBCWT coefficients change approximately linearly with the amount of feature displacement in the spatial domain, motion can be estimated using the phase information from these coefficients. The motion estimation is performed iteratively: first by using coarser level complex coefficients to determine large motion components and then by employing finer level coefficients to refine the motion field. We use a parametric affine model to describe the motion, where the affine parameters are found locally by substituting into an optical flow model and by solving the resulting overdetermined set of equations. From the estimated affine parameters, the motion field between the sensed and the reference data sets can be generated, and the sensed data set then can be shifted and interpolated spatially to align with the reference data set. © 2011 IEEE.
2D PIV measurements in the near field of grid turbulence using stitched fields from multiple cameras
Resumo:
We present measurements of grid turbulence using 2D particle image velocimetry taken immediately downstream from the grid at a Reynolds number of Re M = 16500 where M is the rod spacing. A long field of view of 14M x 4M in the down- and cross-stream directions was achieved by stitching multiple cameras together. Two uniform biplanar grids were selected to have the same M and pressure drop but different rod diameter D and crosssection. A large data set (10 4 vector fields) was obtained to ensure good convergence of second-order statistics. Estimations of the dissipation rate ε of turbulent kinetic energy (TKE) were found to be sensitive to the number of meansquared velocity gradient terms included and not whether the turbulence was assumed to adhere to isotropy or axisymmetry. The resolution dependency of different turbulence statistics was assessed with a procedure that does not rely on the dissipation scale η. The streamwise evolution of the TKE components and ε was found to collapse across grids when the rod diameter was included in the normalisation. We argue that this should be the case between all regular grids when the other relevant dimensionless quantities are matched and the flow has become homogeneous across the stream. Two-point space correlation functions at x/M = 1 show evidence of complex wake interactions which exhibit a strong Reynolds number dependence. However, these changes in initial conditions disappear indicating rapid cross-stream homogenisation. On the other hand, isotropy was, as expected, not found to be established by x/M = 12 for any case studied. © Springer-Verlag 2012.
Resumo:
We present a new co-clustering problem of images and visual features. The problem involves a set of non-object images in addition to a set of object images and features to be co-clustered. Co-clustering is performed in a way that maximises discrimination of object images from non-object images, thus emphasizing discriminative features. This provides a way of obtaining perceptual joint-clusters of object images and features. We tackle the problem by simultaneously boosting multiple strong classifiers which compete for images by their expertise. Each boosting classifier is an aggregation of weak-learners, i.e. simple visual features. The obtained classifiers are useful for object detection tasks which exhibit multimodalities, e.g. multi-category and multi-view object detection tasks. Experiments on a set of pedestrian images and a face data set demonstrate that the method yields intuitive image clusters with associated features and is much superior to conventional boosting classifiers in object detection tasks.
Resumo:
Understanding how and why changes propagate during engineering design is critical because most products and systems emerge from predecessors and not through clean sheet design. This paper applies change propagation analysis methods and extends prior reasoning through examination of a large data set from industry including 41,500 change requests, spanning 8 years during the design of a complex sensor system. Different methods are used to analyze the data and the results are compared to each other and evaluated in the context of previous findings. In particular the networks of connected parent, child and sibling changes are resolved over time and mapped to 46 subsystem areas. A normalized change propagation index (CPI) is then developed, showing the relative strength of each area on the absorber-multiplier spectrum between -1 and +1. Multipliers send out more changes than they receive and are good candidates for more focused change management. Another interesting finding is the quantitative confirmation of the "ripple" change pattern. Unlike the earlier prediction, however, it was found that the peak of cyclical change activity occurred late in the program driven by systems integration and functional testing. Patterns emerged from the data and offer clear implications for technical change management approaches in system design. Copyright © 2007 by ASME.
Resumo:
Atlases and statistical models play important roles in the personalization and simulation of cardiac physiology. For the study of the heart, however, the construction of comprehensive atlases and spatio-temporal models is faced with a number of challenges, in particular the need to handle large and highly variable image datasets, the multi-region nature of the heart, and the presence of complex as well as small cardiovascular structures. In this paper, we present a detailed atlas and spatio-temporal statistical model of the human heart based on a large population of 3D+time multi-slice computed tomography sequences, and the framework for its construction. It uses spatial normalization based on nonrigid image registration to synthesize a population mean image and establish the spatial relationships between the mean and the subjects in the population. Temporal image registration is then applied to resolve each subject-specific cardiac motion and the resulting transformations are used to warp a surface mesh representation of the atlas to fit the images of the remaining cardiac phases in each subject. Subsequently, we demonstrate the construction of a spatio-temporal statistical model of shape such that the inter-subject and dynamic sources of variation are suitably separated. The framework is applied to a 3D+time data set of 138 subjects. The data is drawn from a variety of pathologies, which benefits its generalization to new subjects and physiological studies. The obtained level of detail and the extendability of the atlas present an advantage over most cardiac models published previously. © 1982-2012 IEEE.
Resumo:
OBJECTIVE: This work is concerned with the creation of three-dimensional (3D) extended-field-of-view ultrasound from a set of volumes acquired using a mechanically swept 3D probe. 3D volumes of ultrasound data can be registered by attaching a position sensor to the probe; this can be an inconvenience in a clinical setting. A position sensor can also cause some misalignment due to patient movement and respiratory motion. We propose a combination of three-degrees-of-freedom image registration and an unobtrusively integrated inertial sensor for measuring orientation. The aim of this research is to produce a reliable and portable ultrasound system that is able to register 3D volumes quickly, making it suitable for clinical use. METHOD: As part of a feasibility study we recruited 28 pregnant females attending for routine obstetric scans to undergo 3D extended-field-of-view ultrasound. A total of 49 data sets were recorded. Each registered data set was assessed for correct alignment of each volume by two independent observers. RESULTS: In 77-83% of the data sets more than four consecutive volumes registered. The successful registration relies on good overlap between volumes and is adversely affected by advancing gestational age and foetal movement. CONCLUSION: The development of reliable 3D extended-field-of-view ultrasound may help ultrasound practitioners to demonstrate the anatomical relation of pathology and provide a convenient way to store data.
Resumo:
This paper proposes a hierarchical probabilistic model for ordinal matrix factorization. Unlike previous approaches, we model the ordinal nature of the data and take a principled approach to incorporating priors for the hidden variables. Two algorithms are presented for inference, one based on Gibbs sampling and one based on variational Bayes. Importantly, these algorithms may be implemented in the factorization of very large matrices with missing entries. The model is evaluated on a collaborative filtering task, where users have rated a collection of movies and the system is asked to predict their ratings for other movies. The Netflix data set is used for evaluation, which consists of around 100 million ratings. Using root mean-squared error (RMSE) as an evaluation metric, results show that the suggested model outperforms alternative factorization techniques. Results also show how Gibbs sampling outperforms variational Bayes on this task, despite the large number of ratings and model parameters. Matlab implementations of the proposed algorithms are available from cogsys.imm.dtu.dk/ordinalmatrixfactorization.
Resumo:
This paper addresses the design of mobile sensor networks for optimal data collection. The development is strongly motivated by the application to adaptive ocean sampling for an autonomous ocean observing and prediction system. A performance metric, used to derive optimal paths for the network of mobile sensors, defines the optimal data set as one which minimizes error in a model estimate of the sampled field. Feedback control laws are presented that stably coordinate sensors on structured tracks that have been optimized over a minimal set of parameters. Optimal, closed-loop solutions are computed in a number of low-dimensional cases to illustrate the methodology. Robustness of the performance to the influence of a steady flow field on relatively slow-moving mobile sensors is also explored © 2006 IEEE.
Resumo:
Mitigation plans to combat climate change depend on the combined implementation of many abatement options, but the options interact. Published anthropogenic emissions inventories are disaggregated by gas, sector, country, or final energy form. This allows the assessment of novel energy supply options, but is insufficient for understanding how options for efficiency and demand reduction interact. A consistent framework for understanding the drivers of emissions is therefore developed, with a set of seven complete inventories reflecting all technical options for mitigation connected through lossless allocation matrices. The required data set is compiled and calculated from a wide range of industry, government, and academic reports. The framework is used to create a global Sankey diagram to relate human demand for services to anthropogenic emissions. The application of this framework is demonstrated through a prediction of per-capita emissions based on service demand in different countries, and through an example showing how the "technical potentials" of a set of separate mitigation options should be combined.
Determination of the rheological parameters of self-compacting concrete matrix using slump flow test
Resumo:
The classification of a concrete mixture as self-compacting (SCC) is performed by a series of empirical characterization tests that have been designed to assess not only the flowability of the mixture but also its segregation resistance and filling ability. The objective of the present work is to correlate the rheological parameters of SCC matrix, yield stress and plastic viscosity, to slump flow measurements. The focus of the slump flow test investigation was centered on the fully yielded flow regime and an empirical model relating the yield stress to material and flow parameters is proposed. Our experimental data revealed that the time for a spread of 500 mm which is used in engineering practice as reference for measurement parameters, is an arbitrary choice. Our findings indicate that the non-dimensional final spread is linearly related to the non-dimensional yield-stress. Finally, there are strong indications that the non-dimensional viscosity of the mixture is associated with the non-dimensional final spread as well as the stopping time of the slump flow; this experimental data set suggests an exponential decay of the final spread and stopping time with viscosity. © Appl. Rheol.