993 resultados para data smoothing


Relevância:

70.00% 70.00%

Publicador:

Resumo:

This paper describes a new approach to multivariate scattered data smoothing. It is assumed that the data are generated by a Lipschitz continuous function f, and include random noise to be filtered out. The proposed approach uses known, or estimated value of the Lipschitz constant of f, and forces the data to be consistent with the Lipschitz properties of f. Depending on the assumptions about the distribution of the random noise, smoothing is reduced to a standard quadratic or a linear programming problem. We discuss an efficient algorithm which eliminates the redundant inequality constraints. Numerical experiments illustrate applicability and efficiency of the method. This approach provides an efficient new tool of multivariate scattered data approximation.

Relevância:

70.00% 70.00%

Publicador:

Resumo:

Chromatographic detection responses are recorded digitally. A peak is represented ideally by a Guassian distribution. Raising a Guassian distribution to the power ‘n’ increases the height of the peak to that power, but decreases the standard deviation by √n. Hence there is an increasing disparity in detection responses as the signal moves from low level noise, with a corresponding decrease in peak width. This increases the S/N ratio and increases peak to peak resolution. The ramifications of these factors are that poor resolution in complex chromatographic data can be improved, and low signal responses embedded at near noise levels can be enhanced. The application of this data treatment process is potentially very useful in 2D-HPLC where sample dilution occurs between dimension, reducing signal response, and in the application of post-reaction detection methods, where band broadening is increased by virtue of reaction coils. In this work power functions applied to chromatographic data are discussed in the context of (a) complex separation problems, (b) 2D-HPLC separations, and (c) post-column reaction detectors.

Relevância:

70.00% 70.00%

Publicador:

Resumo:

Twomultidimensional HPLC separations of an Australian red wine are presented, >70% of the available separation space was used. A porous graphitic carbon (PGC) stationary phase was used as the first dimension in both separations with both RP core–shell and hydrophilic interaction chromatography fully porous columns used separately in the second dimension. To overcome peak analysis problems caused by signal noise and low detection limits, the data were pre-processed with penalised least-squares smoothing. The PGC × RP combination separated 85 peaks with a spreading angle of 71 and the PGC × hydrophilic interaction chromatography separated 207 peaks with a spreading angle of 80. Both 2D-HPLC steps were completed in 76 min using a comprehensive stop-and-go approach. A smoothing step was added to peak-picking processes and was able to greatly reduce the number of false peaks present due to noise in the chromatograms. The required thresholds were not able to ignore the noise because of the small magnitude of the peaks; 1874 peaks were located in the non-smoothed PGC × RP separation that reduced to 227 peaks after smoothing was included.

Relevância:

70.00% 70.00%

Publicador:

Resumo:

Accurate measurement of intervertebral kinematics of the cervical spine can support the diagnosis of widespread diseases related to neck pain, such as chronic whiplash dysfunction, arthritis, and segmental degeneration. The natural inaccessibility of the spine, its complex anatomy, and the small range of motion only permit concise measurement in vivo. Low dose X-ray fluoroscopy allows time-continuous screening of cervical spine during patient's spontaneous motion. To obtain accurate motion measurements, each vertebra was tracked by means of image processing along a sequence of radiographic images. To obtain a time-continuous representation of motion and to reduce noise in the experimental data, smoothing spline interpolation was used. Estimation of intervertebral motion for cervical segments was obtained by processing patient's fluoroscopic sequence; intervertebral angle and displacement and the instantaneous centre of rotation were computed. The RMS value of fitting errors resulted in about 0.2 degree for rotation and 0.2 mm for displacements. © 2013 Paolo Bifulco et al.

Relevância:

70.00% 70.00%

Publicador:

Resumo:

The accurate and reliable estimation of travel time based on point detector data is needed to support Intelligent Transportation System (ITS) applications. It has been found that the quality of travel time estimation is a function of the method used in the estimation and varies for different traffic conditions. In this study, two hybrid on-line travel time estimation models, and their corresponding off-line methods, were developed to achieve better estimation performance under various traffic conditions, including recurrent congestion and incidents. The first model combines the Mid-Point method, which is a speed-based method, with a traffic flow-based method. The second model integrates two speed-based methods: the Mid-Point method and the Minimum Speed method. In both models, the switch between travel time estimation methods is based on the congestion level and queue status automatically identified by clustering analysis. During incident conditions with rapidly changing queue lengths, shock wave analysis-based refinements are applied for on-line estimation to capture the fast queue propagation and recovery. Travel time estimates obtained from existing speed-based methods, traffic flow-based methods, and the models developed were tested using both simulation and real-world data. The results indicate that all tested methods performed at an acceptable level during periods of low congestion. However, their performances vary with an increase in congestion. Comparisons with other estimation methods also show that the developed hybrid models perform well in all cases. Further comparisons between the on-line and off-line travel time estimation methods reveal that off-line methods perform significantly better only during fast-changing congested conditions, such as during incidents. The impacts of major influential factors on the performance of travel time estimation, including data preprocessing procedures, detector errors, detector spacing, frequency of travel time updates to traveler information devices, travel time link length, and posted travel time range, were investigated in this study. The results show that these factors have more significant impacts on the estimation accuracy and reliability under congested conditions than during uncongested conditions. For the incident conditions, the estimation quality improves with the use of a short rolling period for data smoothing, more accurate detector data, and frequent travel time updates.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Background: Malignancies arising in the large bowel cause the second largest number of deaths from cancer in the Western World. Despite progresses made during the last decades, colorectal cancer remains one of the most frequent and deadly neoplasias in the western countries. Methods: A genomic study of human colorectal cancer has been carried out on a total of 31 tumoral samples, corresponding to different stages of the disease, and 33 non-tumoral samples. The study was carried out by hybridisation of the tumour samples against a reference pool of non-tumoral samples using Agilent Human 1A 60- mer oligo microarrays. The results obtained were validated by qRT-PCR. In the subsequent bioinformatics analysis, gene networks by means of Bayesian classifiers, variable selection and bootstrap resampling were built. The consensus among all the induced models produced a hierarchy of dependences and, thus, of variables. Results: After an exhaustive process of pre-processing to ensure data quality–lost values imputation, probes quality, data smoothing and intraclass variability filtering–the final dataset comprised a total of 8, 104 probes. Next, a supervised classification approach and data analysis was carried out to obtain the most relevant genes. Two of them are directly involved in cancer progression and in particular in colorectal cancer. Finally, a supervised classifier was induced to classify new unseen samples. Conclusions: We have developed a tentative model for the diagnosis of colorectal cancer based on a biomarker panel. Our results indicate that the gene profile described herein can discriminate between non-cancerous and cancerous samples with 94.45% accuracy using different supervised classifiers (AUC values in the range of 0.997 and 0.955).

Relevância:

60.00% 60.00%

Publicador:

Resumo:

One of the most attractive features of derivative spectrometry is its higher resolving power. In the present power, numerical derivative techniques are evaluated from the viewpoint of increase in selectivity, the latter being expressed in terms of the interferent equivalent concentration (IEC). Typical spectral interferences are covered, including flat background, sloped background, simple curved background and various types of line overlap with different overlapping degrees, which were defined as the ratio of the net interfering signal at the analysis wavelength to the peak signal of the interfering line. the IECs in the derivative spectra are decreased by one to two order of magnitudes compared to those in the original spectra, and in the most cases, assume values below the conventional detection limits. The overlapping degree is the dominant factor that determines whether an analysis line can be resolved from an interfering line with the derivative techniques. Generally, the second derivative technique is effective only for line overlap with an overlapping degree of less than 0.8. The effects of other factors such as line shape, data smoothing, step size and the intensity ratio of analyte to interferent on the performance of the derivative techniques are also discussed. All results are illustrated with practical examples.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Monotonicity preserving interpolation and approximation have received substantial attention in the last thirty years because of their numerous applications in computer aided-design, statistics, and machine learning [9, 10, 19]. Constrained splines are particularly popular because of their flexibility in modeling different geometrical shapes, sound theoretical properties, and availability of numerically stable algorithms [9,10,26]. In this work we examine parallelization and adaptation for GPUs of a few algorithms of monotone spline interpolation and data smoothing, which arose in the context of estimating probability distributions.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

The purpose of this study was to identify the validity of an upper-body mounted accelerometer to measure peak acceleration during high-intensity treadmill running. A twelve camera motion analysis (MA) system was used as the criterion measure with markers placed on and close to the accelerometer. Ten peak impacts per participant were compared (n = 390). All accelerometer values were significantly different between the MA unit and T6 reflective marker’s acceleration data. Smoothing accelerometer data at 8 and 6 Hz provides an acceptable indirect measure of peak impact acceleration performed during high-intensity running. Consequently, smoothing algorithms should be incorporated into the commercially available software that the devices are supplied with.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Three comprehensive one-dimensional simulators were used on the same PC to simulate the dynamics of different electrophoretic configurations, including two migrating hybrid boundaries, an isotachophoretic boundary and the zone electrophoretic separation of ten monovalent anions. Two simulators, SIMUL5 and GENTRANS, use a uniform grid, while SPRESSO uses a dynamic adaptive grid. The simulators differ in the way components are handled. SIMUL5 and SPRESSO feature one equation for all components, whereas GENTRANS is based on the use of separate modules for the different types of monovalent components, a module for multivalent components and a module for proteins. The code for multivalent components is executed more slowly compared to those for monovalent components. Furthermore, with SIMUL5, the computational time interval becomes smaller when it is operated with a reduced calculation space that features moving borders, whereas GENTRANS offers the possibility of using data smoothing (removal of negative concentrations), which can avoid numerical oscillations and speed up a simulation. SPRESSO with its adaptive grid could be employed to simulate the same configurations with smaller numbers of grid points and thus is faster in certain but not all cases. The data reveal that simulations featuring a large number of monovalent components distributed such that a high mesh is required throughout a large proportion of the column are fastest executed with GENTRANS.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Background Malignancies arising in the large bowel cause the second largest number of deaths from cancer in the Western World. Despite progresses made during the last decades, colorectal cancer remains one of the most frequent and deadly neoplasias in the western countries. Methods A genomic study of human colorectal cancer has been carried out on a total of 31 tumoral samples, corresponding to different stages of the disease, and 33 non-tumoral samples. The study was carried out by hybridisation of the tumour samples against a reference pool of non-tumoral samples using Agilent Human 1A 60-mer oligo microarrays. The results obtained were validated by qRT-PCR. In the subsequent bioinformatics analysis, gene networks by means of Bayesian classifiers, variable selection and bootstrap resampling were built. The consensus among all the induced models produced a hierarchy of dependences and, thus, of variables. Results After an exhaustive process of pre-processing to ensure data quality--lost values imputation, probes quality, data smoothing and intraclass variability filtering--the final dataset comprised a total of 8, 104 probes. Next, a supervised classification approach and data analysis was carried out to obtain the most relevant genes. Two of them are directly involved in cancer progression and in particular in colorectal cancer. Finally, a supervised classifier was induced to classify new unseen samples. Conclusions We have developed a tentative model for the diagnosis of colorectal cancer based on a biomarker panel. Our results indicate that the gene profile described herein can discriminate between non-cancerous and cancerous samples with 94.45% accuracy using different supervised classifiers (AUC values in the range of 0.997 and 0.955)

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Background:Malignancies arising in the large bowel cause the second largest number of deaths from cancer in the Western World. Despite progresses made during the last decades, colorectal cancer remains one of the most frequent and deadly neoplasias in the western countries. Methods: A genomic study of human colorectal cancer has been carried out on a total of 31 tumoral samples, corresponding to different stages of the disease, and 33 non-tumoral samples. The study was carried out by hybridisation of the tumour samples against a reference pool of non-tumoral samples using Agilent Human 1A 60-mer oligo microarrays. The results obtained were validated by qRT-PCR. In the subsequent bioinformatics analysis, gene networks by means of Bayesian classifiers, variable selection and bootstrap resampling were built. The consensus among all the induced models produced a hierarchy of dependences and, thus, of variables. Results: After an exhaustive process of pre-processing to ensure data quality--lost values imputation, probes quality, data smoothing and intraclass variability filtering--the final dataset comprised a total of 8, 104 probes. Next, a supervised classification approach and data analysis was carried out to obtain the most relevant genes. Two of them are directly involved in cancer progression and in particular in colorectal cancer. Finally, a supervised classifier was induced to classify new unseen samples. Conclusions: We have developed a tentative model for the diagnosis of colorectal cancer based on a biomarker panel. Our results indicate that the gene profile described herein can discriminate between non-cancerous and cancerous samples with 94.45% accuracy using different supervised classifiers (AUC values in the range of 0.997 and 0.955).

Relevância:

60.00% 60.00%

Publicador:

Resumo:

An investigation is carried out into the design of a small local computer network for eventual implementation on the University of Aston campus. Microprocessors are investigated as a possible choice for use as a node controller for reasons of cost and reliability. Since the network will be local, high speed lines of megabit order are proposed. After an introduction to several well known networks, various aspects of networks are discussed including packet switching, functions of a node and host-node protocol. Chapter three develops the network philosophy with an introduction to microprocessors. Various organisations of microprocessors into multicomputer and multiprocessor systems are discussed, together with methods of achieving reliabls computing. Chapter four presents the simulation model and its implentation as a computer program. The major modelling effort is to study the behaviour of messages queueing for access to the network and the message delay experienced on the network. Use is made of spectral analysis to determine the sampling frequency while Sxponentially Weighted Noving Averages are used for data smoothing.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

2002 Mathematics Subject Classification: 62M10.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Neural data are inevitably contaminated by noise. When such noisy data are subjected to statistical analysis, misleading conclusions can be reached. Here we attempt to address this problem by applying a state-space smoothing method, based on the combined use of the Kalman filter theory and the Expectation–Maximization algorithm, to denoise two datasets of local field potentials recorded from monkeys performing a visuomotor task. For the first dataset, it was found that the analysis of the high gamma band (60–90 Hz) neural activity in the prefrontal cortex is highly susceptible to the effect of noise, and denoising leads to markedly improved results that were physiologically interpretable. For the second dataset, Granger causality between primary motor and primary somatosensory cortices was not consistent across two monkeys and the effect of noise was suspected. After denoising, the discrepancy between the two subjects was significantly reduced.