840 resultados para Image-based cytometry
Resumo:
Synchrotron radiation X-ray tomographic microscopy is a nondestructive method providing ultra-high-resolution 3D digital images of rock microstructures. We describe this method and, to demonstrate its wide applicability, we present 3D images of very different rock types: Berea sandstone, Fontainebleau sandstone, dolomite, calcitic dolomite, and three-phase magmatic glasses. For some samples, full and partial saturation scenarios are considered using oil, water, and air. The rock images precisely reveal the 3D rock microstructure, the pore space morphology, and the interfaces between fluids saturating the same pore. We provide the raw image data sets as online supplementary material, along with laboratory data describing the rock properties. By making these data sets available to other research groups, we aim to stimulate work based on digital rock images of high quality and high resolution. We also discuss and suggest possible applications and research directions that can be pursued on the basis of our data.
Resumo:
BACKGROUND Functional brain images such as Single-Photon Emission Computed Tomography (SPECT) and Positron Emission Tomography (PET) have been widely used to guide the clinicians in the Alzheimer's Disease (AD) diagnosis. However, the subjectivity involved in their evaluation has favoured the development of Computer Aided Diagnosis (CAD) Systems. METHODS It is proposed a novel combination of feature extraction techniques to improve the diagnosis of AD. Firstly, Regions of Interest (ROIs) are selected by means of a t-test carried out on 3D Normalised Mean Square Error (NMSE) features restricted to be located within a predefined brain activation mask. In order to address the small sample-size problem, the dimension of the feature space was further reduced by: Large Margin Nearest Neighbours using a rectangular matrix (LMNN-RECT), Principal Component Analysis (PCA) or Partial Least Squares (PLS) (the two latter also analysed with a LMNN transformation). Regarding the classifiers, kernel Support Vector Machines (SVMs) and LMNN using Euclidean, Mahalanobis and Energy-based metrics were compared. RESULTS Several experiments were conducted in order to evaluate the proposed LMNN-based feature extraction algorithms and its benefits as: i) linear transformation of the PLS or PCA reduced data, ii) feature reduction technique, and iii) classifier (with Euclidean, Mahalanobis or Energy-based methodology). The system was evaluated by means of k-fold cross-validation yielding accuracy, sensitivity and specificity values of 92.78%, 91.07% and 95.12% (for SPECT) and 90.67%, 88% and 93.33% (for PET), respectively, when a NMSE-PLS-LMNN feature extraction method was used in combination with a SVM classifier, thus outperforming recently reported baseline methods. CONCLUSIONS All the proposed methods turned out to be a valid solution for the presented problem. One of the advances is the robustness of the LMNN algorithm that not only provides higher separation rate between the classes but it also makes (in combination with NMSE and PLS) this rate variation more stable. In addition, their generalization ability is another advance since several experiments were performed on two image modalities (SPECT and PET).
Resumo:
Evaluation of segmentation methods is a crucial aspect in image processing, especially in the medical imaging field, where small differences between segmented regions in the anatomy can be of paramount importance. Usually, segmentation evaluation is based on a measure that depends on the number of segmented voxels inside and outside of some reference regions that are called gold standards. Although some other measures have been also used, in this work we propose a set of new similarity measures, based on different features, such as the location and intensity values of the misclassified voxels, and the connectivity and the boundaries of the segmented data. Using the multidimensional information provided by these measures, we propose a new evaluation method whose results are visualized applying a Principal Component Analysis of the data, obtaining a simplified graphical method to compare different segmentation results. We have carried out an intensive study using several classic segmentation methods applied to a set of MRI simulated data of the brain with several noise and RF inhomogeneity levels, and also to real data, showing that the new measures proposed here and the results that we have obtained from the multidimensional evaluation, improve the robustness of the evaluation and provides better understanding about the difference between segmentation methods.
Resumo:
Mosaics have been commonly used as visual maps for undersea exploration and navigation. The position and orientation of an underwater vehicle can be calculated by integrating the apparent motion of the images which form the mosaic. A feature-based mosaicking method is proposed in this paper. The creation of the mosaic is accomplished in four stages: feature selection and matching, detection of points describing the dominant motion, homography computation and mosaic construction. In this work we demonstrate that the use of color and textures as discriminative properties of the image can improve, to a large extent, the accuracy of the constructed mosaic. The system is able to provide 3D metric information concerning the vehicle motion using the knowledge of the intrinsic parameters of the camera while integrating the measurements of an ultrasonic sensor. The experimental results of real images have been tested on the GARBI underwater vehicle
Resumo:
Abstract :This article examines the interplay of text and image in The Fairy Tales of Charles Perrault (1977), translated by Angela Carter and illustrated by Martin Ware, as a form of intersemiotic dialogue that sheds new light on Carter's work. It argues that Ware's highly original artwork based on the translation not only calls into question the association of fairy tales with children's literature (which still characterizes Carter's translation), but also captures an essential if heretofore neglected aspect of Carter's creative process, namely the dynamics between translating, illustrating and rewriting classic tales. Several elements from Ware's illustrations are indeed taken up and elaborated on in The Bloody Chamber and Other Stories (1979), the collection of "stories about fairy stories" that made Carter famous. These include visual details and strategies that she transposed to the realm of writing, giving rise to reflections on the relation between visuality and textuality.RésuméCet article considère l'interaction du texte et de l'image dans les contes de Perrault traduits par Angela Carter et illustrés par Martin Ware (The Fairy Tales of Charles Perrault, 1977) comme une forme de dialogue intersémiotique particulièrement productif. Il démontre que les illustrations originales de Ware ne mettent pas seulement en question l'assimilation des contes à la littérature de jeunesse (qui est encore la perspective adoptée par la traductrice dans ce livre), mais permettent aussi de saisir un aspect essentiel bien que jusque là ignoré du procession de création dans l'oeuvre de Carter, à savoir la dynamique qui lie la traduction, l'illustration et la réécriture des contes classiques. Plusieurs éléments des illustrations de Ware sont ainsi repris et élaborés dans The Bloody Chamber and Other Stories (1979), la collection de "stories about fairy stories" qui rendit Carter célèbre. La transposition de détails et de stratégies visuelles dans l'écriture donnent ainsi l'occasion de réflexions sur les rapports entre la visualité et la textualité.
Resumo:
Generation of tumor-antigen specific CD4(+) T-helper (T(H)) lines through in vitro priming is of interest for adoptive cell therapy of cancer, but the development of this approach has been limited by the lack of appropriate tools to identify and isolate low frequency tumor antigen-specific CD4(+) T cells. Here, we have used recently developed MHC class II/peptide tetramers incorporating an immunodominant peptide from NY-ESO-1 (ESO), a tumor antigen frequently expressed in different human solid and hematologic cancers, to implement an in vitro priming platform allowing the generation of ESO-specific T(H) lines. We isolated phenotypically defined CD4(+) T-cell subpopulations from circulating lymphocytes of DR52b(+) healthy donors by flow cytometry cell sorting and stimulated them in vitro with peptide ESO(119-143), autologous APC and IL-2. We assessed the frequency of ESO-specific cells in the cultures by staining with DR52b/ESO(119-143) tetramers (ESO-tetramers) and TCR repertoire of ESO-tetramer(+) cells by co-staining with TCR variable β chain (BV) specific antibodies. We isolated ESO-tetramer(+) cells by flow cytometry cell sorting and expanded them with PHA, APC and IL-2 to generate ESO-specific T(H) lines. We characterized the lines for antigen recognition, by stimulation with ESO peptide or recombinant protein, cytokine production, by intracellular staining using specific antibodies, and alloreactivity, by stimulation with allo-APC. Using this approach, we could consistently generate ESO-tetramer(+) T(H) lines from conventional CD4(+)CD25(-) naïve and central memory populations, but not from effector memory populations or CD4(+)CD25(+) Treg. In vitro primed T(H) lines recognized ESO with affinities comparable to ESO-tetramer(+) cells from patients immunized with an ESO vaccine and used a similar TCR repertoire. In this study, using MHC class II/ESO tetramers, we have implemented an in vitro priming platform allowing the generation of ESO-monospecific polyclonal T(H) lines from non-immune individuals. This is an approach that is of potential interest for adoptive cell therapy of patients bearing ESO-expressing cancers.
Resumo:
Nowadays, the joint exploitation of images acquired daily by remote sensing instruments and of images available from archives allows a detailed monitoring of the transitions occurring at the surface of the Earth. These modifications of the land cover generate spectral discrepancies that can be detected via the analysis of remote sensing images. Independently from the origin of the images and of type of surface change, a correct processing of such data implies the adoption of flexible, robust and possibly nonlinear method, to correctly account for the complex statistical relationships characterizing the pixels of the images. This Thesis deals with the development and the application of advanced statistical methods for multi-temporal optical remote sensing image processing tasks. Three different families of machine learning models have been explored and fundamental solutions for change detection problems are provided. In the first part, change detection with user supervision has been considered. In a first application, a nonlinear classifier has been applied with the intent of precisely delineating flooded regions from a pair of images. In a second case study, the spatial context of each pixel has been injected into another nonlinear classifier to obtain a precise mapping of new urban structures. In both cases, the user provides the classifier with examples of what he believes has changed or not. In the second part, a completely automatic and unsupervised method for precise binary detection of changes has been proposed. The technique allows a very accurate mapping without any user intervention, resulting particularly useful when readiness and reaction times of the system are a crucial constraint. In the third, the problem of statistical distributions shifting between acquisitions is studied. Two approaches to transform the couple of bi-temporal images and reduce their differences unrelated to changes in land cover are studied. The methods align the distributions of the images, so that the pixel-wise comparison could be carried out with higher accuracy. Furthermore, the second method can deal with images from different sensors, no matter the dimensionality of the data nor the spectral information content. This opens the doors to possible solutions for a crucial problem in the field: detecting changes when the images have been acquired by two different sensors.
Resumo:
In many European countries, image quality for digital x-ray systems used in screening mammography is currently specified using a threshold-detail detectability method. This is a two-part study that proposes an alternative method based on calculated detectability for a model observer: the first part of the work presents a characterization of the systems. Eleven digital mammography systems were included in the study; four computed radiography (CR) systems, and a group of seven digital radiography (DR) detectors, composed of three amorphous selenium-based detectors, three caesium iodide scintillator systems and a silicon wafer-based photon counting system. The technical parameters assessed included the system response curve, detector uniformity error, pre-sampling modulation transfer function (MTF), normalized noise power spectrum (NNPS) and detective quantum efficiency (DQE). Approximate quantum noise limited exposure range was examined using a separation of noise sources based upon standard deviation. Noise separation showed that electronic noise was the dominant noise at low detector air kerma for three systems; the remaining systems showed quantum noise limited behaviour between 12.5 and 380 µGy. Greater variation in detector MTF was found for the DR group compared to the CR systems; MTF at 5 mm(-1) varied from 0.08 to 0.23 for the CR detectors against a range of 0.16-0.64 for the DR units. The needle CR detector had a higher MTF, lower NNPS and higher DQE at 5 mm(-1) than the powder CR phosphors. DQE at 5 mm(-1) ranged from 0.02 to 0.20 for the CR systems, while DQE at 5 mm(-1) for the DR group ranged from 0.04 to 0.41, indicating higher DQE for the DR detectors and needle CR system than for the powder CR phosphor systems. The technical evaluation section of the study showed that the digital mammography systems were well set up and exhibiting typical performance for the detector technology employed in the respective systems.
Resumo:
We propose a method for brain atlas deformation inpresence of large space-occupying tumors, based on an apriori model of lesion growth that assumes radialexpansion of the lesion from its starting point. First,an affine registration brings the atlas and the patientinto global correspondence. Then, the seeding of asynthetic tumor into the brain atlas provides a templatefor the lesion. Finally, the seeded atlas is deformed,combining a method derived from optical flow principlesand a model of lesion growth (MLG). Results show that themethod can be applied to the automatic segmentation ofstructures and substructures in brains with grossdeformation, with important medical applications inneurosurgery, radiosurgery and radiotherapy.
Resumo:
This paper proposes a novel approach for the analysis of illicit tablets based on their visual characteristics. In particular, the paper concentrates on the problem of ecstasy pill seizure profiling and monitoring. The presented method extracts the visual information from pill images and builds a representation of it, i.e. it builds a pill profile based on the pill visual appearance. Different visual features are used to build different image similarity measures, which are the basis for a pill monitoring strategy based on both discriminative and clustering models. The discriminative model permits to infer whether two pills come from the same seizure, while the clustering models groups of pills that share similar visual characteristics. The resulting clustering structure allows to perform a visual identification of the relationships between different seizures. The proposed approach was evaluated using a data set of 621 Ecstasy pill pictures. The results demonstrate that this is a feasible and cost effective method for performing pill profiling and monitoring.
Resumo:
Demosaicking is a particular case of interpolation problems where, from a scalar image in which each pixel has either the red, the green or the blue component, we want to interpolate the full-color image. State-of-the-art demosaicking algorithms perform interpolation along edges, but these edges are estimated locally. We propose a level-set-based geometric method to estimate image edges, inspired by the image in-painting literature. This method has a time complexity of O(S) , where S is the number of pixels in the image, and compares favorably with the state-of-the-art algorithms both visually and in most relevant image quality measures.
Resumo:
PURPOSE: To investigate magnetization transfer (MT) effects as a new source of contrast for imaging and tracking of peripheral foot nerves. MATERIALS AND METHODS: Two sets of 3D spoiled gradient-echo images acquired with and without a saturation pulse were used to generate MT ratio (MTR) maps of 260 μm in-plane resolution for eight volunteers at 3T. Scan parameters were adjusted to minimize signal loss due to T2 dephasing, and a dedicated coil was used to improve the inherently low signal-to-noise ratio of small voxels. Resulting MTR values in foot nerves were compared with those in surrounding muscle tissue. RESULTS: Average MTR values for muscle (45.5 ± 1.4%) and nerve (21.4 ± 3.1%) were significantly different (P < 0.0001). In general, the difference in MTR values was sufficiently large to allow for intensity-based segmentation and tracking of foot nerves in individual subjects. This procedure was termed MT-based 3D visualization. CONCLUSION: The MTR serves as a new source of contrast for imaging of peripheral foot nerves and provides a means for high spatial resolution tracking of these structures. The proposed methodology is directly applicable on standard clinical MR scanners and could be applied to systemic pathologies, such as diabetes.
Resumo:
BACKGROUND: Radiation dose exposure is of particular concern in children due to the possible harmful effects of ionizing radiation. The adaptive statistical iterative reconstruction (ASIR) method is a promising new technique that reduces image noise and produces better overall image quality compared with routine-dose contrast-enhanced methods. OBJECTIVE: To assess the benefits of ASIR on the diagnostic image quality in paediatric cardiac CT examinations. MATERIALS AND METHODS: Four paediatric radiologists based at two major hospitals evaluated ten low-dose paediatric cardiac examinations (80 kVp, CTDI(vol) 4.8-7.9 mGy, DLP 37.1-178.9 mGy·cm). The average age of the cohort studied was 2.6 years (range 1 day to 7 years). Acquisitions were performed on a 64-MDCT scanner. All images were reconstructed at various ASIR percentages (0-100%). For each examination, radiologists scored 19 anatomical structures using the relative visual grading analysis method. To estimate the potential for dose reduction, acquisitions were also performed on a Catphan phantom and a paediatric phantom. RESULTS: The best image quality for all clinical images was obtained with 20% and 40% ASIR (p < 0.001) whereas with ASIR above 50%, image quality significantly decreased (p < 0.001). With 100% ASIR, a strong noise-free appearance of the structures reduced image conspicuity. A potential for dose reduction of about 36% is predicted for a 2- to 3-year-old child when using 40% ASIR rather than the standard filtered back-projection method. CONCLUSION: Reconstruction including 20% to 40% ASIR slightly improved the conspicuity of various paediatric cardiac structures in newborns and children with respect to conventional reconstruction (filtered back-projection) alone.
Resumo:
Introduction: Quantitative measures of degree of lumbar spinal stenosis (LSS) such as antero-posterior diameter of the canal or dural sac cross sectional area vary widely and do not correlate with clinical symptoms or results of surgical decompression. In an effort to improve quantification of stenosis we have developed a grading system based on the morphology of the dural sac and its contents as seen on T2 axial images. The grading comprises seven categories ranging form normal to the most severe stenosis and takes into account the ratio of rootlet/CSF content. Material and methods: Fifty T2 axial MRI images taken at disc level from twenty seven symptomatic lumbar spinal stenosis patients who underwent decompressive surgery were classified into seven categories by five observers and reclassified 2 weeks later by the same investigators. Intra- and inter-observer reliability of the classification were assessed using Cohen's and Fleiss' kappa statistics, respectively. Results: Generally, the morphology grading system itself was well adopted by the observers. Its success in application is strongly influenced by the identification of the dural sac. The average intraobserver Cohen's kappa was 0.53 ± 0.2. The inter-observer Fleiss' kappa was 0.38 ± 0.02 in the first rating and 0.3 ± 0.03 in the second rating repeated after two weeks. Discussion: In this attempt, the teaching of the observers was limited to an introduction to the general idea of the morphology grading system and one example MRI image per category. The identification of the dimension of the dural sac may be a difficult issue in absence of complete T1 T2 MRI image series as it was the case here. The similarity of the CSF to possibly present fat on T2 images was the main reason of mismatch in the assignment of the cases to a category. The Fleiss correlation factors of the five observers are fair and the proposed morphology grading system is promising.
Resumo:
Acute brain slices are slices of brain tissue that are kept vital in vitro for further recordings and analyses. This tool is of major importance in neurobiology and allows the study of brain cells such as microglia, astrocytes, neurons and their inter/intracellular communications via ion channels or transporters. In combination with light/fluorescence microscopies, acute brain slices enable the ex vivo analysis of specific cells or groups of cells inside the slice, e.g. astrocytes. To bridge ex vivo knowledge of a cell with its ultrastructure, we developed a correlative microscopy approach for acute brain slices. The workflow begins with sampling of the tissue and precise trimming of a region of interest, which contains GFP-tagged astrocytes that can be visualised by fluorescence microscopy of ultrathin sections. The astrocytes and their surroundings are then analysed by high resolution scanning transmission electron microscopy (STEM). An important aspect of this workflow is the modification of a commercial cryo-ultramicrotome to observe the fluorescent GFP signal during the trimming process. It ensured that sections contained at least one GFP astrocyte. After cryo-sectioning, a map of the GFP-expressing astrocytes is established and transferred to correlation software installed on a focused ion beam scanning electron microscope equipped with a STEM detector. Next, the areas displaying fluorescence are selected for high resolution STEM imaging. An overview area (e.g. a whole mesh of the grid) is imaged with an automated tiling and stitching process. In the final stitched image, the local organisation of the brain tissue can be surveyed or areas of interest can be magnified to observe fine details, e.g. vesicles or gold labels on specific proteins. The robustness of this workflow is contingent on the quality of sample preparation, based on Tokuyasu's protocol. This method results in a reasonable compromise between preservation of morphology and maintenance of antigenicity. Finally, an important feature of this approach is that the fluorescence of the GFP signal is preserved throughout the entire preparation process until the last step before electron microscopy.