957 resultados para swd: Image segmentation


Relevância:

80.00% 80.00%

Publicador:

Resumo:

Circular shortest paths represent a powerful methodology for image segmentation. The circularity condition ensures that the contour found by the algorithm is closed, a natural requirement for regular objects. Several implementations have been proposed in the past that either promise closure with high probability or ensure closure strictly, but with a mild computational efficiency handicap. Circularity can be viewed as a priori information that helps recover the correct object contour. Our "observation" is that circularity is only one among many possible constraints that can be imposed on shortest paths to guide them to a desirable solution. In this contribution, we illustrate this opportunity under a volume constraint but the concept is generally applicable. We also describe several adornments to the circular shortest path algorithm that proved useful in applications. © 2011 IEEE.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

In studies of germ cell transplantation, measureing tubule diameters and counting cells from different populations using antibodies as markers are very important. Manual measurement of tubule sizes and cell counts is a tedious and sanity grinding work. In this paper, we propose a new boundary weighting based tubule detection method. We first enhance the linear features of the input image and detect the approximate centers of tubules. Next, a boundary weighting transform is applied to the polar transformed image of each tubule region and a circular shortest path is used for the boundary detection. Then, ellipse fitting is carried out for tubule selection and measurement. The algorithm has been tested on a dataset consisting of 20 images, each having about 20 tubules. Experiments show that the detection results of our algorithm are very close to the results obtained manually. © 2013 IEEE.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Numerous efforts have been dedicated to the synthesis of large-volume methacrylate monoliths for large-scale biomolecules purification but most were obstructed by the enormous release of exotherms during preparation, thereby introducing structural heterogeneity in the monolith pore system. A significant radial temperature gradient develops along the monolith thickness, reaching a terminal temperature that supersedes the maximum temperature required for structurally homogenous monoliths preparation. The enormous heat build-up is perceived to encompass the heat associated with initiator decomposition and the heat released from free radical-monomer and monomer-monomer interactions. The heat resulting from the initiator decomposition was expelled along with some gaseous fumes before commencing polymerization in a gradual addition fashion. Characteristics of 80 mL monolith prepared using this technique was compared with that of a similar monolith synthesized in a bulk polymerization mode. An extra similarity in the radial temperature profiles was observed for the monolith synthesized via the heat expulsion technique. A maximum radial temperature gradient of only 4.3°C was recorded at the center and 2.1°C at the monolith peripheral for the combined heat expulsion and gradual addition technique. The comparable radial temperature distributions obtained birthed identical pore size distributions at different radial points along the monolith thickness.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

A novel shape recognition algorithm was developed to autonomously classify the Northern Pacific Sea Star (Asterias amurenis) from benthic images that were collected by the Starbug AUV during 6km of transects in the Derwent estuary. Despite the effects of scattering, attenuation, soft focus and motion blur within the underwater images, an optimal joint classification rate of 77.5% and misclassification rate of 13.5% was achieved. The performance of algorithm was largely attributed to its ability to recognise locally deformed sea star shapes that were created during the segmentation of the distorted images.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

There is a major effort in medical imaging to develop algorithms to extract information from DTI and HARDI, which provide detailed information on brain integrity and connectivity. As the images have recently advanced to provide extraordinarily high angular resolution and spatial detail, including an entire manifold of information at each point in the 3D images, there has been no readily available means to view the results. This impedes developments in HARDI research, which need some method to check the plausibility and validity of image processing operations on HARDI data or to appreciate data features or invariants that might serve as a basis for new directions in image segmentation, registration, and statistics. We present a set of tools to provide interactive display of HARDI data, including both a local rendering application and an off-screen renderer that works with a web-based viewer. Visualizations are presented after registration and averaging of HARDI data from 90 human subjects, revealing important details for which there would be no direct way to appreciate using conventional display of scalar images.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Purpose: To quantify the uncertainties of carotid plaque morphology reconstruction based on patient-specific multispectral in vivo magnetic resonance imaging (MRI) and their impacts on the plaque stress analysis. Materials and Methods: In this study, three independent investigators were invited to reconstruct the carotid bifurcation with plaque based on MR images from two subjects to study the geometry reconstruction reproducibility. Finite element stress analyses were performed on the carotid bifurcations, as well as the models with artificially modified plaque geometries to mimic the image segmentation uncertainties, to study the impacts of the uncertainties to the stress prediction. Results: Plaque reconstruction reproducibility was generally high in the study. The uncertainties among interobservers are around one or the subpixel level. It also shows that the predicted stress is relatively less sensitive to the arterial wall segmentation uncertainties, and more affected by the accuracy of lipid region definition. For a model with lipid core region artificially increased by adding one pixel on the lipid region boundary, it will significantly increase the maximum Von Mises Stress in fibrous cap (>100%) compared with the baseline model for all subjects. Conclusion: The current in vivo MRI in the carotid plaque could provide useful and reliable information for plaque morphology. The accuracy of stress analysis based on plaque geometry is subject to MRI quality. The improved resolution/quality in plaque imaging with newly developed MRI protocols would generate more realistic stress predictions.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

The rupture of atherosclerotic plaques is known to be associated with the stresses that act on or within the arterial wall. The extreme wall tensile stress (WTS) is usually recognized as a primary trigger for the rupture of vulnerable plaque. The present study used the in-vivo high-resolution multi-spectral magnetic resonance imaging (MRI) for carotid arterial plaque morphology reconstruction. Image segmentation of different plaque components was based on the multi-spectral MRI and co-registered with different sequences for the patient. Stress analysis was performed on totally four subjects with different plaque burden by fluid-structure interaction (FSI) simulations. Wall shear stress distributions are highly related to the degree of stenosis, while the level of its magnitude is much lower than the WTS in the fibrous cap. WTS is higher in the luminal wall and lower at the outer wall, with the lowest stress at the lipid region. Local stress concentrations are well confined in the thinner fibrous cap region, and usually locating in the plaque shoulder; the introduction of relative stress variation during a cycle in the fibrous cap can be a potential indicator for plaque fatigue process in the thin fibrous cap. According to stress analysis of the four subjects, a risk assessment in terms of mechanical factors could be made, which may be helpful in clinical practice. However, more subjects with patient specific analysis are desirable for plaque-stability study.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

This thesis examines the feasibility of a forest inventory method based on two-phase sampling in estimating forest attributes at the stand or substand levels for forest management purposes. The method is based on multi-source forest inventory combining auxiliary data consisting of remote sensing imagery or other geographic information and field measurements. Auxiliary data are utilized as first-phase data for covering all inventory units. Various methods were examined for improving the accuracy of the forest estimates. Pre-processing of auxiliary data in the form of correcting the spectral properties of aerial imagery was examined (I), as was the selection of aerial image features for estimating forest attributes (II). Various spatial units were compared for extracting image features in a remote sensing aided forest inventory utilizing very high resolution imagery (III). A number of data sources were combined and different weighting procedures were tested in estimating forest attributes (IV, V). Correction of the spectral properties of aerial images proved to be a straightforward and advantageous method for improving the correlation between the image features and the measured forest attributes. Testing different image features that can be extracted from aerial photographs (and other very high resolution images) showed that the images contain a wealth of relevant information that can be extracted only by utilizing the spatial organization of the image pixel values. Furthermore, careful selection of image features for the inventory task generally gives better results than inputting all extractable features to the estimation procedure. When the spatial units for extracting very high resolution image features were examined, an approach based on image segmentation generally showed advantages compared with a traditional sample plot-based approach. Combining several data sources resulted in more accurate estimates than any of the individual data sources alone. The best combined estimate can be derived by weighting the estimates produced by the individual data sources by the inverse values of their mean square errors. Despite the fact that the plot-level estimation accuracy in two-phase sampling inventory can be improved in many ways, the accuracy of forest estimates based mainly on single-view satellite and aerial imagery is a relatively poor basis for making stand-level management decisions.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Microarrays are high throughput biological assays that allow the screening of thousands of genes for their expression. The main idea behind microarrays is to compute for each gene a unique signal that is directly proportional to the quantity of mRNA that was hybridized on the chip. A large number of steps and errors associated with each step make the generated expression signal noisy. As a result, microarray data need to be carefully pre-processed before their analysis can be assumed to lead to reliable and biologically relevant conclusions. This thesis focuses on developing methods for improving gene signal and further utilizing this improved signal for higher level analysis. To achieve this, first, approaches for designing microarray experiments using various optimality criteria, considering both biological and technical replicates, are described. A carefully designed experiment leads to signal with low noise, as the effect of unwanted variations is minimized and the precision of the estimates of the parameters of interest are maximized. Second, a system for improving the gene signal by using three scans at varying scanner sensitivities is developed. A novel Bayesian latent intensity model is then applied on these three sets of expression values, corresponding to the three scans, to estimate the suitably calibrated true signal of genes. Third, a novel image segmentation approach that segregates the fluorescent signal from the undesired noise is developed using an additional dye, SYBR green RNA II. This technique helped in identifying signal only with respect to the hybridized DNA, and signal corresponding to dust, scratch, spilling of dye, and other noises, are avoided. Fourth, an integrated statistical model is developed, where signal correction, systematic array effects, dye effects, and differential expression, are modelled jointly as opposed to a sequential application of several methods of analysis. The methods described in here have been tested only for cDNA microarrays, but can also, with some modifications, be applied to other high-throughput technologies. Keywords: High-throughput technology, microarray, cDNA, multiple scans, Bayesian hierarchical models, image analysis, experimental design, MCMC, WinBUGS.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Early identification of diabetic foot complications and their precursors is essential in preventing their devastating consequences, such as foot infection and amputation. Frequent, automatic risk assessment by an intelligent telemedicine system might be feasible and cost effective. Infrared thermography is a promising modality for such a system. The temperature differences between corresponding areas on contralateral feet are the clinically significant parameters. This asymmetric analysis is hindered by (1) foot segmentation errors, especially when the foot temperature and the ambient temperature are comparable, and by (2) different shapes and sizes between contralateral feet due to deformities or minor amputations. To circumvent the first problem, we used a color image and a thermal image acquired synchronously. Foot regions, detected in the color image, were rigidly registered to the thermal image. This resulted in 97.8% ± 1.1% sensitivity and 98.4% ± 0.5% specificity over 76 high-risk diabetic patients with manual annotation as a reference. Nonrigid landmark-based registration with Bsplines solved the second problem. Corresponding points in the two feet could be found regardless of the shapes and sizes of the feet. With that, the temperature difference of the left and right feet could be obtained.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Automated image segmentation techniques are useful tools in biological image analysis and are an essential step in tracking applications. Typically, snakes or active contours are used for segmentation and they evolve under the influence of certain internal and external forces. Recently, a new class of shape-specific active contours have been introduced, which are known as Snakuscules and Ovuscules. These contours are based on a pair of concentric circles and ellipses as the shape templates, and the optimization is carried out by maximizing a contrast function between the outer and inner templates. In this paper, we present a unified approach to the formulation and optimization of Snakuscules and Ovuscules by considering a specific form of affine transformations acting on a pair of concentric circles. We show how the parameters of the affine transformation may be optimized for, to generate either Snakuscules or Ovuscules. Our approach allows for a unified formulation and relies only on generic regularization terms and not shape-specific regularization functions. We show how the calculations of the partial derivatives may be made efficient thanks to the Green's theorem. Results on synthesized as well as real data are presented.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

This paper presents a GPU implementation of normalized cuts for road extraction problem using panchromatic satellite imagery. The roads have been extracted in three stages namely pre-processing, image segmentation and post-processing. Initially, the image is pre-processed to improve the tolerance by reducing the clutter (that mostly represents the buildings, vegetation,. and fallow regions). The road regions are then extracted using the normalized cuts algorithm. Normalized cuts algorithm is a graph-based partitioning `approach whose focus lies in extracting the global impression (perceptual grouping) of an image rather than local features. For the segmented image, post-processing is carried out using morphological operations - erosion and dilation. Finally, the road extracted image is overlaid on the original image. Here, a GPGPU (General Purpose Graphical Processing Unit) approach has been adopted to implement the same algorithm on the GPU for fast processing. A performance comparison of this proposed GPU implementation of normalized cuts algorithm with the earlier algorithm (CPU implementation) is presented. From the results, we conclude that the computational improvement in terms of time as the size of image increases for the proposed GPU implementation of normalized cuts. Also, a qualitative and quantitative assessment of the segmentation results has been projected.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Motivated by multi-distribution divergences, which originate in information theory, we propose a notion of `multipoint' kernels, and study their applications. We study a class of kernels based on Jensen type divergences and show that these can be extended to measure similarity among multiple points. We study tensor flattening methods and develop a multi-point (kernel) spectral clustering (MSC) method. We further emphasize on a special case of the proposed kernels, which is a multi-point extension of the linear (dot-product) kernel and show the existence of cubic time tensor flattening algorithm in this case. Finally, we illustrate the usefulness of our contributions using standard data sets and image segmentation tasks.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

This paper investigates the effect of particle size of sand and the surface asperities of reinforcing material on their interlocking mechanism and its influence on the interfacial shear strength under direct sliding condition. Three sands of different sizes with similar morphological characteristics and four different types of reinforcing materials with different surface features were used in this study. Interface direct shear tests on these materials were performed in a specially developed symmetric loading interface direct shear test setup. Morphological characteristics of sand particles were determined from digital image analysis and the surface roughness of the reinforcing materials was measured using an analytical expression developed for this purpose. Interface direct shear tests at three different normal stresses were carried out by shearing the sand on the reinforcing material fixed to a smooth surface. Test results revealed that the peak interfacial friction and dilation angles are hugely dependent upon the interlocking between the sand particles and the asperities of reinforcing material, which in turn depends on the relative size of sand particles and asperities. Asperity ratio (AS/D-50) of interlocking materials, which is defined as the ratio of asperity spacing of the reinforcing material and the mean particle size of sand was found to govern the interfacial shear strength with highest interfacial strength measured when the asperity ratio was equal to one, which represents the closest fitting of sand particles into the asperities. It was also understood that the surface roughness of the reinforcing material influences the shear strength to an extent, the influence being more pronounced in coarser particles. Shear bands in the interface shear tests were analysed through image segmentation technique and it was observed that the ratio of shear band thickness (t) to the median particle size (D-50) was maximum when the AS/D-50 was equal to one. (C) 2015 Elsevier Ltd. All rights reserved.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

O objetivo desta dissertação foi criar uma nova abordagem para identificar de maneira automática feições do tipo edificação em uma imagem digital. Tal identificação seria de interesse de órgãos públicos que lidam com planejamento urbano para fins de controle da ocupação humana irregular. A abordagem criada utilizou agentes de software especialistas para proceder com o processamento da segmentação e reconhecimento de feições na imagem digital. Os agentes foram programados para tratar uma imagem colorida com o padrão Red, Green e Blue (RGB). A criação desta nova abordagem teve como motivação o fato das atuais técnicas existentes de segmentação e classificação de imagens dependerem sobremaneira dos seus usuários. Em outras palavras, pretendeu-se com a abordagem em questão permitir que usuários menos técnicos pudessem interagir com um sistema classificador, sem a necessidade de profundos conhecimentos de processamento digital de imagem. Uma ferramenta protótipo foi desenvolvida para testar essa abordagem, que emprega de forma inusitada, agentes inteligentes, com testes feitos em recortes de ortofotos digitais do Município de Angra dos Reis (RJ).