900 resultados para Textural features


Relevância:

20.00% 20.00%

Publicador:

Resumo:

This paper proposes a region based image retrieval system using the local colour and texture features of image sub regions. The regions of interest (ROI) are roughly identified by segmenting the image into fixed partitions, finding the edge map and applying morphological dilation. The colour and texture features of the ROIs are computed from the histograms of the quantized HSV colour space and Gray Level co- occurrence matrix (GLCM) respectively. Each ROI of the query image is compared with same number of ROIs of the target image that are arranged in the descending order of white pixel density in the regions, using Euclidean distance measure for similarity computation. Preliminary experimental results show that the proposed method provides better retrieving result than retrieval using some of the existing methods.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Grey Level Co-occurrence Matrices (GLCM) are one of the earliest techniques used for image texture analysis. In this paper we defined a new feature called trace extracted from the GLCM and its implications in texture analysis are discussed in the context of Content Based Image Retrieval (CBIR). The theoretical extension of GLCM to n-dimensional gray scale images are also discussed. The results indicate that trace features outperform Haralick features when applied to CBIR.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This paper presents a writer identification scheme for Malayalam documents. As the accomplishment rate of a scheme is highly dependent on the features extracted from the documents, the process of feature selection and extraction is highly relevant. The paper describes a set of novel features exclusively for Malayalam language. The features were studied in detail which resulted in a comparative study of all the features. The features are fused to form the feature vector or knowledge vector. This knowledge vector is then used in all the phases of the writer identification scheme. The scheme has been tested on a test bed of 280 writers of which 50 writers having only one page, 215 writers with at least 2 pages and 15 writers with at least 4 pages. To perform a comparative evaluation of the scheme the test is conducted using WD-LBP method also. A recognition rate of around 95% was obtained for the proposed approach

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The span of writer identification extends to broad domes like digital rights administration, forensic expert decisionmaking systems, and document analysis systems and so on. As the success rate of a writer identification scheme is highly dependent on the features extracted from the documents, the phase of feature extraction and therefore selection is highly significant for writer identification schemes. In this paper, the writer identification in Malayalam language is sought for by utilizing feature extraction technique such as Scale Invariant Features Transform (SIFT).The schemes are tested on a test bed of 280 writers and performance evaluated

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The characterization and grading of glioma tumors, via image derived features, for diagnosis, prognosis, and treatment response has been an active research area in medical image computing. This paper presents a novel method for automatic detection and classification of glioma from conventional T2 weighted MR images. Automatic detection of the tumor was established using newly developed method called Adaptive Gray level Algebraic set Segmentation Algorithm (AGASA).Statistical Features were extracted from the detected tumor texture using first order statistics and gray level co-occurrence matrix (GLCM) based second order statistical methods. Statistical significance of the features was determined by t-test and its corresponding p-value. A decision system was developed for the grade detection of glioma using these selected features and its p-value. The detection performance of the decision system was validated using the receiver operating characteristic (ROC) curve. The diagnosis and grading of glioma using this non-invasive method can contribute promising results in medical image computing

Relevância:

20.00% 20.00%

Publicador:

Resumo:

HINDI

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The evolution of coast through geological time scale is dependent on the transgression-regression event subsequent to the rise or fall of sea level. This event is accounted by investigation of the vertical sediment deposition patterns and their interrelationship for paleo-enviornmental reconstruction. Different methods like sedimentological (grain size and micro-morphological) and geochemical (elemental relationship) analyses as well as radiocarbon dating are generally used to decipher the sea level changes and paleoclimatic conditions of the Quaternary sediment sequence. For the Indian coast with a coastline length of about 7500 km, studies on geological and geomorphological signatures of sea level changes during the Quaternary were reported in general by researchers during the last two decades. However, for the southwest coast of India particularily Kerala which is famous for its coastal landforms comprising of estuaries, lagoons, backwaters, coastal plains, cliffs and barrier beaches, studies pertaining to the marine transgression-regression events in the southern region are limited. The Neendakara-Kayamkulam coastal stretch in central Kerala where the coast is manifested with shore parallel Kayamkulam Lagoon on one side and shore perpendicular Ashtamudi Estuary on the other side indicating existence of an uplifted prograded coastal margin followed by barrier beaches, backwater channels, ridge and runnel topography is an ideal site for studying such events. Hence the present study has been taken up in this context to address the gap area. The location for collection of core samples representing coastal plain, estuarylagoon and offshore regions have been identified based on published literature and available sedimentary records. The objectives of the research work are:  To study the lithological variations and depositional environments of sediment cores along the coastal plain, estuary-lagoon and offshore regions between Kollam and Kayamkulam in the central Kerala coast  To study the transportation and diagenetic history of sediments in the area  To investigate the geochemical characterization of sediments and to elucidate the source-sink relationship  To understand the marine transgression-regression events and to propose a conceptual model for the region The thesis comprises of 8 chapters. The first chapter embodies the preamble for the selection and significance of this research work. The study area is introduced with details on its physiographical, geological, geomorphological, rainfall and climate information. A review of literature, compiling the research on different aspects such as physico-chemical, geomorphological, tectonics, transgression-regression events are presented in the second chapter and they are broadly classified into three viz:- International, National and Kerala. The field data collection and laboratory analyses adopted in the research work are discussed in the third chapter. For collection of sediment core samples from the coastal plains, rotary drilling method was employed whereas for the estuary-lagoon and offshore locations the gravity/piston corer method was adopted. The collected subsurficial samples were analysed for texture, surface micro-texture, elemental analysis, XRD and radiocarbon dating techniques for age determination. The fourth chapter deals with the textural analysis of the core samples collected from various predefined locations of the study area. The result reveals that the Ashtamudi Estuary is composed of silty clay to clayey type of sediments whereas offshore cores are carpeted with silty clay to relict sand. Investigation of the source of sediments deposited in the coastal plain located on either side of the estuary indicates the dominance of terrigenous to marine origin in the southern region whereas it is predominantly of marine origin towards the north. Further the hydrodynamic conditions as well as the depositional enviornment of the sediment cores are elucidated based on statistical parameters that decipher the deposition pattern at various locations viz., coastal plain (open to closed basin), Ashtamudi Estuary (partially open to restricted estuary to closed basin) and offshore (open channel). The intensity of clay minerals is also discussed. From the results of radiocarbon dating the sediment depositional environments were deciphered.The results of the microtextural study of sediment samples (quartz grains) using Scanning Electron Microscope (SEM) are presented in the fifth chapter. These results throw light on the processes of transport and diagenetic history of the detrital sediments. Based on the lithological variations, selected quartz grains of different environments were also analysed. The study indicates that the southern coastal plain sediments were transported and deposited mechanically under fluvial environment followed by diagenesis under prolonged marine incursion. But in the case of the northern coastal plain, the sediments were transported and deposited under littoral environment indicating the dominance of marine incursion through mechanical as well as chemical processes. The quartz grains of the Ashtamudi Estuary indicate fluvial origin. The surface texture features of the offshore sediments suggest that the quartz grains are of littoral origin and represent the relict beach deposits. The geochemical characterisation of sediment cores based on geochemical classification, sediment maturity, palaeo-weathering and provenance in different environments are discussed in the sixth chapter. In the seventh chapter the integration of multiproxies data along with radiocarbon dates are presented and finally evolution and depositional history based on transgression–regression events is deciphered. The eighth chapter summarizes the major findings and conclusions of the study with recommendation for future work.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The consumers are becoming more concerned about food quality, especially regarding how, when and where the foods are produced (Haglund et al., 1999; Kahl et al., 2004; Alföldi, et al., 2006). Therefore, during recent years there has been a growing interest in the methods for food quality assessment, especially in the picture-development methods as a complement to traditional chemical analysis of single compounds (Kahl et al., 2006). The biocrystallization as one of the picture-developing method is based on the crystallographic phenomenon that when crystallizing aqueous solutions of dihydrate CuCl2 with adding of organic solutions, originating, e.g., from crop samples, biocrystallograms are generated with reproducible crystal patterns (Kleber & Steinike-Hartung, 1959). Its output is a crystal pattern on glass plates from which different variables (numbers) can be calculated by using image analysis. However, there is a lack of a standardized evaluation method to quantify the morphological features of the biocrystallogram image. Therefore, the main sakes of this research are (1) to optimize an existing statistical model in order to describe all the effects that contribute to the experiment, (2) to investigate the effect of image parameters on the texture analysis of the biocrystallogram images, i.e., region of interest (ROI), color transformation and histogram matching on samples from the project 020E170/F financed by the Federal Ministry of Food, Agriculture and Consumer Protection(BMELV).The samples are wheat and carrots from controlled field and farm trials, (3) to consider the strongest effect of texture parameter with the visual evaluation criteria that have been developed by a group of researcher (University of Kassel, Germany; Louis Bolk Institute (LBI), Netherlands and Biodynamic Research Association Denmark (BRAD), Denmark) in order to clarify how the relation of the texture parameter and visual characteristics on an image is. The refined statistical model was accomplished by using a lme model with repeated measurements via crossed effects, programmed in R (version 2.1.0). The validity of the F and P values is checked against the SAS program. While getting from the ANOVA the same F values, the P values are bigger in R because of the more conservative approach. The refined model is calculating more significant P values. The optimization of the image analysis is dealing with the following parameters: ROI(Region of Interest which is the area around the geometrical center), color transformation (calculation of the 1 dimensional gray level value out of the three dimensional color information of the scanned picture, which is necessary for the texture analysis), histogram matching (normalization of the histogram of the picture to enhance the contrast and to minimize the errors from lighting conditions). The samples were wheat from DOC trial with 4 field replicates for the years 2003 and 2005, “market samples”(organic and conventional neighbors with the same variety) for 2004 and 2005, carrot where the samples were obtained from the University of Kassel (2 varieties, 2 nitrogen treatments) for the years 2004, 2005, 2006 and “market samples” of carrot for the years 2004 and 2005. The criterion for the optimization was repeatability of the differentiation of the samples over the different harvest(years). For different samples different ROIs were found, which reflect the different pictures. The best color transformation that shows efficiently differentiation is relied on gray scale, i.e., equal color transformation. The second dimension of the color transformation only appeared in some years for the effect of color wavelength(hue) for carrot treated with different nitrate fertilizer levels. The best histogram matching is the Gaussian distribution. The approach was to find a connection between the variables from textural image analysis with the different visual criteria. The relation between the texture parameters and visual evaluation criteria was limited to the carrot samples, especially, as it could be well differentiated by the texture analysis. It was possible to connect groups of variables of the texture analysis with groups of criteria from the visual evaluation. These selected variables were able to differentiate the samples but not able to classify the samples according to the treatment. Contrarily, in case of visual criteria which describe the picture as a whole there is a classification in 80% of the sample cases possible. Herewith, it clearly can find the limits of the single variable approach of the image analysis (texture analysis).

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Almost everyone sketches. People use sketches day in and day out in many different and heterogeneous fields, to share their thoughts and clarify ambiguous interpretations, for example. The media used to sketch varies from analog tools like flipcharts to digital tools like smartboards. Whereas analog tools are usually affected by insufficient editing capabilities like cut/copy/paste, digital tools greatly support these scenarios. Digital tools can be grouped into informal and formal tools. Informal tools can be understood as simple drawing environments, whereas formal tools offer sophisticated support to create, optimize and validate diagrams of a certain application domain. Most digital formal tools force users to stick to a concrete syntax and editing workflow, limiting the user’s creativity. For that reason, a lot of people first sketch their ideas using the flexibility of analog or digital informal tools. Subsequently, the sketch is "portrayed" in an appropriate digital formal tool. This work presents Scribble, a highly configurable and extensible sketching framework which allows to dynamically inject sketching features into existing graphical diagram editors, based on Eclipse GEF. This allows to combine the flexibility of informal tools with the power of formal tools without any effort. No additional code is required to augment a GEF editor with sophisticated sketching features. Scribble recognizes drawn elements as well as handwritten text and automatically generates the corresponding domain elements. A local training data library is created dynamically by incrementally learning shapes, drawn by the user. Training data can be shared with others using the WebScribble web application which has been created as part of this work.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The central challenge in face recognition lies in understanding the role different facial features play in our judgments of identity. Notable in this regard are the relative contributions of the internal (eyes, nose and mouth) and external (hair and jaw-line) features. Past studies that have investigated this issue have typically used high-resolution images or good-quality line drawings as facial stimuli. The results obtained are therefore most relevant for understanding the identification of faces at close range. However, given that real-world viewing conditions are rarely optimal, it is also important to know how image degradations, such as loss of resolution caused by large viewing distances, influence our ability to use internal and external features. Here, we report experiments designed to address this issue. Our data characterize how the relative contributions of internal and external features change as a function of image resolution. While we replicated results of previous studies that have shown internal features of familiar faces to be more useful for recognition than external features at high resolution, we found that the two feature sets reverse in importance as resolution decreases. These results suggest that the visual system uses a highly non-linear cue-fusion strategy in combining internal and external features along the dimension of image resolution and that the configural cues that relate the two feature sets play an important role in judgments of facial identity.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This paper argues that the Japanese business system cannot be adequately understood without extending the focus of analysis beyond the individual firm to the vertical keiretsu, or business group. The vertical group or keiretsu structure was first identified and studied in the auto and electronics industries, where it is most strongly marked, but it characterizes virtually all sectors, service industries as well as manufacturing. Large industrial vertical keiretsu are composed of subsidiaries engaged in three distinct types of activities (manufacturing, marketing, and quasirelated business). The coordination and control systems are built on the flows of products, financial resources, information and technology, and people across formal company boundaries, with the parent firm controlling the key flows. The paper examines the prevailing explanations first for the emergence and then for the persistence of the vertical group structure, and looks at the current pressures for change and adaptation in the system.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The statistical analysis of compositional data is commonly used in geological studies. As is well-known, compositions should be treated using logratios of parts, which are difficult to use correctly in standard statistical packages. In this paper we describe the new features of our freeware package, named CoDaPack, which implements most of the basic statistical methods suitable for compositional data. An example using real data is presented to illustrate the use of the package

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Resumen tomado de la publicaci??n

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Resumen tomado de la publicaci??n

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Knowing the number of words in a Word 2010 document or in a text selection within a file can be very useful. See how to use Word’s word count features. For best viewing Download the video.