45 resultados para MASS CLASSIFICATION SYSTEMS
em CentAUR: Central Archive University of Reading - UK
Resumo:
Automatic generation of classification rules has been an increasingly popular technique in commercial applications such as Big Data analytics, rule based expert systems and decision making systems. However, a principal problem that arises with most methods for generation of classification rules is the overfit-ting of training data. When Big Data is dealt with, this may result in the generation of a large number of complex rules. This may not only increase computational cost but also lower the accuracy in predicting further unseen instances. This has led to the necessity of developing pruning methods for the simplification of rules. In addition, classification rules are used further to make predictions after the completion of their generation. As efficiency is concerned, it is expected to find the first rule that fires as soon as possible by searching through a rule set. Thus a suit-able structure is required to represent the rule set effectively. In this chapter, the authors introduce a unified framework for construction of rule based classification systems consisting of three operations on Big Data: rule generation, rule simplification and rule representation. The authors also review some existing methods and techniques used for each of the three operations and highlight their limitations. They introduce some novel methods and techniques developed by them recently. These methods and techniques are also discussed in comparison to existing ones with respect to efficient processing of Big Data.
Resumo:
Identification of Fusarium species has always been difficult due to confusing phenotypic classification systems. We have developed a fluorescent-based polymerase chain reaction assay that allows for rapid and reliable identification of five toxigenic and pathogenic Fusarium species. The species includes Fusarium avenaceum, F. culmorum, F. equiseti, F. oxysporum and F. sambucinum. The method is based on the PCR amplification of species-specific DNA fragments using fluorescent oligonucleotide primers, which were designed based on sequence divergence within the internal transcribed spacer region of nuclear ribosomal DNA. Besides providing an accurate, reliable, and quick diagnosis of these Fusaria, another advantage with this method is that it reduces the potential for exposure to carcinogenic chemicals as it substitutes the use of fluorescent dyes in place of ethidium, bromide. Apart from its multidisciplinary importance and usefulness, it also obviates the need for gel electrophoresis. (C) 2002 Published by Elsevier Science B.V. on behalf of the Federation of European Microbiological Societies.
Resumo:
This paper presents an approach for automatic classification of pulsed Terahertz (THz), or T-ray, signals highlighting their potential in biomedical, pharmaceutical and security applications. T-ray classification systems supply a wealth of information about test samples and make possible the discrimination of heterogeneous layers within an object. In this paper, a novel technique involving the use of Auto Regressive (AR) and Auto Regressive Moving Average (ARMA) models on the wavelet transforms of measured T-ray pulse data is presented. Two example applications are examined - the classi. cation of normal human bone (NHB) osteoblasts against human osteosarcoma (HOS) cells and the identification of six different powder samples. A variety of model types and orders are used to generate descriptive features for subsequent classification. Wavelet-based de-noising with soft threshold shrinkage is applied to the measured T-ray signals prior to modeling. For classi. cation, a simple Mahalanobis distance classi. er is used. After feature extraction, classi. cation accuracy for cancerous and normal cell types is 93%, whereas for powders, it is 98%.
Resumo:
The North Atlantic Marine Boundary Layer Experiment (NAMBLEX), involving over 50 scientists from 12 institutions, took place at Mace Head, Ireland (53.32° N, 9.90° W), between 23 July and 4 September 2002. A wide range of state-of-the-art instrumentation enabled detailed measurements of the boundary layer structure and atmospheric composition in the gas and aerosol phase to be made, providing one of the most comprehensive in situ studies of the marine boundary layer to date. This overview paper describes the aims of the NAMBLEX project in the context of previous field campaigns in the Marine Boundary Layer (MBL), the overall layout of the site, a summary of the instrumentation deployed, the temporal coverage of the measurement data, and the numerical models used to interpret the field data. Measurements of some trace species were made for the first time during the campaign, which was characterised by predominantly clean air of marine origin, but more polluted air with higher levels of NOx originating from continental regions was also experienced. This paper provides a summary of the meteorological measurements and Planetary Boundary Layer (PBL) structure measurements, presents time series of some of the longer-lived trace species (O3, CO, H2, DMS, CH4, NMHC, NOx, NOy, PAN) and summarises measurements of other species that are described in more detail in other papers within this special issue, namely oxygenated VOCs, HCHO, peroxides, organo-halogenated species, a range of shorter lived halogen species (I2, OIO, IO, BrO), NO3 radicals, photolysis frequencies, the free radicals OH, HO2 and (HO2+Σ RO2), as well as a summary of the aerosol measurements. NAMBLEX was supported by measurements made in the vicinity of Mace Head using the NERC Dornier-228 aircraft. Using ECMWF wind-fields, calculations were made of the air-mass trajectories arriving at Mace Head during NAMBLEX, and were analysed together with both meteorological and trace-gas measurements. In this paper a chemical climatology for the duration of the campaign is presented to interpret the distribution of air-mass origins and emission sources, and to provide a convenient framework of air-mass classification that is used by other papers in this issue for the interpretation of observed variability in levels of trace gases and aerosols.
Resumo:
The SPE taxonomy of evolving software systems, first proposed by Lehman in 1980, is re-examined in this work. The primary concepts of software evolution are related to generic theories of evolution, particularly Dawkins' concept of a replicator, to the hermeneutic tradition in philosophy and to Kuhn's concept of paradigm. These concepts provide the foundations that are needed for understanding the phenomenon of software evolution and for refining the definitions of the SPE categories. In particular, this work argues that a software system should be defined as of type P if its controlling stakeholders have made a strategic decision that the system must comply with a single paradigm in its representation of domain knowledge. The proposed refinement of SPE is expected to provide a more productive basis for developing testable hypotheses and models about possible differences in the evolution of E- and P-type systems than is provided by the original scheme. Copyright (C) 2005 John Wiley & Sons, Ltd.
Resumo:
We extend recent work that included the effect of pressure forces to derive the precession rate of eccentric accretion discs in cataclysmic variables to the case of double degenerate systems. We find that the logical scaling of the pressure force in such systems results in predictions of unrealistically high primary masses. Using the prototype AM CVn as a calibrator for the magnitude of the effect, we find that there is no scaling that applies consistently to all the systems in the class. We discuss the reasons for the lack of a superhump period to mass ratio relationship analogous to that known for SU UMa systems and suggest that this is because these secondaries do not have a single valued mass-radius relationship. We highlight the unreliability of mass-ratios derived by applying the SU UMa expression to the AM CVn binaries.
Resumo:
Within this paper modern techniques such as satellite image analysis and tools provided by geographic information systems (GIS.) are exploited in order to extend and improve existing techniques for mapping the spatial distribution of sediment transport processes. The processes of interest comprise mass movements such as solifluction, slope wash, dirty avalanches and rock- and boulder falls. They differ considerably in nature and therefore different approaches for the derivation of their spatial extent are required. A major challenge is addressing the differences between the comparably coarse resolution of the available satellite data (Landsat TM/ETM+, 30 in x 30 m) and the actual scale of sediment transport in this environment. A three-stepped approach has been developed which is based on the concept of Geomorphic Process Units (GPUs): parameterization, process area delineation and combination. Parameters include land cover from satellite data and digital elevation model derivatives. Process areas are identified using a hierarchical classification scheme utilizing thresholds and definition of topology. The approach has been developed for the Karkevagge in Sweden and could be successfully transferred to the Rabotsbekken catchment at Okstindan, Norway using similar input data. Copyright (C) 2008 John Wiley & Sons, Ltd.
Resumo:
This article begins by identifying a close relationship between the image of children generated by several sociologists working within the new sociology of childhood perspective and the claims and ambitions of the proponents of children's autonomy rights. The image of the child as a competent, self-controlled human agent are then subjected to observation from the perspective of Niklas Luhmann's social systems theory. The new sociology of childhood's constructivist approach is compared and contrasted with Niklas Luhmann's theory of 'operational constructivism'. The article applies tenets of Luhmann's theory, to the emergence of the new childhood sociologist's image of the child as a competent, self-controlled social agent, to the epistemological status of this image and, in particular, to claims that it derives from scientific endeavour. The article proceeds to identify two theoretical developments within sociology - sociology of identity and social agency - which have brought about fundamental changes in what may be considered 'sociological' and so 'scientific' and paved the way for sociological communications about what children,really are'. In conclusion, it argues that the merging of sociology with polemics, ideology, opinion and personal beliefs and, at the level of social systems, between science and politics represents in Luhmann's terms 'dedifferentiation'- a tendency he claims may have serious adverse consequences for modern society. This warning is applied to the scientific status of sociology - its claim to be able to produce 'facts' for society, upon which social systems, such as politics and law, may rely. Like the mass media, sociology may now be capable of producing only information, and not facts, about children.
Resumo:
We have combined several key sample preparation steps for the use of a liquid matrix system to provide high analytical sensitivity in automated ultraviolet -- matrix-assisted laser desorption/ionisation -- mass spectrometry (UV-MALDI-MS). This new sample preparation protocol employs a matrix-mixture which is based on the glycerol matrix-mixture described by Sze et al. The low-femtomole sensitivity that is achievable with this new preparation protocol enables proteomic analysis of protein digests comparable to solid-state matrix systems. For automated data acquisition and analysis, the MALDI performance of this liquid matrix surpasses the conventional solid-state MALDI matrices. Besides the inherent general advantages of liquid samples for automated sample preparation and data acquisition the use of the presented liquid matrix significantly reduces the extent of unspecific ion signals in peptide mass fingerprints compared to typically used solid matrices, such as 2,5-dihydroxybenzoic acid (DHB) or alpha-cyano-hydroxycinnamic acid (CHCA). In particular, matrix and low-mass ion signals and ion signals resulting from cation adduct formation are dramatically reduced. Consequently, the confidence level of protein identification by peptide mass mapping of in-solution and in-gel digests is generally higher.
Resumo:
A homologous series of macrocyclic oligoamides has been prepared in high yield by reaction of isophthaloyl chloride with m-phenylenediamine under pseudo-high-dilution conditions. The products were characterized by infrared and H-1 NMR spectroscopies, matrix assisted laser desorption-ionization time-of-flight mass spectrometry, and gel permeation chromatography (GPC). A series of linear oligomers was prepared for comparison. The macrocycles ranged in size from the cyclic trimer up to at least the cyclic nonamer (90 ring atoms). The same homologous series of macrocyclic oligomers was prepared in high yield by the cyclodepolymerization of poly(m-phenylene isophthalamide) (Nomex). Cyclodepolymerization was best achieved by treating a 1% w/v solution of the polymer in dimethyl sulfoxide containing calcium chloride or lithium chloride with 3-4 mol % of sodium hydride or the sodium salt of benzanilide at 150 degreesC for 70 h. Treatment of a concentrated solution of the macrocyclic oligomers (25% w/v) with 4 mol % of sodium hydride or the sodium salt of benzanilide in a solution of lithium chloride in dimethyl sulfoxide at 170 degreesC for 6 h resulted in efficient entropically driven ring-opening polymerizations to give poly(m-phenylene isophthalamide), characterized by infrared and H-1 NMR spectroscopies and by GPC. The molecular weights obtained were comparable with those of the commercial polymer.
Resumo:
We have combined several key sample preparation steps for the use of a liquid matrix system to provide high analytical sensitivity in automated ultraviolet - matrix-assisted laser desorption/ ionisation - mass spectrometry (UV-MALDI-MS). This new sample preparation protocol employs a matrix-mixture which is based on the glycerol matrix-mixture described by Sze et al. U. Am. Soc. Mass Spectrom. 1998, 9, 166-174). The low-ferntomole sensitivity that is achievable with this new preparation protocol enables proteomic analysis of protein digests comparable to solid-state matrix systems. For automated data acquisition and analysis, the MALDI performance of this liquid matrix surpasses the conventional solid-state MALDI matrices. Besides the inherent general advantages of liquid samples for automated sample preparation and data acquisition the use of the presented liquid matrix significantly reduces the extent of unspecific ion signals in peptide mass fingerprints compared to typically used solid matrices, such as 2,5-dihydrox-ybenzoic acid (DHB) or alpha-cyano-hydroxycinnamic acid (CHCA). In particular, matrix and lowmass ion signals and ion signals resulting from cation adduct formation are dramatically reduced. Consequently, the confidence level of protein identification by peptide mass mapping of in-solution and in-gel digests is generally higher.
Resumo:
The Maillard reaction causes changes to protein structure and occurs in foods mainly during thermal treatment. Melanoidins, the final products of the Maillard reaction, may enter the gastrointestinal tract, which is populated by different species of bacteria. In this study, melanoidins were prepared from gluten and glucose. Their effect on the growth of faecal bacteria was determined in culture with genotype and phenotype probes to identify the different species involved. Analysis of peptic and tryptic digests showed that low molecular mass products are formed from the degradation of melanoidins. Results showed a change in the growth of bacteria. This in vitro study demonstrated that melanoidins, prepared from gluten and glucose, affect the growth of the gut microflora.
Resumo:
We consider a fully complex-valued radial basis function (RBF) network for regression and classification applications. For regression problems, the locally regularised orthogonal least squares (LROLS) algorithm aided with the D-optimality experimental design, originally derived for constructing parsimonious real-valued RBF models, is extended to the fully complex-valued RBF (CVRBF) network. Like its real-valued counterpart, the proposed algorithm aims to achieve maximised model robustness and sparsity by combining two effective and complementary approaches. The LROLS algorithm alone is capable of producing a very parsimonious model with excellent generalisation performance while the D-optimality design criterion further enhances the model efficiency and robustness. By specifying an appropriate weighting for the D-optimality cost in the combined model selecting criterion, the entire model construction procedure becomes automatic. An example of identifying a complex-valued nonlinear channel is used to illustrate the regression application of the proposed fully CVRBF network. The proposed fully CVRBF network is also applied to four-class classification problems that are typically encountered in communication systems. A complex-valued orthogonal forward selection algorithm based on the multi-class Fisher ratio of class separability measure is derived for constructing sparse CVRBF classifiers that generalise well. The effectiveness of the proposed algorithm is demonstrated using the example of nonlinear beamforming for multiple-antenna aided communication systems that employ complex-valued quadrature phase shift keying modulation scheme. (C) 2007 Elsevier B.V. All rights reserved.
Resumo:
In rapid scan Fourier transform spectrometry, we show that the noise in the wavelet coefficients resulting from the filter bank decomposition of the complex insertion loss function is linearly related to the noise power in the sample interferogram by a noise amplification factor. By maximizing an objective function composed of the power of the wavelet coefficients divided by the noise amplification factor, optimal feature extraction in the wavelet domain is performed. The performance of a classifier based on the output of a filter bank is shown to be considerably better than that of an Euclidean distance classifier in the original spectral domain. An optimization procedure results in a further improvement of the wavelet classifier. The procedure is suitable for enhancing the contrast or classifying spectra acquired by either continuous wave or THz transient spectrometers as well as for increasing the dynamic range of THz imaging systems. (C) 2003 Optical Society of America.