190 resultados para Contig Creation Algorithm
Resumo:
Advances in hardware and software in the past decade allow to capture, record and process fast data streams at a large scale. The research area of data stream mining has emerged as a consequence from these advances in order to cope with the real time analysis of potentially large and changing data streams. Examples of data streams include Google searches, credit card transactions, telemetric data and data of continuous chemical production processes. In some cases the data can be processed in batches by traditional data mining approaches. However, in some applications it is required to analyse the data in real time as soon as it is being captured. Such cases are for example if the data stream is infinite, fast changing, or simply too large in size to be stored. One of the most important data mining techniques on data streams is classification. This involves training the classifier on the data stream in real time and adapting it to concept drifts. Most data stream classifiers are based on decision trees. However, it is well known in the data mining community that there is no single optimal algorithm. An algorithm may work well on one or several datasets but badly on others. This paper introduces eRules, a new rule based adaptive classifier for data streams, based on an evolving set of Rules. eRules induces a set of rules that is constantly evaluated and adapted to changes in the data stream by adding new and removing old rules. It is different from the more popular decision tree based classifiers as it tends to leave data instances rather unclassified than forcing a classification that could be wrong. The ongoing development of eRules aims to improve its accuracy further through dynamic parameter setting which will also address the problem of changing feature domain values.
Resumo:
This contribution introduces a new digital predistorter to compensate serious distortions caused by memory high power amplifiers (HPAs) which exhibit output saturation characteristics. The proposed design is based on direct learning using a data-driven B-spline Wiener system modeling approach. The nonlinear HPA with memory is first identified based on the B-spline neural network model using the Gauss-Newton algorithm, which incorporates the efficient De Boor algorithm with both B-spline curve and first derivative recursions. The estimated Wiener HPA model is then used to design the Hammerstein predistorter. In particular, the inverse of the amplitude distortion of the HPA's static nonlinearity can be calculated effectively using the Newton-Raphson formula based on the inverse of De Boor algorithm. A major advantage of this approach is that both the Wiener HPA identification and the Hammerstein predistorter inverse can be achieved very efficiently and accurately. Simulation results obtained are presented to demonstrate the effectiveness of this novel digital predistorter design.
Resumo:
Evolutionary meta-algorithms for pulse shaping of broadband femtosecond duration laser pulses are proposed. The genetic algorithm searching the evolutionary landscape for desired pulse shapes consists of a population of waveforms (genes), each made from two concatenated vectors, specifying phases and magnitudes, respectively, over a range of frequencies. Frequency domain operators such as mutation, two-point crossover average crossover, polynomial phase mutation, creep and three-point smoothing as well as a time-domain crossover are combined to produce fitter offsprings at each iteration step. The algorithm applies roulette wheel selection; elitists and linear fitness scaling to the gene population. A differential evolution (DE) operator that provides a source of directed mutation and new wavelet operators are proposed. Using properly tuned parameters for DE, the meta-algorithm is used to solve a waveform matching problem. Tuning allows either a greedy directed search near the best known solution or a robust search across the entire parameter space.
Resumo:
This paper proposes a framework to support Customer Relationship Management (CRM) implementation in nursing homes. The work extends research by Cheng et al. (2005) who conducted in-depth questionnaires to identify critical features (termed value-characteristics), which are areas identified as adding the most value if implemented. Although Cheng et al. did proposed an implementation framework, summary of, and inconsistent inclusion of value-characteristics, limits the practical use of this contribution during implementation. In this paper we adapt the original framework to correct perceived deficiencies. We link the value characteristics to operational, analytical, strategic and/or collaborative CRM solution types, to allow consideration in context of practical implementation solutions. The outcome of this paper shows that, practically, a 'one solution meets all characteristic' approach to CRM implementation within nursing homes is inappropriate. Our framework, however, supports implementers in identifying how value can be gained when implementing a specific CRM solution within nursing homes; which subsequently support project management and expectation management.
Resumo:
This paper explores the nature of private social and environmental reporting (SER). From interviews with UK institutional investors, we show that both investors and investees employ Goffmanesque, staged impression management as a means of creating and disseminating a dual myth of social and environmental accountability. The interviewees’ utterances unveil private meetings imbued with theatrical verbal and physical impression management. Most of the time, the investors’ shared awareness of reality belongs to a Goffmanesque frame whereby they accept no intentionality, misrepresentation or fabrication, believing instead that the ‘performers’ (investees) are not intending to deceive them. A shared perception that social and environmental considerations are subordinated to financial issues renders private SER an empty encounter characterised as a relationship-building exercise with seldom any impact on investment decision-making. Investors spoke of occasional instances of fabrication but these were insufficient to break the frame of dual myth creation. They only identified a handful of instances where intentional misrepresentation had been significant enough to alter their reality and behaviour. Only in the most extreme cases of fabrication and lying did the staged meeting break frame and become a genuine occasion of accountability, where investors demanded greater transparency, further meetings and at the extreme, divested shares. We conclude that the frontstage, ritualistic impression management in private SER is inconsistent with backstage activities within financial institutions where private financial reporting is prioritised. The investors appeared to be in a double bind whereby they devoted resources to private SER but were simultaneously aware that these efforts may be at best subordinated, at worst ignored, rendering private SER a predominantly cosmetic, theatrical and empty exercise.
Resumo:
This paper analyze and study a pervasive computing system in a mining environment to track people based on RFID (radio frequency identification) technology. In first instance, we explain the RFID fundamentals and the LANDMARC (location identification based on dynamic active RFID calibration) algorithm, then we present the proposed algorithm combining LANDMARC and trilateration technique to collect the coordinates of the people inside the mine, next we generalize a pervasive computing system that can be implemented in mining, and finally we show the results and conclusions.
The role of double-stranded break repair in the creation of phenotypic diversity at cereal VRN1 loci
Resumo:
Nonhomologous repair of double-stranded breaks, although fundamental to the maintenance of genomic integrity in all eukaryotes, has received little attention as to its evolutionary consequences in the generation and selection of phenotypic diversity. Here we document the role of illegitimate recombination in the creation of novel alleles in VRN1 orthologs selected to confer adaptation to annual cropping systems in barley and wheat.
Resumo:
For Northern Hemisphere extra-tropical cyclone activity, the dependency of a potential anthropogenic climate change signal on the identification method applied is analysed. This study investigates the impact of the used algorithm on the changing signal, not the robustness of the climate change signal itself. Using one single transient AOGCM simulation as standard input for eleven state-of-the-art identification methods, the patterns of model simulated present day climatologies are found to be close to those computed from re-analysis, independent of the method applied. Although differences in the total number of cyclones identified exist, the climate change signals (IPCC SRES A1B) in the model run considered are largely similar between methods for all cyclones. Taking into account all tracks, decreasing numbers are found in the Mediterranean, the Arctic in the Barents and Greenland Seas, the mid-latitude Pacific and North America. Changing patterns are even more similar, if only the most severe systems are considered: the methods reveal a coherent statistically significant increase in frequency over the eastern North Atlantic and North Pacific. We found that the differences between the methods considered are largely due to the different role of weaker systems in the specific methods.
Resumo:
Northern Hemisphere cyclone activity is assessed by applying an algorithm for the detection and tracking of synoptic scale cyclones to mean sea level pressure data. The method, originally developed for the Southern Hemisphere, is adapted for application in the Northern Hemisphere winter season. NCEP-Reanalysis data from 1958/59 to 1997/98 are used as input. The sensitivities of the results to particular parameters of the algorithm are discussed for both case studies and from a climatological point of view. Results show that the choice of settings is of major relevance especially for the tracking of smaller scale and fast moving systems. With an appropriate setting the algorithm is capable of automatically tracking different types of cyclones at the same time: Both fast moving and developing systems over the large ocean basins and smaller scale cyclones over the Mediterranean basin can be assessed. The climatology of cyclone variables, e.g., cyclone track density, cyclone counts, intensification rates, propagation speeds and areas of cyclogenesis and -lysis gives detailed information on typical cyclone life cycles for different regions. The lowering of the spatial and temporal resolution of the input data from full resolution T62/06h to T42/12h decreases the cyclone track density and cyclone counts. Reducing the temporal resolution alone contributes to a decline in the number of fast moving systems, which is relevant for the cyclone track density. Lowering spatial resolution alone mainly reduces the number of weak cyclones.
Resumo:
For an increasing number of applications, mesoscale modelling systems now aim to better represent urban areas. The complexity of processes resolved by urban parametrization schemes varies with the application. The concept of fitness-for-purpose is therefore critical for both the choice of parametrizations and the way in which the scheme should be evaluated. A systematic and objective model response analysis procedure (Multiobjective Shuffled Complex Evolution Metropolis (MOSCEM) algorithm) is used to assess the fitness of the single-layer urban canopy parametrization implemented in the Weather Research and Forecasting (WRF) model. The scheme is evaluated regarding its ability to simulate observed surface energy fluxes and the sensitivity to input parameters. Recent amendments are described, focussing on features which improve its applicability to numerical weather prediction, such as a reduced and physically more meaningful list of input parameters. The study shows a high sensitivity of the scheme to parameters characterizing roof properties in contrast to a low response to road-related ones. Problems in partitioning of energy between turbulent sensible and latent heat fluxes are also emphasized. Some initial guidelines to prioritize efforts to obtain urban land-cover class characteristics in WRF are provided. Copyright © 2010 Royal Meteorological Society and Crown Copyright.
Resumo:
In this paper a modified algorithm is suggested for developing polynomial neural network (PNN) models. Optimal partial description (PD) modeling is introduced at each layer of the PNN expansion, a task accomplished using the orthogonal least squares (OLS) method. Based on the initial PD models determined by the polynomial order and the number of PD inputs, OLS selects the most significant regressor terms reducing the output error variance. The method produces PNN models exhibiting a high level of accuracy and superior generalization capabilities. Additionally, parsimonious models are obtained comprising a considerably smaller number of parameters compared to the ones generated by means of the conventional PNN algorithm. Three benchmark examples are elaborated, including modeling of the gas furnace process as well as the iris and wine classification problems. Extensive simulation results and comparison with other methods in the literature, demonstrate the effectiveness of the suggested modeling approach.
Resumo:
A new sparse kernel density estimator is introduced. Our main contribution is to develop a recursive algorithm for the selection of significant kernels one at time using the minimum integrated square error (MISE) criterion for both kernel selection. The proposed approach is simple to implement and the associated computational cost is very low. Numerical examples are employed to demonstrate that the proposed approach is effective in constructing sparse kernel density estimators with competitive accuracy to existing kernel density estimators.
Resumo:
This paper explores the nature of private social and environmental reporting (SER). From interviews with UK institutional investors, we show that both investors and investees employ Goffmanesque, staged impression management as a means of creating and disseminating a dual myth of social and environmental accountability. The interviewees’ utterances unveil private meetings imbued with theatrical verbal and physical impression management. Most of the time, the investors’ shared awareness of reality belongs to a Goffmanesque frame whereby they accept no intentionality, misrepresentation or fabrication, believing instead that the ‘performers’ (investees) are not intending to deceive them. A shared perception that social and environmental considerations are subordinated to financial issues renders private SER an empty encounter characterised as a relationship-building exercise with seldom any impact on investment decision-making. Investors spoke of occasional instances of fabrication but these were insufficient to break the frame of dual myth creation. They only identified a handful of instances where intentional misrepresentation had been significant enough to alter their reality and behaviour. Only in the most extreme cases of fabrication and lying did the staged meeting break frame and become a genuine occasion of accountability, where investors demanded greater transparency, further meetings and at the extreme, divested shares. We conclude that the frontstage, ritualistic impression management in private SER is inconsistent with backstage activities within financial institutions where private financial reporting is prioritised. The investors appeared to be in a double bind whereby they devoted resources to private SER but were simultaneously aware that these efforts may be at best subordinated, at worst ignored, rendering private SER a predominantly cosmetic, theatrical and empty exercise.