59 resultados para Associative Classifiers
Resumo:
Two studies investigated participants' sensitivity to the amount and diversity of the evidence when reasoning inductively about categories. Both showed that participants are more sensitive to characteristics of the evidence for arguments with general rather than specific conclusions. Both showed an association between cognitive ability and sensitivity to these evidence characteristics, particularly when the conclusion category was general. These results suggest that a simple associative process may not be sufficient to capture some key phenomena of category-based induction. They also support the claim that the need to generate a superordinate category is a complicating factor in category-based reasoning and that adults' tendency to generate such categories while reasoning has been overestimated.
Resumo:
The identification and classification of network traffic and protocols is a vital step in many quality of service and security systems. Traffic classification strategies must evolve, alongside the protocols utilising the Internet, to overcome the use of ephemeral or masquerading port numbers and transport layer encryption. This research expands the concept of using machine learning on the initial statistics of flow of packets to determine its underlying protocol. Recognising the need for efficient training/retraining of a classifier and the requirement for fast classification, the authors investigate a new application of k-means clustering referred to as 'two-way' classification. The 'two-way' classification uniquely analyses a bidirectional flow as two unidirectional flows and is shown, through experiments on real network traffic, to improve classification accuracy by as much as 18% when measured against similar proposals. It achieves this accuracy while generating fewer clusters, that is, fewer comparisons are needed to classify a flow. A 'two-way' classification offers a new way to improve accuracy and efficiency of machine learning statistical classifiers while still maintaining the fast training times associated with the k-means.
Resumo:
Since the 'completion' of Histoire(s) du cinema (1988-1998), Jean-Luc Godard's work has become increasingly mosaic-like in its forms and configurations, and markedly elegiac in its ruminations on history, cinema, art, and thought. While his associative aesthetic and citational method –including his choice of ‘actors’, and the fragmentariness of his ‘soundtracks’ – can combine to create a distinctive cinematic event, the films themselves refuse to cohere around a unifying concern, or yield to a thematic schema. Not surprisingly, Film Socialisme does not offer us the illusion of narrative or structural integrity anymore than it contributes to the quotidian rhetoric of political and moral argument. It is, however, a political film in the sense that it alters something more fundamental than opinions and points of view. It transforms a way of seeing and understanding reality and history, fiction and documentary, images, and images of images. If anything, it belongs to that dissident or ‘dissensual’ category of artwork capable of ‘emancipating the spectator’ by disturbing what Jacques Rancière terms ‘the distribution of the sensible’ in that it generates gaps, openings, and spaces, poses questions, invites associations without positing a fixed position, imposing an interpretation, or allowing itself to invest in the illusion of expressive objectivity and the stability of meaning. The myriad citations and fragments that comprise the film are never intended to culminate into anything cohesive, never mind conclusive. In one sense, they have no source and no context beyond their moment in the film itself, and what we make of that moment. This article studies the degree to which Godard allows these images and sounds to combine and collide, associate and dissolve in this film, arguing that Film Socialisme is both an important intervention in the history of contemporary cinema, and necessary point of reference in any serious discussion of the relations between that cinema and political reality.
Resumo:
Regarding the Real: Cinema, Documentary, and the Visual Arts develops an interdisciplinary approach to documentary film, focusing on its cultural and formal relations to other visual arts, such as animation, assemblage, photography, painting, sculpture, and architecture. The book considers the work of figures whose preferred film language is associative and fragmentary, and for whom the documentary is an endlessly open form, an unstable expressive phenomenon that cannot but interrogate the validity of its own narratives and representational modes. Combining close analysis with cultural history, Regarding the Real calls for a re-assessment of the influence of the modern arts in subverting the structures of realism typically associated with documentary filmmaking.
Resumo:
Bank conflicts can severely reduce the bandwidth of an interleaved multibank memory and conflict misses increase the miss rate of a cache or a predictor. Both occurrences are manifestations of the same problem: Objects which should be mapped to different indices are accidentally mapped to the same index. Suitable chosen hash functions can avoid conflicts in each of these situations by mapping the most frequently occurring patterns conflict-free. A particularly interesting class of hash functions are the XOR-based hash functions, which compute each set index bit as the exclusive-or of a subset of the address bits. When implementing an XOR-based hash function, it is extremely important to understand what patterns are mapped conflict-free and how a hash function can be constructed to map the most frequently occurring patterns without conflicts. Hereto, this paper presents two ways to reason about hash functions: by their null space and by their column space. The null space helps to quickly determine whether a pattern is mapped conflict-free. The column space is more useful for other purposes, e. g., to reduce the fan-in of the XOR-gates without introducing conflicts or to evaluate interbank dispersion in skewed-associative caches. Examples illustrate how these ideas can be applied to construct conflict-free hash functions.
Resumo:
Caches hide the growing latency of accesses to the main memory from the processor by storing the most recently used data on-chip. To limit the search time through the caches, they are organized in a direct mapped or set-associative way. Such an organization introduces many conflict misses that hamper performance. This paper studies randomizing set index functions, a technique to place the data in the cache in such a way that conflict misses are avoided. The performance of such a randomized cache strongly depends on the randomization function. This paper discusses a methodology to generate randomization functions that perform well over a broad range of benchmarks. The methodology uses profiling information to predict the conflict miss rate of randomization functions. Then, using this information, a search algorithm finds the best randomization function. Due to implementation issues, it is preferable to use a randomization function that is extremely simple and can be evaluated in little time. For these reasons, we use randomization functions where each randomized address bit is computed as the XOR of a subset of the original address bits. These functions are chosen such that they operate on as few address bits as possible and have few inputs to each XOR. This paper shows that to index a 2(m)-set cache, it suffices to randomize m+2 or m+3 address bits and to limit the number of inputs to each XOR to 2 bits to obtain the full potential of randomization. Furthermore, it is shown that the randomization function that we generate for one set of benchmarks also works well for an entirely different set of benchmarks. Using the described methodology, it is possible to reduce the implementation cost of randomization functions with only an insignificant loss in conflict reduction.
Resumo:
Embedded processors are used in numerous devices executing dedicated applications. This setting makes it worthwhile to optimize the processor to the application it executes, in order to increase its power-efficiency. This paper proposes to enhance direct mapped data caches with automatically tuned randomized set index functions to achieve that goal. We show how randomization functions can be automatically generated and compare them to traditional set-associative caches in terms of performance and energy consumption. A 16 kB randomized direct mapped cache consumes 22% less energy than a 2-way set-associative cache, while it is less than 3% slower. When the randomization function is made configurable (i.e., it can be adapted to the program), the additional reduction of conflicts outweighs the added complexity of the hardware, provided there is a sufficient amount of conflict misses.
Resumo:
The concentration of organic acids in anaerobic digesters is one of the most critical parameters for monitoring and advanced control of anaerobic digestion processes. Thus, a reliable online-measurement system is absolutely necessary. A novel approach to obtaining these measurements indirectly and online using UV/vis spectroscopic probes, in conjunction with powerful pattern recognition methods, is presented in this paper. An UV/vis spectroscopic probe from S::CAN is used in combination with a custom-built dilution system to monitor the absorption of fully fermented sludge at a spectrum from 200 to 750 nm. Advanced pattern recognition methods are then used to map the non-linear relationship between measured absorption spectra to laboratory measurements of organic acid concentrations. Linear discriminant analysis, generalized discriminant analysis (GerDA), support vector machines (SVM), relevance vector machines, random forest and neural networks are investigated for this purpose and their performance compared. To validate the approach, online measurements have been taken at a full-scale 1.3-MW industrial biogas plant. Results show that whereas some of the methods considered do not yield satisfactory results, accurate prediction of organic acid concentration ranges can be obtained with both GerDA and SVM-based classifiers, with classification rates in excess of 87% achieved on test data.
Resumo:
This study highlights the potential associated with utilising multi-component polymeric gels to formulate materials that possess unique rheological and mechanical properties. The synergistic effect* and interaction between hydroxyethylcellulose (HEC) and sodium carboxymethylcellulose (NaCMC), polymers which are commonly employed as drug delivery platforms for implantable medical devices (1), have been determined using dynamic, continuous shear and texture profile analysis. * The difference between the actual response of a binary mixture and the sum of the two components comprising the mixture Increases in polymer concentration resulted in an increase in G', G? and ?' whereas tan d decreased. Similarly, significant increases were also apparent in continuous shear and texture analysis. All binary mixtures showed positive synergy values which may suggest associative interaction between the two components.
Resumo:
This study explored the pattern of memory functioning in 58 patients with chronic schizophrenia and compared their performance with 53 normal controls. Multiple domains of memory were assessed, including verbal and nonverbal memory span, verbal and non-verbal paired associate learning, verbal and visual long-term memory, spatial and non-spatial conditional associative learning, recognition memory and memory for temporal order. Consistent with previous studies, substantial deficits in long-term memory were observed, with relative preservation of memory span. Memory for temporal order and recognition memory was intact, although significant deficits were observed on the conditional associative learning tasks. There was no evidence of lateralized memory impairment. In these respects, the pattern of memory impairment in schizophrenia is more similar in nature to that found in patients with memory dysfunction following mesiotemporal lobe lesions, rather than that associated with focal frontal lobe damage. (C) 1999 Elsevier Science B.V. All rights reserved.
Resumo:
Background: Ineffective risk stratification can delay diagnosis of serious disease in patients with hematuria. We applied a systems biology approach to analyze clinical, demographic and biomarker measurements (n = 29) collected from 157 hematuric patients: 80 urothelial cancer (UC) and 77 controls with confounding pathologies.
Methods: On the basis of biomarkers, we conducted agglomerative hierarchical clustering to identify patient and biomarker clusters. We then explored the relationship between the patient clusters and clinical characteristics using Chi-square analyses. We determined classification errors and areas under the receiver operating curve of Random Forest Classifiers (RFC) for patient subpopulations using the biomarker clusters to reduce the dimensionality of the data.
Results: Agglomerative clustering identified five patient clusters and seven biomarker clusters. Final diagnoses categories were non-randomly distributed across the five patient clusters. In addition, two of the patient clusters were enriched with patients with ‘low cancer-risk’ characteristics. The biomarkers which contributed to the diagnostic classifiers for these two patient clusters were similar. In contrast, three of the patient clusters were significantly enriched with patients harboring ‘high cancer-risk” characteristics including proteinuria, aggressive pathological stage and grade, and malignant cytology. Patients in these three clusters included controls, that is, patients with other serious disease and patients with cancers other than UC. Biomarkers which contributed to the diagnostic classifiers for the largest ‘high cancer- risk’ cluster were different than those contributing to the classifiers for the ‘low cancer-risk’ clusters. Biomarkers which contributed to subpopulations that were split according to smoking status, gender and medication were different.
Conclusions: The systems biology approach applied in this study allowed the hematuric patients to cluster naturally on the basis of the heterogeneity within their biomarker data, into five distinct risk subpopulations. Our findings highlight an approach with the promise to unlock the potential of biomarkers. This will be especially valuable in the field of diagnostic bladder cancer where biomarkers are urgently required. Clinicians could interpret risk classification scores in the context of clinical parameters at the time of triage. This could reduce cystoscopies and enable priority diagnosis of aggressive diseases, leading to improved patient outcomes at reduced costs. © 2013 Emmert-Streib et al; licensee BioMed Central Ltd.
Resumo:
The temporal analysis of products (TAP) technique was successfully applied for the first time to investigate the reverse water-gas shift (RWGS) reaction over a 2% Pt/CeO2 catalyst. The adsorption/desorption rate constants for CO2 and H-2 were determined in separate TAP pulse-response experiments, and the number of H-containing exchangeable species was determined using D-2 multipulse TAP experiments. This number is similar to the amount of active sites observed in previous SSITKA experiments. The CO production in the RWGS reaction was studied in a TAP experiment using separate (sequential) and simultaneous pulsing Of CO2 and H-2. A small yield of CO was observed when CO2 was pulsed alone over the reduced catalyst, whereas a much higher CO yield was observed when CO2 and H-2 were pulsed consecutively. The maximum CO yield was observed when the CO2 pulse was followed by a H-2 pulse with only a short (1 s) delay. Based on these findings, we conclude that an associative reaction mechanism dominates the RWGS reaction under these experimental conditions. The rate constants for several elementary steps can be determined from the TAP data. In addition, using a difference in the time scale of the separate reaction steps identified in the TAP experiments, it is possible to distinguish a number of possible reaction pathways. (c) 2005 Elsevier Inc. All rights reserved.
Resumo:
The association between poor metabolic control and the microvascular complications of diabetes is now well established, but the relationship between long-term metabolic control and the accelerated atherosclerosis of diabetes is as yet poorly defined. Hyperglycemia is the standard benchmark by which metabolic control is assessed. One mechanism by which elevated glucose levels may mediate vascular injury is through early and advanced glycation reactions affecting a wide variety of target molecules. The "glycation hypothesis'' has developed over the past 30 years, evolving gradually into a "carbonyl stress hypothesis'' and taking into account not only the modification of proteins by glucose, but also the roles of oxidative stress, a wide range of reactive carbonyl-containing intermediates (derived not only from glucose but also from lipids), and a variety of extra- and intracellular target molecules. The final products of these reactions may now be termed "Either Advanced Glycation or Lipoxidation End-Products'' or "EAGLEs.'' The ubiquity of carbonyl stress within the body, the complexity of the reactions involved, the variety of potential carbonyl intermediates and target molecules and their differing half-lives, and the slow development of the complications of diabetes all pose major challenges in dissecting the significance of these processes. The extent of the reactions tends to correlate with overall metabolic control, creating pitfalls in the interpretation of associative data. Many animal and cell culture studies, while supporting the hypothesis, must be viewed with caution in terms of relevance to human diabetes. In this article, the development of the carbonyl stress hypothesis is reviewed, and implications for present and future treatments to prevent complications are discussed.
Resumo:
The operations and processes that the human brain employs to achieve fast visual categorization remain a matter of debate. A first issue concerns the timing and place of rapid visual categorization and to what extent it can be performed with an early feed-forward pass of information through the visual system. A second issue involves the categorization of stimuli that do not reach visual awareness. There is disagreement over the degree to which these stimuli activate the same early mechanisms as stimuli that are consciously perceived. We employed continuous flash suppression (CFS), EEG recordings, and machine learning techniques to study visual categorization of seen and unseen stimuli. Our classifiers were able to predict from the EEG recordings the category of stimuli on seen trials but not on unseen trials. Rapid categorization of conscious images could be detected around 100?ms on the occipital electrodes, consistent with a fast, feed-forward mechanism of target detection. For the invisible stimuli, however, CFS eliminated all traces of early processing. Our results support the idea of a fast mechanism of categorization and suggest that this early categorization process plays an important role in later, more subtle categorizations, and perceptual processes.
Resumo:
Background: There has been an explosion of interest in methods of exogenous brain stimulation that induce changes in the excitability of human cerebral cortex. The expectation is that these methods may promote recovery of function following brain injury. To assess their effects on motor output, it is typical to assess the state of corticospinal projections from primary motor cortex to muscles of the hand, via electromyographic responses to transcranial magnetic stimulation. If a range of stimulation intensities is employed, the recruitment curves (RCs) obtained can, at least for intrinsic hand muscles, be fitted by a sigmoid function.
Objective/hypothesis: To establish whether sigmoid fits provide a reliable basis upon which to characterize the input–output properties of the corticospinal pathway for muscles proximal to the hand, and to assess as an alternative the area under the (recruitment) curve (AURC).
Methods: A comparison of the reliability of these measures, using RCs obtained for muscles that are frequently the targets of rehabilitation.
Results: The AURC is an extremely reliable measure of the state of corticospinal projections to hand and forearm muscles, which has both face and concurrent validity. Construct validity is demonstrated by detection of widely distributed (across muscles) changes in corticospinal excitability induced by paired associative stimulation (PAS).
Conclusion(s): The parameters derived from sigmoid fits are unlikely to provide an adequate means to assess the effectiveness of therapeutic regimes. The AURC can be employed to characterize corticospinal projections to a range of muscles, and gauge the efficacy of longitudinal interventions in clinical rehabilitation.