862 resultados para 280401 Analysis of Algorithms and Complexity
Resumo:
The initial aim of the CareMan project was to develop a joint degree programme that combined and utilised the strengths of the five collaborating universities that were already involved in delivering social and health care management education. Because the project was to be implemented in collaboration between education- al institutions, the collaboration had to be based on a detailed understanding of the national and institutional specifics of each of the individual academic enti- ties. During this process it was recognised that, due to a number of regulation issues, achieving the original aim would not be possible; ultimately, following a series of analytical works, which are presented below, it was decided that a set of three master’s level modules should be developed. One of the reasons was that the Finnish law on master’s degrees at universities of applied sciences (UAS) stated that the requirement for entry to a UAS master’s programme was a bachelor degree from a UAS or equivalent, plus a minimum of three years of work experience in an appropriate field. The three years’ work experience is also required from international students. In practice this meant that the participating Finnish UASs, Lahti and HAMK, could not award a diploma for foreign students without this work experience. The other European universities do not have the work experience requirement, although some take it as a bonus for admission (FHS UK). There were also other differences in law (e.g., requirements for mini - mum standards in Social Work education at FHS UK) that could not have been overcome during the period of project realisation. Consequently, the outcome was the development of only three common educational modules, each for 10 ECTS, which were developed, delivered and assessed during the lifetime of the project. The intention was that these would be integrated into the current masters’ level provision in each of the universities
Resumo:
ABSTRACT: This text reports an analysis of teaching activity with basis in the category of human activity. Refers to a theoretical research that focuses the essential nexus generators of objectification of educational practice both in the material aspect as the ideal. It is assumed that this practice, in dialectical unity between those aspects, constitutes in the material relations of social life, is structured and institutionalized as universal model of education (ideally) and becomes the main determinant of teacher performance. Methodologically, this study analyzes the educational practice in the totality of its relations.Therefore, the actions and operations of teaching, as well as, the reason and purposes that moved and guide this activity are determined by the objective conditions current of the school and the effects generated by the position of the teacher in these relations as an active subject. The guiding worry of this research focuses on the characterization of teaching activity in educational performance: in its accomplishment. When to investigate the evidence of these connections is necessary to reflect on the correlative links between the structural elements of the teaching activity and the natural practice of teachers under the determinations derived from the first. This reflection, arise the need to investigate the process of appropriation of the activity by the teacher. Thereby, concludes that the teacher not only appropriates the specific knowledge of the disciplines that teaches and the means of education which is available in their educational practice, but also the elements that structure their activity. This evinced the importance of investigations that worry with: the significance of the purpose of education for the teacher; the necessities that move in your performance; the meaning of your activity; among others. In addition to the responses of teachers is necessary to analyze what the content and the origin of appropriation. KEYWORDS: Teacher; Teaching activity; Educational practice.
Resumo:
We introduce K-tree in an information retrieval context. It is an efficient approximation of the k-means clustering algorithm. Unlike k-means it forms a hierarchy of clusters. It has been extended to address issues with sparse representations. We compare performance and quality to CLUTO using document collections. The K-tree has a low time complexity that is suitable for large document collections. This tree structure allows for efficient disk based implementations where space requirements exceed that of main memory.
Resumo:
This paper describes the approach taken to the XML Mining track at INEX 2008 by a group at the Queensland University of Technology. We introduce the K-tree clustering algorithm in an Information Retrieval context by adapting it for document clustering. Many large scale problems exist in document clustering. K-tree scales well with large inputs due to its low complexity. It offers promising results both in terms of efficiency and quality. Document classification was completed using Support Vector Machines.
Resumo:
Random Indexing K-tree is the combination of two algorithms suited for large scale document clustering.
Resumo:
For most of the work done in developing association rule mining, the primary focus has been on the efficiency of the approach and to a lesser extent the quality of the derived rules has been emphasized. Often for a dataset, a huge number of rules can be derived, but many of them can be redundant to other rules and thus are useless in practice. The extremely large number of rules makes it difficult for the end users to comprehend and therefore effectively use the discovered rules and thus significantly reduces the effectiveness of rule mining algorithms. If the extracted knowledge can’t be effectively used in solving real world problems, the effort of extracting the knowledge is worth little. This is a serious problem but not yet solved satisfactorily. In this paper, we propose a concise representation called Reliable Approximate basis for representing non-redundant approximate association rules. We prove that the redundancy elimination based on the proposed basis does not reduce the belief to the extracted rules. We also prove that all approximate association rules can be deduced from the Reliable Approximate basis. Therefore the basis is a lossless representation of approximate association rules.
Resumo:
Streaming SIMD Extensions (SSE) is a unique feature embedded in the Pentium III and IV classes of microprocessors. By fully exploiting SSE, parallel algorithms can be implemented on a standard personal computer and a theoretical speedup of four can be achieved. In this paper, we demonstrate the implementation of a parallel LU matrix decomposition algorithm for solving linear systems with SSE and discuss advantages and disadvantages of this approach based on our experimental study.
Resumo:
This paper introduces PartSS, a new partition-based fil- tering for tasks performing string comparisons under edit distance constraints. PartSS offers improvements over the state-of-the-art method NGPP with the implementation of a new partitioning scheme and also improves filtering abil- ities by exploiting theoretical results on shifting and scaling ranges, thus accelerating the rate of calculating edit distance between strings. PartSS filtering has been implemented within two major tasks of data integration: similarity join and approximate membership extraction under edit distance constraints. The evaluation on an extensive range of real-world datasets demonstrates major gain in efficiency over NGPP and QGrams approaches.
Resumo:
Sfinks is a shift register based stream cipher designed for hardware implementation and submitted to the eSTREAM project. In this paper, we analyse the initialisation process of Sfinks. We demonstrate a slid property of the loaded state of the Sfinks cipher, where multiple key-IV pairs may produce phase shifted keystream sequences. The state update functions of both the initialisation process and keystream generation and also the pattern of the padding affect generation of the slid pairs.
Resumo:
Topic modeling has been widely utilized in the fields of information retrieval, text mining, text classification etc. Most existing statistical topic modeling methods such as LDA and pLSA generate a term based representation to represent a topic by selecting single words from multinomial word distribution over this topic. There are two main shortcomings: firstly, popular or common words occur very often across different topics that bring ambiguity to understand topics; secondly, single words lack coherent semantic meaning to accurately represent topics. In order to overcome these problems, in this paper, we propose a two-stage model that combines text mining and pattern mining with statistical modeling to generate more discriminative and semantic rich topic representations. Experiments show that the optimized topic representations generated by the proposed methods outperform the typical statistical topic modeling method LDA in terms of accuracy and certainty.
Resumo:
The count-min sketch is a useful data structure for recording and estimating the frequency of string occurrences, such as passwords, in sub-linear space with high accuracy. However, it cannot be used to draw conclusions on groups of strings that are similar, for example close in Hamming distance. This paper introduces a variant of the count-min sketch which allows for estimating counts within a specified Hamming distance of the queried string. This variant can be used to prevent users from choosing popular passwords, like the original sketch, but it also allows for a more efficient method of analysing password statistics.
Resumo:
The 3-Hitting Set problem involves a family of subsets F of size at most three over an universe U. The goal is to find a subset of U of the smallest possible size that intersects every set in F. The version of the problem with parity constraints asks for a subset S of size at most k that, in addition to being a hitting set, also satisfies certain parity constraints on the sizes of the intersections of S with each set in the family F. In particular, an odd (even) set is a hitting set that hits every set at either one or three (two) elements, and a perfect code is a hitting set that intersects every set at exactly one element. These questions are of fundamental interest in many contexts for general set systems. Just as for Hitting Set, we find these questions to be interesting for the case of families consisting of sets of size at most three. In this work, we initiate an algorithmic study of these problems in this special case, focusing on a parameterized analysis. We show, for each problem, efficient fixed-parameter tractable algorithms using search trees that are tailor-made to the constraints in question, and also polynomial kernels using sunflower-like arguments in a manner that accounts for equivalence under the additional parity constraints.
Resumo:
"The Structure and Interpretation of Computer Programs" is the entry-level subject in Computer Science at the Massachusetts Institute of Technology. It is required of all students at MIT who major in Electrical Engineering or in Computer Science, as one fourth of the "common core curriculum," which also includes two subjects on circuits and linear systems and a subject on the design of digital systems. We have been involved in the development of this subject since 1978, and we have taught this material in its present form since the fall of 1980 to approximately 600 students each year. Most of these students have had little or no prior formal training in computation, although most have played with computers a bit and a few have had extensive programming or hardware design experience. Our design of this introductory Computer Science subject reflects two major concerns. First we want to establish the idea that a computer language is not just a way of getting a computer to perform operations, but rather that it is a novel formal medium for expressing ideas about methodology. Thus, programs must be written for people to read, and only incidentally for machines to execute. Secondly, we believe that the essential material to be addressed by a subject at this level, is not the syntax of particular programming language constructs, nor clever algorithms for computing particular functions of efficiently, not even the mathematical analysis of algorithms and the foundations of computing, but rather the techniques used to control the intellectual complexity of large software systems.