899 resultados para Multiple kernel learning
Resumo:
To enhance the global search ability of Population Based Incremental Learning (PBIL) methods, It Is proposed that multiple probability vectors are to be Included on available PBIL algorithms. As a result, the strategy for updating those probability vectors and the negative learning and mutation operators are redefined as reported. Numerical examples are reported to demonstrate the pros and cons of the newly Implemented algorithm. ©2006 IEEE.
Resumo:
This study aimed to verify the effects of a metatextual intervention program, in the elaboration of stories written by students with learning difficulties. Four students were included in the sample of both genders, with ages ranging between eight years and four months and ten years and two months of age. The program was implemented at the participant schools, using an approach of multiple baseline within-subjects, with two conditions: baseline and intervention. Data analysis was based on the classification of stories produced by the students. Mann-Whitney testing was also applied, to analyze whether there have been significant changes in these productions. The results indicated that all students have improved performance in relation to the categories of produced stories, from elementary schemas (33%), for a more elaborate scheme (77%), with a better structuring of the elements that constitute a story. Statistical analysis also showed that the intervention has produced significant results for all variables analyzed. The data obtained have shown that the program was effective.
Resumo:
This paper describes a 3D virtual lab environment that was developed using OpenSim software integrated into Moodle. Virtuald software tool was used to provide pedagogical support to the lab by enabling to create online texts and delivering them to the students. The courses taught in this virtual lab are methodologically in conformity to theory of multiple intelligences. Some results are presented.
Resumo:
In the pattern recognition research field, Support Vector Machines (SVM) have been an effectiveness tool for classification purposes, being successively employed in many applications. The SVM input data is transformed into a high dimensional space using some kernel functions where linear separation is more likely. However, there are some computational drawbacks associated to SVM. One of them is the computational burden required to find out the more adequate parameters for the kernel mapping considering each non-linearly separable input data space, which reflects the performance of SVM. This paper introduces the Polynomial Powers of Sigmoid for SVM kernel mapping, and it shows their advantages over well-known kernel functions using real and synthetic datasets.
Resumo:
The dorsolateral column of the periaqueductal gray (dlPAG) integrates aversive emotional experiences and represents an important site responding to life threatening situations, such as hypoxia, cardiac pain and predator threats. Previous studies have shown that the dorsal PAG also supports fear learning; and we have currently explored how the dlPAG influences associative learning. We have first shown that N-methyl-D-aspartate (NMDA) 100 pmol injection in the dlPAG works as a valuable unconditioned stimulus (US) for the acquisition of olfactory fear conditioning (OFC) using amyl acetate odor as conditioned stimulus (CS). Next, we revisited the ascending projections of the dlPAG to the thalamus and hypothalamus to reveal potential paths that could mediate associative learning during OFC. Accordingly, the most important ascending target of the dlPAG is the hypothalamic defensive circuit, and we were able to show that pharmacological inactivation using beta-adrenoceptor blockade of the dorsal premammillary nucleus, the main exit way for the hypothalamic defensive circuit to thalamo-cortical circuits involved in fear learning, impaired the acquisition of the OFC promoted by NMDA stimulation of the dlPAG. Moreover, our tracing study revealed multiple parallel paths from the dlPAG to several thalamic targets linked to cortical-hippocampal-amygdalar circuits involved in fear learning. Overall, the results point to a major role of the dlPAG in the mediation of aversive associative learning via ascending projections to the medial hypothalamic defensive circuit, and perhaps, to other thalamic targets, as well. These results provide interesting perspectives to understand how life threatening events impact on fear learning, and should be useful to understand pathological fear memory encoding in anxiety disorders.
Resumo:
Item response theory (IRT) comprises a set of statistical models which are useful in many fields, especially when there is an interest in studying latent variables (or latent traits). Usually such latent traits are assumed to be random variables and a convenient distribution is assigned to them. A very common choice for such a distribution has been the standard normal. Recently, Azevedo et al. [Bayesian inference for a skew-normal IRT model under the centred parameterization, Comput. Stat. Data Anal. 55 (2011), pp. 353-365] proposed a skew-normal distribution under the centred parameterization (SNCP) as had been studied in [R. B. Arellano-Valle and A. Azzalini, The centred parametrization for the multivariate skew-normal distribution, J. Multivariate Anal. 99(7) (2008), pp. 1362-1382], to model the latent trait distribution. This approach allows one to represent any asymmetric behaviour concerning the latent trait distribution. Also, they developed a Metropolis-Hastings within the Gibbs sampling (MHWGS) algorithm based on the density of the SNCP. They showed that the algorithm recovers all parameters properly. Their results indicated that, in the presence of asymmetry, the proposed model and the estimation algorithm perform better than the usual model and estimation methods. Our main goal in this paper is to propose another type of MHWGS algorithm based on a stochastic representation (hierarchical structure) of the SNCP studied in [N. Henze, A probabilistic representation of the skew-normal distribution, Scand. J. Statist. 13 (1986), pp. 271-275]. Our algorithm has only one Metropolis-Hastings step, in opposition to the algorithm developed by Azevedo et al., which has two such steps. This not only makes the implementation easier but also reduces the number of proposal densities to be used, which can be a problem in the implementation of MHWGS algorithms, as can be seen in [R.J. Patz and B.W. Junker, A straightforward approach to Markov Chain Monte Carlo methods for item response models, J. Educ. Behav. Stat. 24(2) (1999), pp. 146-178; R. J. Patz and B. W. Junker, The applications and extensions of MCMC in IRT: Multiple item types, missing data, and rated responses, J. Educ. Behav. Stat. 24(4) (1999), pp. 342-366; A. Gelman, G.O. Roberts, and W.R. Gilks, Efficient Metropolis jumping rules, Bayesian Stat. 5 (1996), pp. 599-607]. Moreover, we consider a modified beta prior (which generalizes the one considered in [3]) and a Jeffreys prior for the asymmetry parameter. Furthermore, we study the sensitivity of such priors as well as the use of different kernel densities for this parameter. Finally, we assess the impact of the number of examinees, number of items and the asymmetry level on the parameter recovery. Results of the simulation study indicated that our approach performed equally as well as that in [3], in terms of parameter recovery, mainly using the Jeffreys prior. Also, they indicated that the asymmetry level has the highest impact on parameter recovery, even though it is relatively small. A real data analysis is considered jointly with the development of model fitting assessment tools. The results are compared with the ones obtained by Azevedo et al. The results indicate that using the hierarchical approach allows us to implement MCMC algorithms more easily, it facilitates diagnosis of the convergence and also it can be very useful to fit more complex skew IRT models.
Resumo:
Support Vector Machines (SVMs) have achieved very good performance on different learning problems. However, the success of SVMs depends on the adequate choice of the values of a number of parameters (e.g., the kernel and regularization parameters). In the current work, we propose the combination of meta-learning and search algorithms to deal with the problem of SVM parameter selection. In this combination, given a new problem to be solved, meta-learning is employed to recommend SVM parameter values based on parameter configurations that have been successfully adopted in previous similar problems. The parameter values returned by meta-learning are then used as initial search points by a search technique, which will further explore the parameter space. In this proposal, we envisioned that the initial solutions provided by meta-learning are located in good regions of the search space (i.e. they are closer to optimum solutions). Hence, the search algorithm would need to evaluate a lower number of candidate solutions when looking for an adequate solution. In this work, we investigate the combination of meta-learning with two search algorithms: Particle Swarm Optimization and Tabu Search. The implemented hybrid algorithms were used to select the values of two SVM parameters in the regression domain. These combinations were compared with the use of the search algorithms without meta-learning. The experimental results on a set of 40 regression problems showed that, on average, the proposed hybrid methods obtained lower error rates when compared to their components applied in isolation.
Resumo:
In multi-label classification, examples can be associated with multiple labels simultaneously. The task of learning from multi-label data can be addressed by methods that transform the multi-label classification problem into several single-label classification problems. The binary relevance approach is one of these methods, where the multi-label learning task is decomposed into several independent binary classification problems, one for each label in the set of labels, and the final labels for each example are determined by aggregating the predictions from all binary classifiers. However, this approach fails to consider any dependency among the labels. Aiming to accurately predict label combinations, in this paper we propose a simple approach that enables the binary classifiers to discover existing label dependency by themselves. An experimental study using decision trees, a kernel method as well as Naive Bayes as base-learning techniques shows the potential of the proposed approach to improve the multi-label classification performance.
Resumo:
Abstract Background Educational computer games are examples of computer-assisted learning objects, representing an educational strategy of growing interest. Given the changes in the digital world over the last decades, students of the current generation expect technology to be used in advancing their learning requiring a need to change traditional passive learning methodologies to an active multisensory experimental learning methodology. The objective of this study was to compare a computer game-based learning method with a traditional learning method, regarding learning gains and knowledge retention, as means of teaching head and neck Anatomy and Physiology to Speech-Language and Hearing pathology undergraduate students. Methods Students were randomized to participate to one of the learning methods and the data analyst was blinded to which method of learning the students had received. Students’ prior knowledge (i.e. before undergoing the learning method), short-term knowledge retention and long-term knowledge retention (i.e. six months after undergoing the learning method) were assessed with a multiple choice questionnaire. Students’ performance was compared considering the three moments of assessment for both for the mean total score and for separated mean scores for Anatomy questions and for Physiology questions. Results Students that received the game-based method performed better in the pos-test assessment only when considering the Anatomy questions section. Students that received the traditional lecture performed better in both post-test and long-term post-test when considering the Anatomy and Physiology questions. Conclusions The game-based learning method is comparable to the traditional learning method in general and in short-term gains, while the traditional lecture still seems to be more effective to improve students’ short and long-term knowledge retention.
Resumo:
This thesis is a collection of five independent but closely related studies. The overall purpose is to approach the analysis of learning outcomes from a perspective that combines three major elements, namely lifelonglifewide learning, human capital, and the benefits of learning. The approach is based on an interdisciplinary perspective of the human capital paradigm. It considers the multiple learning contexts that are responsible for the development of embodied potential – including formal, nonformal and informal learning – and the multiple outcomes – including knowledge, skills, economic, social and others– that result from learning. The studies also seek to examine the extent and relative influence of learning in different contexts on the formation of embodied potential and how in turn that affects economic and social well being. The first study combines the three major elements, lifelonglifewide learning, human capital, and the benefits of learning into one common conceptual framework. This study forms a common basis for the four empirical studies that follow. All four empirical studies use data from the International Adult Literacy Survey (IALS) to investigate the relationships among the major elements of the conceptual framework presented in the first study. Study I. A conceptual framework for the analysis of learning outcomes This study brings together some key concepts and theories that are relevant for the analysis of learning outcomes. Many of the concepts and theories have emerged from varied disciplines including economics, educational psychology, cognitive science and sociology, to name only a few. Accordingly, some of the research questions inherent in the framework relate to different disciplinary perspectives. The primary purpose is to create a common basis for formulating and testing hypotheses as well as to interpret the findings in the empirical studies that follow. In particular, the framework facilitates the process of theorizing and hypothesizing on the relationships and processes concerning lifelong learning as well as their antecedents and consequences. Study II. Determinants of literacy proficiency: A lifelong-lifewide learning perspective This study investigates lifelong and lifewide processes of skill formation. In particular, it seeks to estimate the substitutability and complementarity effects of learning in multiple settings over the lifespan on literacy skill formation. This is done by investigating the predictive capacity of major determinants of literacy proficiency that are associated with a variety of learning contexts including school, home, work, community and leisure. An identical structural model based on previous research is fitted to the IALS data for 18 countries. The results show that even after accounting for all factors, education remains the most important predictor of literacy proficiency. In all countries, however, the total effect of education is significantly mediated through further learning occurring at work, at home and in the community. Therefore, the job and other literacy related factors complement education in predicting literacy proficiency. This result points to a virtual cycle of lifelong learning, particularly to how educational attainment influences other learning behaviours throughout life. In addition, results show that home background as measured by parents’ education is also a strong predictor of literacy proficiency, but in many countries this occurs only if a favourable home background is complemented with some post-secondary education. Study III. The effect of literacy proficiency on earnings: An aggregated occupational approach using the Canadian IALS data This study uses data from the Canadian Adult Literacy Survey to estimate the earnings return to literacy skills. The approach adapts a labour segmented view of the labour market by aggregating occupations into seven types, enabling the estimation of the variable impact of literacy proficiency on earnings, both within and between different types of occupations. This is done using Hierarchical Linear Modeling (HLM). The method used to construct the aggregated occupational classification is based on analysis that considers the role of cognitive and other skills in relation to the nature of occupational tasks. Substantial premiums are found to be associated with some occupational types even after adjusting for within occupational differences in individual characteristics such as schooling, literacy proficiency, labour force experience and gender. Average years of schooling and average levels of literacy proficiency at the between level account for over two-thirds of the premiums. Within occupations, there are significant returns to schooling but they vary depending on the type of occupations. In contrast, the within occupational return of literacy proficiency is not necessarily significant. The latter depends on the type of occupation. Study IV: Determinants of economic and social outcomes from a lifewide learning perspective in Canada In this study the relationship between learning in different contexts, which span the lifewide learning dimension, and individual earnings on the one hand and community participation on the other are examined in separate but comparable models. Data from the Canadian Adult Literacy Survey are used to estimate structural models, which correspond closely to the common conceptual framework outlined in Study I. The findings suggest that the relationship between formal education and economic and social outcomes is complex with confounding effects. The results indicate that learning occurring in different contexts and for different reasons leads to different kinds of benefits. The latter finding suggests a potential trade-off between realizing economic and social benefits through learning that are taken for either job-related or personal-interest related reasons. Study V: The effects of learning on economic and social well being: A comparative analysis Using the same structural model as in Study IV, hypotheses are comparatively examined using the International Adult Literacy Survey data for Canada, Denmark, the Netherlands, Norway, the United Kingdom, and the United States. The main finding from Study IV is confirmed for an additional five countries, namely that the effect of initial schooling on well being is more complex than a direct one and it is significantly mediated by subsequent learning. Additionally, findings suggest that people who devote more time to learning for job-related reasons than learning for personal-interest related reasons experience higher levels of economic well being. Moreover, devoting too much time to learning for personal-interest related reasons has a negative effect on earnings except in Denmark. But the more time people devote to learning for personal-interest related reasons tends to contribute to higher levels of social well being. These results again suggest a trade-off in learning for different reasons and in different contexts.
Resumo:
The wide use of e-technologies represents a great opportunity for underserved segments of the population, especially with the aim of reintegrating excluded individuals back into society through education. This is particularly true for people with different types of disabilities who may have difficulties while attending traditional on-site learning programs that are typically based on printed learning resources. The creation and provision of accessible e-learning contents may therefore become a key factor in enabling people with different access needs to enjoy quality learning experiences and services. Another e-learning challenge is represented by m-learning (which stands for mobile learning), which is emerging as a consequence of mobile terminals diffusion and provides the opportunity to browse didactical materials everywhere, outside places that are traditionally devoted to education. Both such situations share the need to access materials in limited conditions and collide with the growing use of rich media in didactical contents, which are designed to be enjoyed without any restriction. Nowadays, Web-based teaching makes great use of multimedia technologies, ranging from Flash animations to prerecorded video-lectures. Rich media in e-learning can offer significant potential in enhancing the learning environment, through helping to increase access to education, enhance the learning experience and support multiple learning styles. Moreover, they can often be used to improve the structure of Web-based courses. These highly variegated and structured contents may significantly improve the quality and the effectiveness of educational activities for learners. For example, rich media contents allow us to describe complex concepts and process flows. Audio and video elements may be utilized to add a “human touch” to distance-learning courses. Finally, real lectures may be recorded and distributed to integrate or enrich on line materials. A confirmation of the advantages of these approaches can be seen in the exponential growth of video-lecture availability on the net, due to the ease of recording and delivering activities which take place in a traditional classroom. Furthermore, the wide use of assistive technologies for learners with disabilities injects new life into e-learning systems. E-learning allows distance and flexible educational activities, thus helping disabled learners to access resources which would otherwise present significant barriers for them. For instance, students with visual impairments have difficulties in reading traditional visual materials, deaf learners have trouble in following traditional (spoken) lectures, people with motion disabilities have problems in attending on-site programs. As already mentioned, the use of wireless technologies and pervasive computing may really enhance the educational learner experience by offering mobile e-learning services that can be accessed by handheld devices. This new paradigm of educational content distribution maximizes the benefits for learners since it enables users to overcome constraints imposed by the surrounding environment. While certainly helpful for users without disabilities, we believe that the use of newmobile technologies may also become a fundamental tool for impaired learners, since it frees them from sitting in front of a PC. In this way, educational activities can be enjoyed by all the users, without hindrance, thus increasing the social inclusion of non-typical learners. While the provision of fully accessible and portable video-lectures may be extremely useful for students, it is widely recognized that structuring and managing rich media contents for mobile learning services are complex and expensive tasks. Indeed, major difficulties originate from the basic need to provide a textual equivalent for each media resource composing a rich media Learning Object (LO). Moreover, tests need to be carried out to establish whether a given LO is fully accessible to all kinds of learners. Unfortunately, both these tasks are truly time-consuming processes, depending on the type of contents the teacher is writing and on the authoring tool he/she is using. Due to these difficulties, online LOs are often distributed as partially accessible or totally inaccessible content. Bearing this in mind, this thesis aims to discuss the key issues of a system we have developed to deliver accessible, customized or nomadic learning experiences to learners with different access needs and skills. To reduce the risk of excluding users with particular access capabilities, our system exploits Learning Objects (LOs) which are dynamically adapted and transcoded based on the specific needs of non-typical users and on the barriers that they can encounter in the environment. The basic idea is to dynamically adapt contents, by selecting them from a set of media resources packaged in SCORM-compliant LOs and stored in a self-adapting format. The system schedules and orchestrates a set of transcoding processes based on specific learner needs, so as to produce a customized LO that can be fully enjoyed by any (impaired or mobile) student.
Resumo:
[EN]In the framework of the European Higher Education Area, the assessment has been one of the most important aspects considered. In the Spanish Universities, one of the main differences with regard to the previous system is the incorporation of the continuous assessment to the evaluation process that is understood in several ways depending on the Universities, the courses and the lecturers. Focus on our context, a course of Mathematics of the first academic year in the Faculty of Business Administration at the University of Las Palmas de Gran Canaria (Spain), the continuous assessment has brought the preparation of a large amount of different tests to evaluate the students enrolled in it, therefore the incorporation of new tools and skills in order to make the teaching–learning process easier and more dynamic has become a need mainly in those degrees with a large number of students as the case we consider here. In this work we provide an efficient and effective way to elaborate random multiple-choice examina tion tests (although essay exams are also possible) by using Mathematica package and LATEXin order to make easier to the lectures the preparation of a large number of mid-term tests for a large number of students.
Resumo:
Machine learning comprises a series of techniques for automatic extraction of meaningful information from large collections of noisy data. In many real world applications, data is naturally represented in structured form. Since traditional methods in machine learning deal with vectorial information, they require an a priori form of preprocessing. Among all the learning techniques for dealing with structured data, kernel methods are recognized to have a strong theoretical background and to be effective approaches. They do not require an explicit vectorial representation of the data in terms of features, but rely on a measure of similarity between any pair of objects of a domain, the kernel function. Designing fast and good kernel functions is a challenging problem. In the case of tree structured data two issues become relevant: kernel for trees should not be sparse and should be fast to compute. The sparsity problem arises when, given a dataset and a kernel function, most structures of the dataset are completely dissimilar to one another. In those cases the classifier has too few information for making correct predictions on unseen data. In fact, it tends to produce a discriminating function behaving as the nearest neighbour rule. Sparsity is likely to arise for some standard tree kernel functions, such as the subtree and subset tree kernel, when they are applied to datasets with node labels belonging to a large domain. A second drawback of using tree kernels is the time complexity required both in learning and classification phases. Such a complexity can sometimes prevents the kernel application in scenarios involving large amount of data. This thesis proposes three contributions for resolving the above issues of kernel for trees. A first contribution aims at creating kernel functions which adapt to the statistical properties of the dataset, thus reducing its sparsity with respect to traditional tree kernel functions. Specifically, we propose to encode the input trees by an algorithm able to project the data onto a lower dimensional space with the property that similar structures are mapped similarly. By building kernel functions on the lower dimensional representation, we are able to perform inexact matchings between different inputs in the original space. A second contribution is the proposal of a novel kernel function based on the convolution kernel framework. Convolution kernel measures the similarity of two objects in terms of the similarities of their subparts. Most convolution kernels are based on counting the number of shared substructures, partially discarding information about their position in the original structure. The kernel function we propose is, instead, especially focused on this aspect. A third contribution is devoted at reducing the computational burden related to the calculation of a kernel function between a tree and a forest of trees, which is a typical operation in the classification phase and, for some algorithms, also in the learning phase. We propose a general methodology applicable to convolution kernels. Moreover, we show an instantiation of our technique when kernels such as the subtree and subset tree kernels are employed. In those cases, Direct Acyclic Graphs can be used to compactly represent shared substructures in different trees, thus reducing the computational burden and storage requirements.
Resumo:
This study focuses on the processes of change that firms undertake to overcome conditions of organizational rigidity and develop new dynamic capabilities, thanks to the contribution of external knowledge. When external contingencies highlight firms’ core rigidities, external actors can intervene in change projects, providing new competences to firms’ managers. Knowledge transfer and organizational learning processes can lead to the development of new dynamic capabilities. Existing literature does not completely explain how these processes develop and how external knowledge providers, as management consultants, influence them. Dynamic capabilities literature has become very rich in the last years; however, the models that explain how dynamic capabilities evolve are not particularly investigated. Adopting a qualitative approach, this research proposes four relevant case studies in which external actors introduce new knowledge within organizations, activating processes of change. Each case study consists of a management consulting project. Data are collected through in-depth interviews with consultants and managers. A large amount of documents supports evidences from interviews. A narrative approach is adopted to account for change processes and a synthetic approach is proposed to compare case studies along relevant dimensions. This study presents a model of capabilities evolution, supported by empirical evidence, to explain how external knowledge intervenes in capabilities evolution processes: first, external actors solve gaps between environmental demands and firms’ capabilities, changing organizational structures and routines; second, a knowledge transfer between consultants and managers leads to the creation of new ordinary capabilities; third, managers can develop new dynamic capabilities through a deliberate learning process that internalizes new tacit knowledge from consultants. After the end of the consulting project, two elements can influence the deliberate learning process: new external contingencies and changes in the perceptions about external actors.
Resumo:
In many application domains data can be naturally represented as graphs. When the application of analytical solutions for a given problem is unfeasible, machine learning techniques could be a viable way to solve the problem. Classical machine learning techniques are defined for data represented in a vectorial form. Recently some of them have been extended to deal directly with structured data. Among those techniques, kernel methods have shown promising results both from the computational complexity and the predictive performance point of view. Kernel methods allow to avoid an explicit mapping in a vectorial form relying on kernel functions, which informally are functions calculating a similarity measure between two entities. However, the definition of good kernels for graphs is a challenging problem because of the difficulty to find a good tradeoff between computational complexity and expressiveness. Another problem we face is learning on data streams, where a potentially unbounded sequence of data is generated by some sources. There are three main contributions in this thesis. The first contribution is the definition of a new family of kernels for graphs based on Directed Acyclic Graphs (DAGs). We analyzed two kernels from this family, achieving state-of-the-art results from both the computational and the classification point of view on real-world datasets. The second contribution consists in making the application of learning algorithms for streams of graphs feasible. Moreover,we defined a principled way for the memory management. The third contribution is the application of machine learning techniques for structured data to non-coding RNA function prediction. In this setting, the secondary structure is thought to carry relevant information. However, existing methods considering the secondary structure have prohibitively high computational complexity. We propose to apply kernel methods on this domain, obtaining state-of-the-art results.