774 resultados para real world learning


Relevância:

90.00% 90.00%

Publicador:

Resumo:

Background: A genetic network can be represented as a directed graph in which a node corresponds to a gene and a directed edge specifies the direction of influence of one gene on another. The reconstruction of such networks from transcript profiling data remains an important yet challenging endeavor. A transcript profile specifies the abundances of many genes in a biological sample of interest. Prevailing strategies for learning the structure of a genetic network from high-dimensional transcript profiling data assume sparsity and linearity. Many methods consider relatively small directed graphs, inferring graphs with up to a few hundred nodes. This work examines large undirected graphs representations of genetic networks, graphs with many thousands of nodes where an undirected edge between two nodes does not indicate the direction of influence, and the problem of estimating the structure of such a sparse linear genetic network (SLGN) from transcript profiling data. Results: The structure learning task is cast as a sparse linear regression problem which is then posed as a LASSO (l1-constrained fitting) problem and solved finally by formulating a Linear Program (LP). A bound on the Generalization Error of this approach is given in terms of the Leave-One-Out Error. The accuracy and utility of LP-SLGNs is assessed quantitatively and qualitatively using simulated and real data. The Dialogue for Reverse Engineering Assessments and Methods (DREAM) initiative provides gold standard data sets and evaluation metrics that enable and facilitate the comparison of algorithms for deducing the structure of networks. The structures of LP-SLGNs estimated from the INSILICO1, INSILICO2 and INSILICO3 simulated DREAM2 data sets are comparable to those proposed by the first and/or second ranked teams in the DREAM2 competition. The structures of LP-SLGNs estimated from two published Saccharomyces cerevisae cell cycle transcript profiling data sets capture known regulatory associations. In each S. cerevisiae LP-SLGN, the number of nodes with a particular degree follows an approximate power law suggesting that its degree distributions is similar to that observed in real-world networks. Inspection of these LP-SLGNs suggests biological hypotheses amenable to experimental verification. Conclusion: A statistically robust and computationally efficient LP-based method for estimating the topology of a large sparse undirected graph from high-dimensional data yields representations of genetic networks that are biologically plausible and useful abstractions of the structures of real genetic networks. Analysis of the statistical and topological properties of learned LP-SLGNs may have practical value; for example, genes with high random walk betweenness, a measure of the centrality of a node in a graph, are good candidates for intervention studies and hence integrated computational – experimental investigations designed to infer more realistic and sophisticated probabilistic directed graphical model representations of genetic networks. The LP-based solutions of the sparse linear regression problem described here may provide a method for learning the structure of transcription factor networks from transcript profiling and transcription factor binding motif data.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

This presentation discusses and critiques a current case study of a project in which Early Childhood preservice teachers are working in partnership with Design students to develop principles and concepts for the design and construction of an early childhood centre. This centre, to be built on the grounds of the iconic Lone Pine Koala Sanctuary in Brisbane , focuses on Education for Sustainability (EfS), sustainable design and sustainable business. Interdisciplinary initiatives between QUT staff and students from two Faculties (Education and Creative Industries) have been situated in the realworld context of this project. This practical, authentic project has seen stakeholders take an interdisciplinary approach to sustainability, opening up new ways of thinking about early childhood centre design, particularly with respect to operation and function. Interdisciplinarity and a commitment to genuine partnerships have created intellectual spaces to re-think the potential of the disciplines to be interwoven so that future professionals from different fields might come together to learn from each other and to address the sustainability imperative. The case study documents and explores the possibilities that the Lone Pine project offers for academics and students from Early Childhood and Design to collaboratively inform the Sanctuary’s vision for the Centre. The research examines how students benefit from practical, real world, community-integrated learning; how academic staff across two disciplines are able to work collaboratively within a real-world context; and how external stakeholders experience and benefit from the partnership with university staff and students. Data were collected via a series of focus group and individual interviews designed to explore how the various stakeholders (staff, students, business partners) experienced their involvement in the interdisciplinary project. Inductive and deductive thematic analysis of these data suggest many benefits for participants as well as a number of challenges. Findings suggest that the project has provided students with ‘real world’ partnerships that reposition early childhood students’ identities from ‘novice’ to ‘professional’, where their knowledge, expertise and perspectives are simultaneously validated and challenged in their work with designers. These partnerships are enabling preservice teachers to practice a new model of early childhood leadership in sustainability, one that is vital for leading for change in an increasingly complex world. This presentation celebrates, critiques and problematises this project, exploring wider implications for other contexts in which university staff and students may seek to work across traditional boundaries, thus building partnerships for change.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

BACKGROUND OR CONTEXT The higher education sector plays an important role in encouraging students into the STEM pipeline through fostering partnerships with schools, building on universities long tradition in engagement and outreach to secondary schools. Numerous activities focus on integrated STEM learning experiences aimed at developing conceptual scientific and mathematical knowledge with opportunities for students to show and develop skills in working with each other and actively engaging in discussion, decision making and collaborative problem solving. (NAS, 2013; AIG, 2015; OCS, 2014). This highlights the importance of the development and delivery of engaging integrated STEM activities connected to the curriculum to inspire the next generation of scientists and engineers and generally preparing students for post-secondary success. The broad research objective is to gain insight into which engagement activities and to what level they influence secondary school students’ selection of STEM-related career choices at universities. PURPOSE OR GOAL To evaluate and determine the effectiveness of STEM engagement activities impacting student decision making in choosing a STEM-related degree choice at university. APPROACH A survey was conducted with first-year domestic students studying STEM-related fieldswithin the Science and Engineering Faculty at Queensland University of Technology. Of the domestic students commencing in 2015, 29% responded to the survey. The survey was conducted using Survey Monkey and included a variety of questions ranging from academic performance at school to inspiration for choosing a STEM degree. Responses were analysed on a range of factors to evaluate the influence on students’ decisions to study STEM and whether STEM high school engagement activities impacted these decisions. To achieve this the timing of decision making for students choice in study area, degree, and university is compared with the timing of STEM engagement activities. DISCUSSION Statistical analysis using SPSS was carried out on survey data looking at reasons for choosing STEM degrees in terms of gender, academic performance and major influencers in their decision making. It was found that students choose their university courses based on what subjects they enjoyed and exceled at in school. These results found a high correlation between enjoyment of a school subject and their interest in pursuing this subject at university and beyond. Survey results indicated students are heavily influenced by their subject teachers and parents in their choice of STEM-related disciplines. In terms of career choice and when students make their decision, 60% have decided on a broad area of study by year 10, whilst only 15% had decided on a specific course and 10% had decided on which university. The timing of secondary STEM engagement activities is seen as a critical influence on choosing STEM disciplines or selection of senior school subjects with 80% deciding on specific degree between year 11 and 12 and 73% making a decision on which university in year 12. RECOMMENDATIONS/IMPLICATIONS/CONCLUSION Although the data does not support that STEM engagement activities increase the likelihood of STEM-related degree choice, the evidence suggests the students who have participated in STEM activities associate their experiences with their choice to pursue a STEM-related course. It is important for universities to continue to provide quality engaging and inspirational learning experiences in STEM, to identify and build on students’ early interest and engagement, increase STEM knowledge and awareness, engage them in interdisciplinary project-based STEM practices, and provide them with real-world application experiences to sustain their interest.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

This presentation discusses and critiques a current case study of a project in which Early Childhood preservice teachers are working in partnership with Design students to develop principles and concepts for the design and construction of an early childhood centre. This centre, to be built on the grounds of the iconic Lone Pine Koala Sanctuary in Brisbane , focuses on Education for Sustainability (EfS), sustainable design and sustainable business. Interdisciplinary initiatives between QUT staff and students from two Faculties (Education and Creative Industries) have been situated in the realworld context of this project. This practical, authentic project has seen stakeholders take an interdisciplinary approach to sustainability, opening up new ways of thinking about early childhood centre design, particularly with respect to operation and function. Interdisciplinarity and a commitment to genuine partnerships have created intellectual spaces to re-think the potential of the disciplines to be interwoven so that future professionals from different fields might come together to learn from each other and to address the sustainability imperative. The case study documents and explores the possibilities that the Lone Pine project offers for academics and students from Early Childhood and Design to collaboratively inform the Sanctuary’s vision for the Centre. The research examines how students benefit from practical, real world, community-integrated learning; how academic staff across two disciplines are able to work collaboratively within a real-world context; and how external stakeholders experience and benefit from the partnership with university staff and students. Data were collected via a series of focus group and individual interviews designed to explore how the various stakeholders (staff, students, business partners) experienced their involvement in the interdisciplinary project. Inductive and deductive thematic analysis of these data suggest many benefits for participants as well as a number of challenges. Findings suggest that the project has provided students with ‘real world’ partnerships that reposition early childhood students’ identities from ‘novice’ to ‘professional’, where their knowledge, expertise and perspectives are simultaneously validated and challenged in their work with designers. These partnerships are enabling preservice teachers to practice a new model of early childhood leadership in sustainability, one that is vital for leading for change in an increasingly complex world. This presentation celebrates, critiques and problematises this project, exploring wider implications for other contexts in which university staff and students may seek to work across traditional boundaries, thus building partnerships for change.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Video surveillance infrastructure has been widely installed in public places for security purposes. However, live video feeds are typically monitored by human staff, making the detection of important events as they occur difficult. As such, an expert system that can automatically detect events of interest in surveillance footage is highly desirable. Although a number of approaches have been proposed, they have significant limitations: supervised approaches, which can detect a specific event, ideally require a large number of samples with the event spatially and temporally localised; while unsupervised approaches, which do not require this demanding annotation, can only detect whether an event is abnormal and not specific event types. To overcome these problems, we formulate a weakly-supervised approach using Kullback-Leibler (KL) divergence to detect rare events. The proposed approach leverages the sparse nature of the target events to its advantage, and we show that this data imbalance guarantees the existence of a decision boundary to separate samples that contain the target event from those that do not. This trait, combined with the coarse annotation used by weakly supervised learning (that only indicates approximately when an event occurs), greatly reduces the annotation burden while retaining the ability to detect specific events. Furthermore, the proposed classifier requires only a decision threshold, simplifying its use compared to other weakly supervised approaches. We show that the proposed approach outperforms state-of-the-art methods on a popular real-world traffic surveillance dataset, while preserving real time performance.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

This commentary was stimulated by Yeping Li's first editorial (2014) citing one of the journal's goals as adding multidisciplinary perspectives to current studies of single disciplines comprising the focus of other journals. In this commentary I argue for a greater focus on STEM integration, with a more equitable representation of the four disciplines in studies purporting to advance STEM learning. The STEM acronym is often used in reference to just one of the disciplines, commonly science. Although the integration of STEM disciplines is increasingly advocated in the literature, studies that address multiple disciplines appear scant with mixed findings and inadequate directions for STEM advancement. Perspectives on how discipline integration can be achieved are varied, with reference to multidisciplinary, interdisciplinary, and transdisciplinary approaches adding to the debates. Such approaches include core concepts and skills being taught separately in each discipline but housed within a common theme; the introduction of closely linked concepts and skills from two or more disciplines with the aim of deepening understanding and skills; and the adoption of a transdisciplinary approach, where knowledge and skills from two or more disciplines are applied to real-world problems and projects with the aim of shaping the total learning experience. Research that targets STEM integration is an embryonic field with respect to advancing curriculum development and various student outcomes. For example, we still need more studies on how student learning outcomes arise not only from different forms of STEM integration but also from the particular disciplines that are being integrated. As noted in this commentary, it seems that mathematics learning benefits less than the other disciplines in programs claiming to focus on STEM integration. Factors contributing to this finding warrant more scrutiny. Likewise, learning outcomes for engineering within K-12 integrated STEM programs appear under-researched. This commentary advocates a greater focus on these two disciplines within integrated STEM education research. Drawing on recommendations from the literature, suggestions are offered for addressing the challenges of integrating multiple disciplines faced by the STEM community.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Gaussian Processes (GPs) are promising Bayesian methods for classification and regression problems. They have also been used for semi-supervised learning tasks. In this paper, we propose a new algorithm for solving semi-supervised binary classification problem using sparse GP regression (GPR) models. It is closely related to semi-supervised learning based on support vector regression (SVR) and maximum margin clustering. The proposed algorithm is simple and easy to implement. It gives a sparse solution directly unlike the SVR based algorithm. Also, the hyperparameters are estimated easily without resorting to expensive cross-validation technique. Use of sparse GPR model helps in making the proposed algorithm scalable. Preliminary results on synthetic and real-world data sets demonstrate the efficacy of the new algorithm.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

This paper studies the problem of constructing robust classifiers when the training is plagued with uncertainty. The problem is posed as a Chance-Constrained Program (CCP) which ensures that the uncertain data points are classified correctly with high probability. Unfortunately such a CCP turns out to be intractable. The key novelty is in employing Bernstein bounding schemes to relax the CCP as a convex second order cone program whose solution is guaranteed to satisfy the probabilistic constraint. Prior to this work, only the Chebyshev based relaxations were exploited in learning algorithms. Bernstein bounds employ richer partial information and hence can be far less conservative than Chebyshev bounds. Due to this efficient modeling of uncertainty, the resulting classifiers achieve higher classification margins and hence better generalization. Methodologies for classifying uncertain test data points and error measures for evaluating classifiers robust to uncertain data are discussed. Experimental results on synthetic and real-world datasets show that the proposed classifiers are better equipped to handle data uncertainty and outperform state-of-the-art in many cases.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

We study the problem of uncertainty in the entries of the Kernel matrix, arising in SVM formulation. Using Chance Constraint Programming and a novel large deviation inequality we derive a formulation which is robust to such noise. The resulting formulation applies when the noise is Gaussian, or has finite support. The formulation in general is non-convex, but in several cases of interest it reduces to a convex program. The problem of uncertainty in kernel matrix is motivated from the real world problem of classifying proteins when the structures are provided with some uncertainty. The formulation derived here naturally incorporates such uncertainty in a principled manner leading to significant improvements over the state of the art. 1.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Even though several techniques have been proposed in the literature for achieving multiclass classification using Support Vector Machine(SVM), the scalability aspect of these approaches to handle large data sets still needs much of exploration. Core Vector Machine(CVM) is a technique for scaling up a two class SVM to handle large data sets. In this paper we propose a Multiclass Core Vector Machine(MCVM). Here we formulate the multiclass SVM problem as a Quadratic Programming(QP) problem defining an SVM with vector valued output. This QP problem is then solved using the CVM technique to achieve scalability to handle large data sets. Experiments done with several large synthetic and real world data sets show that the proposed MCVM technique gives good generalization performance as that of SVM at a much lesser computational expense. Further, it is observed that MCVM scales well with the size of the data set.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Support Vector Clustering has gained reasonable attention from the researchers in exploratory data analysis due to firm theoretical foundation in statistical learning theory. Hard Partitioning of the data set achieved by support vector clustering may not be acceptable in real world scenarios. Rough Support Vector Clustering is an extension of Support Vector Clustering to attain a soft partitioning of the data set. But the Quadratic Programming Problem involved in Rough Support Vector Clustering makes it computationally expensive to handle large datasets. In this paper, we propose Rough Core Vector Clustering algorithm which is a computationally efficient realization of Rough Support Vector Clustering. Here Rough Support Vector Clustering problem is formulated using an approximate Minimum Enclosing Ball problem and is solved using an approximate Minimum Enclosing Ball finding algorithm. Experiments done with several Large Multi class datasets such as Forest cover type, and other Multi class datasets taken from LIBSVM page shows that the proposed strategy is efficient, finds meaningful soft cluster abstractions which provide a superior generalization performance than the SVM classifier.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

The questions that one should answer in engineering computations - deterministic, probabilistic/randomized, as well as heuristic - are (i) how good the computed results/outputs are and (ii) how much the cost in terms of amount of computation and the amount of storage utilized in getting the outputs is. The absolutely errorfree quantities as well as the completely errorless computations done in a natural process can never be captured by any means that we have at our disposal. While the computations including the input real quantities in nature/natural processes are exact, all the computations that we do using a digital computer or are carried out in an embedded form are never exact. The input data for such computations are also never exact because any measuring instrument has inherent error of a fixed order associated with it and this error, as a matter of hypothesis and not as a matter of assumption, is not less than 0.005 per cent. Here by error we imply relative error bounds. The fact that exact error is never known under any circumstances and any context implies that the term error is nothing but error-bounds. Further, in engineering computations, it is the relative error or, equivalently, the relative error-bounds (and not the absolute error) which is supremely important in providing us the information regarding the quality of the results/outputs. Another important fact is that inconsistency and/or near-consistency in nature, i.e., in problems created from nature is completely nonexistent while in our modelling of the natural problems we may introduce inconsistency or near-inconsistency due to human error or due to inherent non-removable error associated with any measuring device or due to assumptions introduced to make the problem solvable or more easily solvable in practice. Thus if we discover any inconsistency or possibly any near-inconsistency in a mathematical model, it is certainly due to any or all of the three foregoing factors. We do, however, go ahead to solve such inconsistent/near-consistent problems and do get results that could be useful in real-world situations. The talk considers several deterministic, probabilistic, and heuristic algorithms in numerical optimisation, other numerical and statistical computations, and in PAC (probably approximately correct) learning models. It highlights the quality of the results/outputs through specifying relative error-bounds along with the associated confidence level, and the cost, viz., amount of computations and that of storage through complexity. It points out the limitation in error-free computations (wherever possible, i.e., where the number of arithmetic operations is finite and is known a priori) as well as in the usage of interval arithmetic. Further, the interdependence among the error, the confidence, and the cost is discussed.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

In this paper we study the problem of designing SVM classifiers when the kernel matrix, K, is affected by uncertainty. Specifically K is modeled as a positive affine combination of given positive semi definite kernels, with the coefficients ranging in a norm-bounded uncertainty set. We treat the problem using the Robust Optimization methodology. This reduces the uncertain SVM problem into a deterministic conic quadratic problem which can be solved in principle by a polynomial time Interior Point (IP) algorithm. However, for large-scale classification problems, IP methods become intractable and one has to resort to first-order gradient type methods. The strategy we use here is to reformulate the robust counterpart of the uncertain SVM problem as a saddle point problem and employ a special gradient scheme which works directly on the convex-concave saddle function. The algorithm is a simplified version of a general scheme due to Juditski and Nemirovski (2011). It achieves an O(1/T-2) reduction of the initial error after T iterations. A comprehensive empirical study on both synthetic data and real-world protein structure data sets show that the proposed formulations achieve the desired robustness, and the saddle point based algorithm outperforms the IP method significantly.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

In many real world prediction problems the output is a structured object like a sequence or a tree or a graph. Such problems range from natural language processing to compu- tational biology or computer vision and have been tackled using algorithms, referred to as structured output learning algorithms. We consider the problem of structured classifi- cation. In the last few years, large margin classifiers like sup-port vector machines (SVMs) have shown much promise for structured output learning. The related optimization prob -lem is a convex quadratic program (QP) with a large num-ber of constraints, which makes the problem intractable for large data sets. This paper proposes a fast sequential dual method (SDM) for structural SVMs. The method makes re-peated passes over the training set and optimizes the dual variables associated with one example at a time. The use of additional heuristics makes the proposed method more efficient. We present an extensive empirical evaluation of the proposed method on several sequence learning problems.Our experiments on large data sets demonstrate that the proposed method is an order of magnitude faster than state of the art methods like cutting-plane method and stochastic gradient descent method (SGD). Further, SDM reaches steady state generalization performance faster than the SGD method. The proposed SDM is thus a useful alternative for large scale structured output learning.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Ranking problems have become increasingly important in machine learning and data mining in recent years, with applications ranging from information retrieval and recommender systems to computational biology and drug discovery. In this paper, we describe a new ranking algorithm that directly maximizes the number of relevant objects retrieved at the absolute top of the list. The algorithm is a support vector style algorithm, but due to the different objective, it no longer leads to a quadratic programming problem. Instead, the dual optimization problem involves l1, ∞ constraints; we solve this dual problem using the recent l1, ∞ projection method of Quattoni et al (2009). Our algorithm can be viewed as an l∞-norm extreme of the lp-norm based algorithm of Rudin (2009) (albeit in a support vector setting rather than a boosting setting); thus we refer to the algorithm as the ‘Infinite Push’. Experiments on real-world data sets confirm the algorithm’s focus on accuracy at the absolute top of the list.