903 resultados para Learning set
Resumo:
questions of forming of learning sets for artificial neural networks in problems of lossless data compression are considered. Methods of construction and use of learning sets are studied. The way of forming of learning set during training an artificial neural network on the data stream is offered.
Resumo:
Context traditionally has been regarded in vision research as a determinant for the interpretation of sensory information on the basis of previously acquired knowledge. Here we propose a novel, complementary perspective by showing that context also specifically affects visual category learning. In two experiments involving sets of Compound Gabor patterns we explored how context, as given by the stimulus set to be learned, affects the internal representation of pattern categories. In Experiment 1, we changed the (local) context of the individual signal classes by changing the configuration of the learning set. In Experiment 2, we varied the (global) context of a fixed class configuration by changing the degree of signal accentuation. Generalization performance was assessed in terms of the ability to recognize contrast-inverted versions of the learning patterns. Both contextual variations yielded distinct effects on learning and generalization thus indicating a change in internal category representation. Computer simulations suggest that the latter is related to changes in the set of attributes underlying the production rules of the categories. The implications of these findings for phenomena of contrast (in)variance in visual perception are discussed.
Resumo:
Approximate models (proxies) can be employed to reduce the computational costs of estimating uncertainty. The price to pay is that the approximations introduced by the proxy model can lead to a biased estimation. To avoid this problem and ensure a reliable uncertainty quantification, we propose to combine functional data analysis and machine learning to build error models that allow us to obtain an accurate prediction of the exact response without solving the exact model for all realizations. We build the relationship between proxy and exact model on a learning set of geostatistical realizations for which both exact and approximate solvers are run. Functional principal components analysis (FPCA) is used to investigate the variability in the two sets of curves and reduce the dimensionality of the problem while maximizing the retained information. Once obtained, the error model can be used to predict the exact response of any realization on the basis of the sole proxy response. This methodology is purpose-oriented as the error model is constructed directly for the quantity of interest, rather than for the state of the system. Also, the dimensionality reduction performed by FPCA allows a diagnostic of the quality of the error model to assess the informativeness of the learning set and the fidelity of the proxy to the exact model. The possibility of obtaining a prediction of the exact response for any newly generated realization suggests that the methodology can be effectively used beyond the context of uncertainty quantification, in particular for Bayesian inference and optimization.
Resumo:
Notre consommation en eau souterraine, en particulier comme eau potable ou pour l'irrigation, a considérablement augmenté au cours des années. De nombreux problèmes font alors leur apparition, allant de la prospection de nouvelles ressources à la remédiation des aquifères pollués. Indépendamment du problème hydrogéologique considéré, le principal défi reste la caractérisation des propriétés du sous-sol. Une approche stochastique est alors nécessaire afin de représenter cette incertitude en considérant de multiples scénarios géologiques et en générant un grand nombre de réalisations géostatistiques. Nous rencontrons alors la principale limitation de ces approches qui est le coût de calcul dû à la simulation des processus d'écoulements complexes pour chacune de ces réalisations. Dans la première partie de la thèse, ce problème est investigué dans le contexte de propagation de l'incertitude, oú un ensemble de réalisations est identifié comme représentant les propriétés du sous-sol. Afin de propager cette incertitude à la quantité d'intérêt tout en limitant le coût de calcul, les méthodes actuelles font appel à des modèles d'écoulement approximés. Cela permet l'identification d'un sous-ensemble de réalisations représentant la variabilité de l'ensemble initial. Le modèle complexe d'écoulement est alors évalué uniquement pour ce sousensemble, et, sur la base de ces réponses complexes, l'inférence est faite. Notre objectif est d'améliorer la performance de cette approche en utilisant toute l'information à disposition. Pour cela, le sous-ensemble de réponses approximées et exactes est utilisé afin de construire un modèle d'erreur, qui sert ensuite à corriger le reste des réponses approximées et prédire la réponse du modèle complexe. Cette méthode permet de maximiser l'utilisation de l'information à disposition sans augmentation perceptible du temps de calcul. La propagation de l'incertitude est alors plus précise et plus robuste. La stratégie explorée dans le premier chapitre consiste à apprendre d'un sous-ensemble de réalisations la relation entre les modèles d'écoulement approximé et complexe. Dans la seconde partie de la thèse, cette méthodologie est formalisée mathématiquement en introduisant un modèle de régression entre les réponses fonctionnelles. Comme ce problème est mal posé, il est nécessaire d'en réduire la dimensionnalité. Dans cette optique, l'innovation du travail présenté provient de l'utilisation de l'analyse en composantes principales fonctionnelles (ACPF), qui non seulement effectue la réduction de dimensionnalités tout en maximisant l'information retenue, mais permet aussi de diagnostiquer la qualité du modèle d'erreur dans cet espace fonctionnel. La méthodologie proposée est appliquée à un problème de pollution par une phase liquide nonaqueuse et les résultats obtenus montrent que le modèle d'erreur permet une forte réduction du temps de calcul tout en estimant correctement l'incertitude. De plus, pour chaque réponse approximée, une prédiction de la réponse complexe est fournie par le modèle d'erreur. Le concept de modèle d'erreur fonctionnel est donc pertinent pour la propagation de l'incertitude, mais aussi pour les problèmes d'inférence bayésienne. Les méthodes de Monte Carlo par chaîne de Markov (MCMC) sont les algorithmes les plus communément utilisés afin de générer des réalisations géostatistiques en accord avec les observations. Cependant, ces méthodes souffrent d'un taux d'acceptation très bas pour les problèmes de grande dimensionnalité, résultant en un grand nombre de simulations d'écoulement gaspillées. Une approche en deux temps, le "MCMC en deux étapes", a été introduite afin d'éviter les simulations du modèle complexe inutiles par une évaluation préliminaire de la réalisation. Dans la troisième partie de la thèse, le modèle d'écoulement approximé couplé à un modèle d'erreur sert d'évaluation préliminaire pour le "MCMC en deux étapes". Nous démontrons une augmentation du taux d'acceptation par un facteur de 1.5 à 3 en comparaison avec une implémentation classique de MCMC. Une question reste sans réponse : comment choisir la taille de l'ensemble d'entrainement et comment identifier les réalisations permettant d'optimiser la construction du modèle d'erreur. Cela requiert une stratégie itérative afin que, à chaque nouvelle simulation d'écoulement, le modèle d'erreur soit amélioré en incorporant les nouvelles informations. Ceci est développé dans la quatrième partie de la thèse, oú cette méthodologie est appliquée à un problème d'intrusion saline dans un aquifère côtier. -- Our consumption of groundwater, in particular as drinking water and for irrigation, has considerably increased over the years and groundwater is becoming an increasingly scarce and endangered resource. Nofadays, we are facing many problems ranging from water prospection to sustainable management and remediation of polluted aquifers. Independently of the hydrogeological problem, the main challenge remains dealing with the incomplete knofledge of the underground properties. Stochastic approaches have been developed to represent this uncertainty by considering multiple geological scenarios and generating a large number of realizations. The main limitation of this approach is the computational cost associated with performing complex of simulations in each realization. In the first part of the thesis, we explore this issue in the context of uncertainty propagation, where an ensemble of geostatistical realizations is identified as representative of the subsurface uncertainty. To propagate this lack of knofledge to the quantity of interest (e.g., the concentration of pollutant in extracted water), it is necessary to evaluate the of response of each realization. Due to computational constraints, state-of-the-art methods make use of approximate of simulation, to identify a subset of realizations that represents the variability of the ensemble. The complex and computationally heavy of model is then run for this subset based on which inference is made. Our objective is to increase the performance of this approach by using all of the available information and not solely the subset of exact responses. Two error models are proposed to correct the approximate responses follofing a machine learning approach. For the subset identified by a classical approach (here the distance kernel method) both the approximate and the exact responses are knofn. This information is used to construct an error model and correct the ensemble of approximate responses to predict the "expected" responses of the exact model. The proposed methodology makes use of all the available information without perceptible additional computational costs and leads to an increase in accuracy and robustness of the uncertainty propagation. The strategy explored in the first chapter consists in learning from a subset of realizations the relationship between proxy and exact curves. In the second part of this thesis, the strategy is formalized in a rigorous mathematical framework by defining a regression model between functions. As this problem is ill-posed, it is necessary to reduce its dimensionality. The novelty of the work comes from the use of functional principal component analysis (FPCA), which not only performs the dimensionality reduction while maximizing the retained information, but also allofs a diagnostic of the quality of the error model in the functional space. The proposed methodology is applied to a pollution problem by a non-aqueous phase-liquid. The error model allofs a strong reduction of the computational cost while providing a good estimate of the uncertainty. The individual correction of the proxy response by the error model leads to an excellent prediction of the exact response, opening the door to many applications. The concept of functional error model is useful not only in the context of uncertainty propagation, but also, and maybe even more so, to perform Bayesian inference. Monte Carlo Markov Chain (MCMC) algorithms are the most common choice to ensure that the generated realizations are sampled in accordance with the observations. Hofever, this approach suffers from lof acceptance rate in high dimensional problems, resulting in a large number of wasted of simulations. This led to the introduction of two-stage MCMC, where the computational cost is decreased by avoiding unnecessary simulation of the exact of thanks to a preliminary evaluation of the proposal. In the third part of the thesis, a proxy is coupled to an error model to provide an approximate response for the two-stage MCMC set-up. We demonstrate an increase in acceptance rate by a factor three with respect to one-stage MCMC results. An open question remains: hof do we choose the size of the learning set and identify the realizations to optimize the construction of the error model. This requires devising an iterative strategy to construct the error model, such that, as new of simulations are performed, the error model is iteratively improved by incorporating the new information. This is discussed in the fourth part of the thesis, in which we apply this methodology to a problem of saline intrusion in a coastal aquifer.
Resumo:
O presente estudo avaliou a formação de classes de estímulos através de um treino de duas discriminações condicionais (AB e AC) com pareamento consistente modelo-comparação correta, sem conseqüências diferenciais imediatas e fading, aplicando-se testes de simetria e de equivalência após cada bloco de treino. Participaram quatro crianças do pré-escolar que foram expostas ao procedimento de ensino, em duas etapas: na Etapa 1, com estímulos usuais e na Etapa 2 com estímulos não usuais Para cada modelo, três estímulos de comparação foram apresentados simultaneamente. Cada modelo foi emparelhado consistentemente, com os estímulos de comparação, sendo que o estímulo de comparação correto e o modelo apareceram em fading ao longo do treino. Relações simétricas foram demonstradas com dois participantes na Etapa 1, mas não ocorreram relações emergentes na Etapa 2. Dois participantes transferiram o desempenho obtido de uma etapa para outra com o treino discriminativo, como uma espécie de learning set arbitrário. Os resultados indicam que a sequência de treino com estímulos usuais e não usuais e a natureza dos estímulos na Etapa 1 foram variáveis relevantes.
Resumo:
Estudiosos questionam se a proficiência de macacos-prego (Cebus spp.) no uso de objetos como ferramentas, seria fruto de descobertas arbitrárias decorrentes dos frequentes comportamentos exploratórios desses primatas ou se seria devida à compreensão da função das ferramentas. Considerando que tais animais são capazes de modificar, transportar e fabricar ferramentas é possível propor que algum nível de compreensão esteja envolvido, ainda que não seja desvinculado de sua história de vida e sim construído a partir de uma série de interações com situações relevantes para a aquisição de um repertório generalizado de uso de ferramentas. A fim de investigar tal proposta foi realizada uma série de experimentos com dois grupos de macacos-prego (Cebus cf. apella) em que foi manipulada a história experimental desses animais. Todos os sujeitos passaram por repetidas exposições a um problema em que deveriam encaixar seis blocos de brinquedo para construir uma torre, usá-la para ter acesso a uma vareta distante, com essa vareta chegar a uma segunda vareta e encaixá-las formando uma haste longa o suficiente para permitir a recolha de pelotas de alimento em um equipamento. Enquanto dois sujeitos foram repetidamente expostos ao referido problema sem receber nenhum treino adicional, outros dois sujeitos tiveram uma rica história experimental construída, passando pelo treino em tarefas indiretamente relacionadas ao problema final entre as reapresentações do mesmo. Os sujeitos do primeiro grupo não foram capazes de resolver o problema, ao passo que os do segundo grupo o fizeram, ainda que não tenham sido diretamente treinados a isso. Concluiu-se que uma história relevante é fundamental para a chamada compreensão da solução de um problema e que essa compreensão ou insight são processos comportamentais adaptativos, em que habilidades aprendidas em um contexto específicosão transferidas para novos contextos a partir de processos básicos como a Generalização de Estímulos, a Generalização Funcional e o Learning Set.
Identificação automática das primeiras quebras em traços sísmicos por meio de uma rede neural direta
Resumo:
Apesar do avanço tecnológico ocorrido na prospecção sísmica, com a rotina dos levantamentos 2D e 3D, e o significativo aumento na quantidade de dados, a identificação dos tempos de chegada da onda sísmica direta (primeira quebra), que se propaga diretamente do ponto de tiro até a posição dos arranjos de geofones, permanece ainda dependente da avaliação visual do intérprete sísmico. O objetivo desta dissertação, insere-se no processamento sísmico com o intuito de buscar um método eficiente, tal que possibilite a simulação computacional do comportamento visual do intérprete sísmico, através da automação dos processos de tomada de decisão envolvidos na identificação das primeiras quebras em um traço sísmico. Visando, em última análise, preservar o conhecimento intuitivo do intérprete para os casos complexos, nos quais o seu conhecimento será, efetivamente, melhor aproveitado. Recentes descobertas na tecnologia neurocomputacional produziram técnicas que possibilitam a simulação dos aspectos qualitativos envolvidos nos processos visuais de identificação ou interpretação sísmica, com qualidade e aceitabilidade dos resultados. As redes neurais artificiais são uma implementação da tecnologia neurocomputacional e foram, inicialmente, desenvolvidas por neurobiologistas como modelos computacionais do sistema nervoso humano. Elas diferem das técnicas computacionais convencionais pela sua habilidade em adaptar-se ou aprender através de uma repetitiva exposição a exemplos, pela sua tolerância à falta de alguns dos componentes dos dados e pela sua robustez no tratamento com dados contaminados por ruído. O método aqui apresentado baseia-se na aplicação da técnica das redes neurais artificiais para a identificação das primeiras quebras nos traços sísmicos, a partir do estabelecimento de uma conveniente arquitetura para a rede neural artificial do tipo direta, treinada com o algoritmo da retro-propagação do erro. A rede neural artificial é entendida aqui como uma simulação computacional do processo intuitivo de tomada de decisão realizado pelo intérprete sísmico para a identificação das primeiras quebras nos traços sísmicos. A aplicabilidade, eficiência e limitações desta abordagem serão avaliadas em dados sintéticos obtidos a partir da teoria do raio.
Resumo:
Background. Previous knowledge of cervical lymph node compromise may be crucial to choose the best treatment strategy in oral squamous cell carcinoma (OSCC). Here we propose a set four genes, whose mRNA expression in the primary tumor predicts nodal status in OSCC, excluding tongue. Material and methods. We identified differentially expressed genes in OSCC with and without compromised lymph nodes using Differential Display RT-PCR. Known genes were chosen to be validated by means of Northern blotting or real time RT-PCR (qRT-PCR). Thereafter we constructed a Nodal Index (NI) using discriminant analysis in a learning set of 35 patients, which was further validated in a second independent group of 20 patients. Results. Of the 63 differentially expressed known genes identified comparing three lymph node positive (pN+) and three negative (pN0) primary tumors, 23 were analyzed by Northern analysis or RT-PCR in 49 primary tumors. Six genes confirmed as differentially expressed were used to construct a NI, as the best set predictive of lymph nodal status, with the final result including four genes. The NI was able to correctly classify 32 of 35 patients comprising the learning group (88.6%; p = 0.009). Casein kinase 1alpha1 and scavenger receptor class B, member 2 were found to be up regulated in pN + group in contrast to small proline-rich protein 2B and Ras-GTPase activating protein SH3 domain-binding protein 2 which were upregulated in the pN0 group. We validated further our NI in an independent set of 20 primary tumors, 11 of them pN0 and nine pN+ with an accuracy of 80.0% (p = 0.012). Conclusions. The NI was an independent predictor of compromised lymph nodes, taking into the consideration tumor size and histological grade. The genes identified here that integrate our "Nodal Index" model are predictive of lymph node metastasis in OSCC.
Resumo:
Approximate models (proxies) can be employed to reduce the computational costs of estimating uncertainty. The price to pay is that the approximations introduced by the proxy model can lead to a biased estimation. To avoid this problem and ensure a reliable uncertainty quantification, we propose to combine functional data analysis and machine learning to build error models that allow us to obtain an accurate prediction of the exact response without solving the exact model for all realizations. We build the relationship between proxy and exact model on a learning set of geostatistical realizations for which both exact and approximate solvers are run. Functional principal components analysis (FPCA) is used to investigate the variability in the two sets of curves and reduce the dimensionality of the problem while maximizing the retained information. Once obtained, the error model can be used to predict the exact response of any realization on the basis of the sole proxy response. This methodology is purpose-oriented as the error model is constructed directly for the quantity of interest, rather than for the state of the system. Also, the dimensionality reduction performed by FPCA allows a diagnostic of the quality of the error model to assess the informativeness of the learning set and the fidelity of the proxy to the exact model. The possibility of obtaining a prediction of the exact response for any newly generated realization suggests that the methodology can be effectively used beyond the context of uncertainty quantification, in particular for Bayesian inference and optimization.
Resumo:
Children aged between 3 and 7 years were taught simple and dimension-abstracted oddity discrimination using learning-set training techniques, in which isomorphic problems with varying content were presented with verbal explanation and feedback. Following the training phase, simple oddity (SO), dimension-abstracted oddity with one or two irrelevant dimensions, and non-oddity (NO) tasks were presented (without feedback) to determine the basis of solution. Although dimension-abstracted oddity requires discrimination based on a stimulus that is different from the others, which are all the same as each other on the relevant dimension, this was not the major strategy. The data were more consistent with use of a simple oddity strategy by 3- to 4-year-olds, and a most different strategy by 6- to 7-year-olds. These strategies are interpreted as reducing task complexity. (C) 2002 Elsevier Science Inc. All rights reserved.
Resumo:
To investigate the control mechanisms used in adapting to position-dependent forces, subjects performed 150 horizontal reaching movements over 25 cm in the presence of a position-dependent parabolic force field (PF). The PF acted only over the first 10 cm of the movement. On every fifth trial, a virtual mechanical guide (double wall) constrained subjects to move along a straight-line path between the start and target positions. Its purpose was to register lateral force to track formation of an internal model of the force field, and to look for evidence of possible alternative adaptive strategies. The force field produced a force to the right, which initially caused subjects to deviate in that direction. They reacted by producing deviations to the left, into the force field, as early as the second trial. Further adaptation resulted in rapid exponential reduction of kinematic error in the latter portion of the movement, where the greatest perturbation to the handpath was initially observed, whereas there was little modification of the handpath in the region where the PF was active. Significant force directed to counteract the PF was measured on the first guided trial, and was modified during the first half of the learning set. The total force impulse in the region of the PF increased throughout the learning trials, but it always remained less than that produced by the PF. The force profile did not resemble a mirror image of the PF in that it tended to be more trapezoidal than parabolic in shape. As in previous studies of force-field adaptation, we found that changes in muscle activation involved a general increase in the activity of all muscles, which increased arm stiffness, and selectively-greater increases in the activation of muscles which counteracted the PF. With training, activation was exponentially reduced, albeit more slowly than kinematic error. Progressive changes in kinematics and EMG occurred predominantly in the region of the workspace beyond the force field. We suggest that constraints on muscle mechanics limit the ability of the central nervous system to employ an inverse dynamics model to nullify impulse-like forces by generating mirror-image forces. Consequently, subjects adopted a strategy of slightly overcompensating for the first half of the force field, then allowing the force field to push them in the opposite direction. Muscle activity patterns in the region beyond the boundary of the force field were subsequently adjusted because of the relatively-slow response of the second-order mechanics of muscle impedance to the force impulse.
Resumo:
This work presents a model for development of project proposals by students as an approach to teaching information technology while promoting entrepreneurship and reflection. In teams of 3 to 5 participants, students elaborate a project proposal on a topic they have negotiated with each other and with the teacher. The project domain is related to the practical application of state-of-theart information technology in areas of substantial public interest or of immediate interest to the participants. This gives them ample opportunities for reflection not only on technical but also on social, economic, environmental and other dimensions of information technology. This approach has long been used with students of different years and programs of study at the Faculty of Mathematics and Informatics, Plovdiv University “Paisiy Hilendarski”. It has been found to develop all eight key competences for lifelong learning set forth in the Reference Framework and procedural skills required in real life.
Resumo:
In this paper, we present one approach for extending the learning set of a classification algorithm with additional metadata. It is used as a base for giving appropriate names to found regularities. The analysis of correspondence between connections established in the attribute space and existing links between concepts can be used as a test for creation of an adequate model of the observed world. Meta-PGN classifier is suggested as a possible tool for establishing these connections. Applying this approach in the field of content-based image retrieval of art paintings provides a tool for extracting specific feature combinations, which represent different sides of artists' styles, periods and movements.
Resumo:
Study abroad is a highly encouraged component of the undergrad business program at Bella Lake University, with the large majority of undergraduate students choosing to study and live in a foreign country to expand on their international experiences. Currently, there is no learning structure or learning outcome expectation for students that will participate in the study abroad experience. This project focuses on the development of a course that supports students in the pre-departure phase of their study abroad journey and prepares them to set goals, understand the learning process and the practices of experiential learning to encourage students to achieve both personal and professional goals, and encourage the development of intercultural competence. The discourse surrounding the perceptions and efficacy of the course development is based on a self-assessment survey completed by 121 undergraduate business students that participated in the pre-departure sessions prior to leaving for study abroad in March 2016. The self-assessment results overall showed that the course achieved its aims with the majority of students rating that they were more likely to understand and engage in experiential learning, set goals for their study abroad experience and felt more prepared for study abroad after attending the pre-departure sessions. The project concludes that in order for the pre-departure course to maintain its value, the conversation with students surrounding experiential learning in study abroad needs to continue with further course development focusing on both during and post-study abroad. Further exploration can also be done to find varying ways to motivate different students to engage in the learning potential of study abroad.
Resumo:
Employing critical pedagogy and transformative theory as a theoretical framework, I examined a learning process associated with building capacity in community-based organizations (CBOs) through an investigation of the Institutional Capacity Building Program (ICBP) initiated by a Foundation. The study sought to: (a) examine the importance of institutional capacity building for individual and community development; (b) investigate elements of a process associated with a program and characteristics of a learning process for building capacity in CBOs; and (c) analyze the Foundation’s approach to synthesizing, systematizing, and sharing learning. The study used a narrative research design that included 3 one-on-one, hour-long interviews with 2 women having unique vantage points in ICBP: one is a program facilitator working at the Foundation and the other runs a CBO supported by the Foundation. The interviews’ semistructured questions allowed interviewees to share stories regarding their experience with the learning process of ICB and enabled themes to emerge from their day-to-day experience. Through the analysis of this learning process for institutional capacity building, a few lessons can be drawn from the experience of the Foundation.