786 resultados para Modeling Non-Verbal Behaviors Using Machine Learning
Resumo:
Requirement engineering is a key issue in the development of a software project. Like any other development activity it is not without risks. This work is about the empirical study of risks of requirements by applying machine learning techniques, specifically Bayesian networks classifiers. We have defined several models to predict the risk level for a given requirement using three dataset that collect metrics taken from the requirement specifications of different projects. The classification accuracy of the Bayesian models obtained is evaluated and compared using several classification performance measures. The results of the experiments show that the Bayesians networks allow obtaining valid predictors. Specifically, a tree augmented network structure shows a competitive experimental performance in all datasets. Besides, the relations established between the variables collected to determine the level of risk in a requirement, match with those set by requirement engineers. We show that Bayesian networks are valid tools for the automation of risks assessment in requirement engineering.
Resumo:
Pretendeu-se com este projecto de investigação estudar a interação didática co-construída por alunos do ensino superior em moldes de aprendizagem colaborativa na aula de Inglês língua estrangeira, com enfoque na dimensão sócio-afetiva da aprendizagem. Na base do quadro teórico encontra-se o pressuposto de que o conhecimento é algo dinâmico e construído colaborativamente, e que é na interação didática que emergem os comportamentos verbais reveladores do Saber―Ser/Estar/Aprender dos sujeitos, nomeadamente através da coconstrução e negociação de sentidos. Subjacente portanto ao estudo está a convicção de que “o trabalho crítico sobre a interação permite entender os modos relacionais entre os sujeitos pedagógicos, as relações interpessoais que se estabelecem e articular o desenvolvimento linguístico-comunicativo com o desenvolvimento pessoal e social dos alunos” (Araújo e Sá & Andrade, 2002, p. 82). Esta investigação centra-se exclusivamente nos aprendentes, na sequência de indicações provenientes da revisão de literatura, as quais apontam para uma lacuna nas investigações efetuadas até à data, referente ao número insuficiente de estudos dedicado à interação didática interpares, já que a grande maioria dos estudos se dirige para a relação professor-aluno (cf. Baker & Clark, 2010; Hellermann, 2008; O'Donnell & King, 2014). Por outro lado, o estado da arte relativo às investigações focalizadas na interacção entre aprendentes permite concluir que a melhor forma de exponenciar esta interação será através da aprendizagem colaborativa (cf. Johnson, Johnson, & Stanne, 2000; Slavin, 2014; Smith, Sheppard, Johnson, & Johnson, 2005). Circunscrevemos o nosso estudo à dimensão sócio-afetiva das estratégias de aprendizagem que ocorrem nessas interações, já que a revisão da literatura fez evidenciar a correlação positiva da aprendizagem colaborativa com as dimensões social e afetiva da interação (cf. Byun et al., 2012): por um lado, a dinâmica de grupo numa aula de língua estrangeira contribui grandemente para uma perceção afetiva favorável do processo de aprendizagem, incrementando igualmente a quantidade e a qualidade da interação (cf. Felder & Brent, 2007); por outro lado, a existência, na aprendizagem colaborativa, dos fenómenos de correção dos pares e de negociação de sentidos estimula a emergência da dimensão sócio-afetiva da aprendizagem de uma língua estrangeira (cf. Campbell & Kryszewska,1992; Hadfield, 1992; Macaro, 2005). É neste enquadramento teórico que se situam as nossas questões e objetivos de investigação. Em primeiro lugar procurámos saber como é que um grupo de aprendentes de Inglês língua estrangeira do ensino superior perceciona as estratégias de aprendizagem sócio-afetivas que utiliza em contexto de sala de aula, no âmbito da aprendizagem colaborativa e nãocolaborativa. Procurámos igualmente indagar quais as estratégias de aprendizagem sócio-afetivas passíveis de serem identificadas neste grupo de aprendentes, em situação de interação didática, em contexto de aprendizagem colaborativa. Finalmente, questionámo-nos sobre a relação entre a perceção que estes alunos possuem das estratégias de aprendizagem sócio-afetivas que empregam nas aulas de Inglês língua estrangeira e as estratégias sócio-afetivas identificadas em situação de interação didática, em contexto de aprendizagem colaborativa. No que respeita à componente empírica do nosso projecto, norteámo-nos pelo paradigma qualitativo, no contexto do qual efetuámos um estudo de caso, a partir de uma abordagem tendencialmente etnográfica, por tal nos parecer mais consentâneo, quer com a nossa problemática, quer com a natureza complexa dos processos interativos em sala de aula. A metodologia quantitativa está igualmente presente, pretendendo-se que tenha adicionado mais dimensionalidade à investigação, contribuindo para a triangulação dos resultados. A investigação, que se desenvolveu ao longo de 18 semanas, teve a sala de aula como local privilegiado para obter grande parte da informação. Os participantes do estudo de caso foram 24 alunos do primeiro ano de uma turma de Inglês Língua Estrangeira de um Instituto Politécnico, sendo a investigadora a docente da disciplina. A informação proveio primordialmente de um corpus de interações didáticas colaborativas audiogravadas e posteriormente transcritas, constituído por 8 sessões com uma duração aproximada de uma hora, e das respostas a um inquérito por questionário − construído a partir da taxonomia de Oxford (1990) − relativo à dimensão sócio-afetiva das estratégias de aprendizagem do Inglês língua estrangeira. O corpus gravado e transcrito foi analisado através da categorização por indicadores, com o objetivo de se detetarem as marcas sócio-afetivas das estratégias de aprendizagem mobilizadas pelos alunos. As respostas ao questionário foram tratadas quantitativamente numa primeira fase, e os resultados foram posteriormente triangulados com os provenientes da análise do corpus de interações. Este estudo permitiu: i) elencar as estratégias de aprendizagem que os aprendentes referem utilizar em situação de aprendizagem colaborativa e não colaborativa, ii) detetar quais destas estratégias são efetivamente utilizadas na aprendizagem colaborativa, iii) e concluir que existe, na maioria dos casos, um desfasamento entre o autoconceito do aluno relativamente ao seu perfil de aprendente de línguas estrangeiras, mais concretamente às dimensões afetiva e social das estratégias de aprendizagem que mobiliza, e a forma como este aprendente recorre a estas mesma estratégias na sala de aula. Concluímos igualmente que, em termos globais, existem diferenças, por vezes significativas, entre as representações que os sujeitos possuem da aprendizagem colaborativa e aquelas que detêm acerca da aprendizagem não colaborativa.
Resumo:
A variety of physical and biomedical imaging techniques, such as digital holography, interferometric synthetic aperture radar (InSAR), or magnetic resonance imaging (MRI) enable measurement of the phase of a physical quantity additionally to its amplitude. However, the phase can commonly only be measured modulo 2π, as a so called wrapped phase map. Phase unwrapping is the process of obtaining the underlying physical phase map from the wrapped phase. Tile-based phase unwrapping algorithms operate by first tessellating the phase map, then unwrapping individual tiles, and finally merging them to a continuous phase map. They can be implemented computationally efficiently and are robust to noise. However, they are prone to failure in the presence of phase residues or erroneous unwraps of single tiles. We tried to overcome these shortcomings by creating novel tile unwrapping and merging algorithms as well as creating a framework that allows to combine them in modular fashion. To increase the robustness of the tile unwrapping step, we implemented a model-based algorithm that makes efficient use of linear algebra to unwrap individual tiles. Furthermore, we adapted an established pixel-based unwrapping algorithm to create a quality guided tile merger. These original algorithms as well as previously existing ones were implemented in a modular phase unwrapping C++ framework. By examining different combinations of unwrapping and merging algorithms we compared our method to existing approaches. We could show that the appropriate choice of unwrapping and merging algorithms can significantly improve the unwrapped result in the presence of phase residues and noise. Beyond that, our modular framework allows for efficient design and test of new tile-based phase unwrapping algorithms. The software developed in this study is freely available.
Resumo:
SQL Injection Attack (SQLIA) remains a technique used by a computer network intruder to pilfer an organisation’s confidential data. This is done by an intruder re-crafting web form’s input and query strings used in web requests with malicious intent to compromise the security of an organisation’s confidential data stored at the back-end database. The database is the most valuable data source, and thus, intruders are unrelenting in constantly evolving new techniques to bypass the signature’s solutions currently provided in Web Application Firewalls (WAF) to mitigate SQLIA. There is therefore a need for an automated scalable methodology in the pre-processing of SQLIA features fit for a supervised learning model. However, obtaining a ready-made scalable dataset that is feature engineered with numerical attributes dataset items to train Artificial Neural Network (ANN) and Machine Leaning (ML) models is a known issue in applying artificial intelligence to effectively address ever evolving novel SQLIA signatures. This proposed approach applies numerical attributes encoding ontology to encode features (both legitimate web requests and SQLIA) to numerical data items as to extract scalable dataset for input to a supervised learning model in moving towards a ML SQLIA detection and prevention model. In numerical attributes encoding of features, the proposed model explores a hybrid of static and dynamic pattern matching by implementing a Non-Deterministic Finite Automaton (NFA). This combined with proxy and SQL parser Application Programming Interface (API) to intercept and parse web requests in transition to the back-end database. In developing a solution to address SQLIA, this model allows processed web requests at the proxy deemed to contain injected query string to be excluded from reaching the target back-end database. This paper is intended for evaluating the performance metrics of a dataset obtained by numerical encoding of features ontology in Microsoft Azure Machine Learning (MAML) studio using Two-Class Support Vector Machines (TCSVM) binary classifier. This methodology then forms the subject of the empirical evaluation.
Resumo:
Recent years have seen an astronomical rise in SQL Injection Attacks (SQLIAs) used to compromise the confidentiality, authentication and integrity of organisations’ databases. Intruders becoming smarter in obfuscating web requests to evade detection combined with increasing volumes of web traffic from the Internet of Things (IoT), cloud-hosted and on-premise business applications have made it evident that the existing approaches of mostly static signature lack the ability to cope with novel signatures. A SQLIA detection and prevention solution can be achieved through exploring an alternative bio-inspired supervised learning approach that uses input of labelled dataset of numerical attributes in classifying true positives and negatives. We present in this paper a Numerical Encoding to Tame SQLIA (NETSQLIA) that implements a proof of concept for scalable numerical encoding of features to a dataset attributes with labelled class obtained from deep web traffic analysis. In the numerical attributes encoding: the model leverages proxy in the interception and decryption of web traffic. The intercepted web requests are then assembled for front-end SQL parsing and pattern matching by applying traditional Non-Deterministic Finite Automaton (NFA). This paper is intended for a technique of numerical attributes extraction of any size primed as an input dataset to an Artificial Neural Network (ANN) and statistical Machine Learning (ML) algorithms implemented using Two-Class Averaged Perceptron (TCAP) and Two-Class Logistic Regression (TCLR) respectively. This methodology then forms the subject of the empirical evaluation of the suitability of this model in the accurate classification of both legitimate web requests and SQLIA payloads.
Resumo:
The social landscape is filled with an intricate web of species-specific desired objects and course of actions. Humans are highly social animals and, as they navigate this landscape, they need to produce adapted decision-making behaviour. Traditionally social and non-social neural mechanisms affecting choice have been investigated using different approaches. Recently, in an effort to unite these findings, two main theories have been proposed to explain how the brain might encode social and non-social motivational decision-making: the extended common currency and the social valuation specific schema (Ruff & Fehr 2014). One way to test these theories is to directly compare neural activity related to social and non-social decision outcomes within the same experimental setting. Here we address this issue by focusing on the neural substrates of social and non-social forms of uncertainty. Using functional magnetic resonance imaging (fMRI) we directly compared the neural representations of reward and risk prediction and errors (RePE and RiPE) in social and non- social situations using gambling games. We used a trust betting game to vary uncertainty along a social dimension (trustworthiness), and a card game (Preuschoff et al. 2006) to vary uncertainty along a non-social dimension (pure risk). The trust game was designed to maintain the same structure of the card game. In a first study, we exposed a divide between subcortical and cortical regions when comparing the way these regions process social and non-social forms of uncertainty during outcome anticipation. Activity in subcortical regions reflected social and non-social RePE, while activity in cortical regions correlated with social RePE and non-social RiPE. The second study focused on outcome delivery and integrated the concept of RiPE in non-social settings with that of fairness and monetary utility maximisation in social settings. In particular these results corroborate recent models of anterior insula function (Singer et al. 2009; Seth 2013), and expose a possible neural mechanism that weights fairness and uncertainty but not monetary utility. The third study focused on functionally defined regions of the early visual cortex (V1) showing how activity in these areas, traditionally considered only visual, might reflect motivational prediction errors in addition to known perceptual prediction mechanisms (den Ouden et al 2012). On the whole, while our results do not support unilaterally one or the other theory modeling the underlying neural dynamics of social and non-social forms of decision making, they provide a working framework where both general mechanisms might coexist.
Resumo:
Nigerian scam, also known as advance fee fraud or 419 scam, is a prevalent form of online fraudulent activity that causes financial loss to individuals and businesses. Nigerian scam has evolved from simple non-targeted email messages to more sophisticated scams targeted at users of classifieds, dating and other websites. Even though such scams are observed and reported by users frequently, the community’s understanding of Nigerian scams is limited since the scammers operate “underground”. To better understand the underground Nigerian scam ecosystem and seek effective methods to deter Nigerian scam and cybercrime in general, we conduct a series of active and passive measurement studies. Relying upon the analysis and insight gained from the measurement studies, we make four contributions: (1) we analyze the taxonomy of Nigerian scam and derive long-term trends in scams; (2) we provide an insight on Nigerian scam and cybercrime ecosystems and their underground operation; (3) we propose a payment intervention as a potential deterrent to cybercrime operation in general and evaluate its effectiveness; and (4) we offer active and passive measurement tools and techniques that enable in-depth analysis of cybercrime ecosystems and deterrence on them. We first created and analyze a repository of more than two hundred thousand user-reported scam emails, stretching from 2006 to 2014, from four major scam reporting websites. We select ten most commonly observed scam categories and tag 2,000 scam emails randomly selected from our repository. Based upon the manually tagged dataset, we train a machine learning classifier and cluster all scam emails in the repository. From the clustering result, we find a strong and sustained upward trend for targeted scams and downward trend for non-targeted scams. We then focus on two types of targeted scams: sales scams and rental scams targeted users on Craigslist. We built an automated scam data collection system and gathered large-scale sales scam emails. Using the system we posted honeypot ads on Craigslist and conversed automatically with the scammers. Through the email conversation, the system obtained additional confirmation of likely scam activities and collected additional information such as IP addresses and shipping addresses. Our analysis revealed that around 10 groups were responsible for nearly half of the over 13,000 total scam attempts we received. These groups used IP addresses and shipping addresses in both Nigeria and the U.S. We also crawled rental ads on Craigslist, identified rental scam ads amongst the large number of benign ads and conversed with the potential scammers. Through in-depth analysis of the rental scams, we found seven major scam campaigns employing various operations and monetization methods. We also found that unlike sales scammers, most rental scammers were in the U.S. The large-scale scam data and in-depth analysis provide useful insights on how to design effective deterrence techniques against cybercrime in general. We study underground DDoS-for-hire services, also known as booters, and measure the effectiveness of undermining a payment system of DDoS Services. Our analysis shows that the payment intervention can have the desired effect of limiting cybercriminals’ ability and increasing the risk of accepting payments.
Resumo:
Humans have a high ability to extract visual data information acquired by sight. Trought a learning process, which starts at birth and continues throughout life, image interpretation becomes almost instinctively. At a glance, one can easily describe a scene with reasonable precision, naming its main components. Usually, this is done by extracting low-level features such as edges, shapes and textures, and associanting them to high level meanings. In this way, a semantic description of the scene is done. An example of this, is the human capacity to recognize and describe other people physical and behavioral characteristics, or biometrics. Soft-biometrics also represents inherent characteristics of human body and behaviour, but do not allow unique person identification. Computer vision area aims to develop methods capable of performing visual interpretation with performance similar to humans. This thesis aims to propose computer vison methods which allows high level information extraction from images in the form of soft biometrics. This problem is approached in two ways, unsupervised and supervised learning methods. The first seeks to group images via an automatic feature extraction learning , using both convolution techniques, evolutionary computing and clustering. In this approach employed images contains faces and people. Second approach employs convolutional neural networks, which have the ability to operate on raw images, learning both feature extraction and classification processes. Here, images are classified according to gender and clothes, divided into upper and lower parts of human body. First approach, when tested with different image datasets obtained an accuracy of approximately 80% for faces and non-faces and 70% for people and non-person. The second tested using images and videos, obtained an accuracy of about 70% for gender, 80% to the upper clothes and 90% to lower clothes. The results of these case studies, show that proposed methods are promising, allowing the realization of automatic high level information image annotation. This opens possibilities for development of applications in diverse areas such as content-based image and video search and automatica video survaillance, reducing human effort in the task of manual annotation and monitoring.
Resumo:
Este trabajo se enfoca en la implementación de un detector de arrecife de coral de desempeño rápido que se utiliza para un vehículo autónomo submarino (Autonomous Underwater Vehicle, AUV, por sus siglas en inglés). Una detección rápida de la presencia de coral asegura la estabilización del AUV frente al arrecife en el menor tiempo posible, evitando colisiones con el coral. La detección de coral se hace en una imagen que captura la escena que percibe la cámara del AUV. Se realiza una clasificación píxel por píxel entre dos clases: arrecife de coral y el plano de fondo que no es coral. A cada píxel de la imagen se le asigna un vector característico, el mismo que se genera mediante el uso de filtros Gabor Wavelets. Éstos son implementados en C++ y la librería OpenCV. Los vectores característicos son clasificados a través de nueve algoritmos de máquinas de aprendizaje. El desempeño de cada algoritmo se compara mediante la precisión y el tiempo de ejecución. El algoritmo de Árboles de Decisión resultó ser el más rápido y preciso de entre todos los algoritmos. Se creó una base de datos de 621 imágenes de corales de Belice (110 imágenes de entrenamiento y 511 imágenes de prueba).
Resumo:
L’objectif de la présente thèse est de générer des connaissances sur les contributions possibles d’une formation continue à l’évolution des perspectives et pratiques des professionnels de la santé buccodentaire. Prônant une approche centrée sur le patient, la formation vise à sensibiliser les professionnels à la pauvreté et à encourager des pratiques qui se veulent inclusives et qui tiennent compte du contexte social des patients. L’évaluation de la formation s’inscrit dans le contexte d’une recherche-action participative de développement d’outils éducatifs et de transfert des connaissances sur la pauvreté. Cette recherche-action aspire à contribuer à la lutte contre les iniquités sociales de santé et d’accès aux soins au Québec; elle reflète une préoccupation pour une plus grande justice sociale ainsi qu’une prise de position pour une santé publique critique fondée sur une « science des solutions » (Potvin, 2013). Quatre articles scientifiques, ancrés dans une philosophie constructiviste et dans les concepts et principes de l’apprentissage transformationnel (Mezirow, 1991), constituent le cœur de cette thèse. Le premier article présente une revue critique de la littérature portant sur l’enseignement de l’approche de soins centrés sur le patient. Prenant appui sur le concept d’une « épistémologie partagée », des principes éducatifs porteurs d’une transformation de perspective à l’égard de la relation professionnel-patient ont été identifiés et analysés. Le deuxième article de thèse s’inscrit dans le cadre du développement participatif d’outils de formation sur la pauvreté et illustre le processus de co-construction d’un scénario de court-métrage social réaliste portant sur la pauvreté et l’accès aux soins. L’article décrit et apporte une réflexion, notamment sur la dimension de co-formation entre les différents acteurs des milieux académique, professionnel et citoyen qui ont constitué le collectif À l’écoute les uns des autres. Nous y découvrons la force du croisement des savoirs pour générer des prises de conscience sur soi et sur ses préjugés. Les outils développés par le collectif ont été intégrés à une formation continue axée sur la réflexion critique et l’apprentissage transformationnel, et conçue pour être livrée en cabinet dentaire privé. Les deux derniers articles de thèse présentent les résultats d’une étude de cas instrumentale évaluative centrée sur cette formation continue et visant donc à répondre à l’objectif premier de cette thèse. Le premier consiste en une analyse des transformations de perspectives et d’action au sein d’une équipe de 15 professionnels dentaires ayant participé à la formation continue sur une période de trois mois. L’article décrit, entre autres, une plus grande ouverture, chez certains participants, sur les causes structurelles de la pauvreté et une plus grande sensibilité au vécu au quotidien des personnes prestataires de l’aide sociale. L’article comprend également une exploration des effets paradoxaux dans l’apprentissage, notamment le renforcement, chez certains, de perceptions négatives à l’égard des personnes prestataires de l’aide sociale. Le quatrième article fait état de barrières idéologiques contraignant la transformation des pratiques professionnelles : 1) l’identification à l’idéologie du marché privé comme véhicule d’organisation des soins; 2) l’attachement au concept d’égalité dans les pratiques, au détriment de l’équité; 3) la prédominance du modèle biomédical, contraignant l’adoption de pratiques centrées sur la personne et 4) la catégorisation sociale des personnes prestataires de l’aide sociale. L’analyse des perceptions, mais aussi de l’expérience vécue de ces barrières démontre comment des facteurs systémiques et sociaux influent sur le rapport entre professionnel dentaire et personne prestataire de l’aide sociale. Les conséquences pour la recherche, l’éducation dentaire, le transfert des connaissances, ainsi que pour la régulation professionnelle et les politiques de santé buccodentaire, sont examinées à partir de cette perspective.
Resumo:
Dissertação (mestrado)—Universidade de Brasília, Instituto de Física, Programa de Pós-Graduação em Física, 2015.
Resumo:
Dissertação de Mestrado apresentada ao Instituto Superior de Psicologia Aplicada para obtenção de grau de Mestre na especialidade de Psicologia Social e das Organizações.
Resumo:
Les métaheuristiques sont très utilisées dans le domaine de l'optimisation discrète. Elles permettent d’obtenir une solution de bonne qualité en un temps raisonnable, pour des problèmes qui sont de grande taille, complexes, et difficiles à résoudre. Souvent, les métaheuristiques ont beaucoup de paramètres que l’utilisateur doit ajuster manuellement pour un problème donné. L'objectif d'une métaheuristique adaptative est de permettre l'ajustement automatique de certains paramètres par la méthode, en se basant sur l’instance à résoudre. La métaheuristique adaptative, en utilisant les connaissances préalables dans la compréhension du problème, des notions de l'apprentissage machine et des domaines associés, crée une méthode plus générale et automatique pour résoudre des problèmes. L’optimisation globale des complexes miniers vise à établir les mouvements des matériaux dans les mines et les flux de traitement afin de maximiser la valeur économique du système. Souvent, en raison du grand nombre de variables entières dans le modèle, de la présence de contraintes complexes et de contraintes non-linéaires, il devient prohibitif de résoudre ces modèles en utilisant les optimiseurs disponibles dans l’industrie. Par conséquent, les métaheuristiques sont souvent utilisées pour l’optimisation de complexes miniers. Ce mémoire améliore un procédé de recuit simulé développé par Goodfellow & Dimitrakopoulos (2016) pour l’optimisation stochastique des complexes miniers stochastiques. La méthode développée par les auteurs nécessite beaucoup de paramètres pour fonctionner. Un de ceux-ci est de savoir comment la méthode de recuit simulé cherche dans le voisinage local de solutions. Ce mémoire implémente une méthode adaptative de recherche dans le voisinage pour améliorer la qualité d'une solution. Les résultats numériques montrent une augmentation jusqu'à 10% de la valeur de la fonction économique.
Resumo:
Les métaheuristiques sont très utilisées dans le domaine de l'optimisation discrète. Elles permettent d’obtenir une solution de bonne qualité en un temps raisonnable, pour des problèmes qui sont de grande taille, complexes, et difficiles à résoudre. Souvent, les métaheuristiques ont beaucoup de paramètres que l’utilisateur doit ajuster manuellement pour un problème donné. L'objectif d'une métaheuristique adaptative est de permettre l'ajustement automatique de certains paramètres par la méthode, en se basant sur l’instance à résoudre. La métaheuristique adaptative, en utilisant les connaissances préalables dans la compréhension du problème, des notions de l'apprentissage machine et des domaines associés, crée une méthode plus générale et automatique pour résoudre des problèmes. L’optimisation globale des complexes miniers vise à établir les mouvements des matériaux dans les mines et les flux de traitement afin de maximiser la valeur économique du système. Souvent, en raison du grand nombre de variables entières dans le modèle, de la présence de contraintes complexes et de contraintes non-linéaires, il devient prohibitif de résoudre ces modèles en utilisant les optimiseurs disponibles dans l’industrie. Par conséquent, les métaheuristiques sont souvent utilisées pour l’optimisation de complexes miniers. Ce mémoire améliore un procédé de recuit simulé développé par Goodfellow & Dimitrakopoulos (2016) pour l’optimisation stochastique des complexes miniers stochastiques. La méthode développée par les auteurs nécessite beaucoup de paramètres pour fonctionner. Un de ceux-ci est de savoir comment la méthode de recuit simulé cherche dans le voisinage local de solutions. Ce mémoire implémente une méthode adaptative de recherche dans le voisinage pour améliorer la qualité d'une solution. Les résultats numériques montrent une augmentation jusqu'à 10% de la valeur de la fonction économique.
Resumo:
Fully articulated hand tracking promises to enable fundamentally new interactions with virtual and augmented worlds, but the limited accuracy and efficiency of current systems has prevented widespread adoption. Today's dominant paradigm uses machine learning for initialization and recovery followed by iterative model-fitting optimization to achieve a detailed pose fit. We follow this paradigm, but make several changes to the model-fitting, namely using: (1) a more discriminative objective function; (2) a smooth-surface model that provides gradients for non-linear optimization; and (3) joint optimization over both the model pose and the correspondences between observed data points and the model surface. While each of these changes may actually increase the cost per fitting iteration, we find a compensating decrease in the number of iterations. Further, the wide basin of convergence means that fewer starting points are needed for successful model fitting. Our system runs in real-time on CPU only, which frees up the commonly over-burdened GPU for experience designers. The hand tracker is efficient enough to run on low-power devices such as tablets. We can track up to several meters from the camera to provide a large working volume for interaction, even using the noisy data from current-generation depth cameras. Quantitative assessments on standard datasets show that the new approach exceeds the state of the art in accuracy. Qualitative results take the form of live recordings of a range of interactive experiences enabled by this new approach.