967 resultados para Capture-recapture Data
Resumo:
The tongue is the most important and dynamic articulator for speech formation, because of its anatomic aspects (particularly, the large volume of this muscular organ comparatively to the surrounding organs of the vocal tract) and also due to the wide range of movements and flexibility that are involved. In speech communication research, a variety of techniques have been used for measuring the three-dimensional vocal tract shapes. More recently, magnetic resonance imaging (MRI) becomes common; mainly, because this technique allows the collection of a set of static and dynamic images that can represent the entire vocal tract along any orientation. Over the years, different anatomical organs of the vocal tract have been modelled; namely, 2D and 3D tongue models, using parametric or statistical modelling procedures. Our aims are to present and describe some 3D reconstructed models from MRI data, for one subject uttering sustained articulations of some typical Portuguese sounds. Thus, we present a 3D database of the tongue obtained by stack combinations with the subject articulating Portuguese vowels. This 3D knowledge of the speech organs could be very important; especially, for clinical purposes (for example, for the assessment of articulatory impairments followed by tongue surgery in speech rehabilitation), and also for a better understanding of acoustic theory in speech formation.
Resumo:
The hand is one of the most important instruments of the human body, mainly due to the possibility of grip movements. Grip strength has been described as an important predictor of functional capacity. There are several factors that may influence it, such as gender, age and anthropometric characteristics. Functional capacity refers to the ability to perform daily activities which allow the individual to self-care and to live with autonomy. Composite Physical Function (CPF) scale is an evaluation tool for functional capacity that includes daily activities, self-care, sports activities, upper limb function and gait capacity. In 2011, Portugal had 15% of young population (0-14years) and 19% of elderly population (over 65 years). Considering the double-ageing phenomen, it is important to understand the effect of the grip strength in elderly individuals, considering their characteristics, as the need to maintainin dependency as long as possible.
Resumo:
A estimativa da idade gestacional (IG) em restos cadavéricos fetais é importante em contextos forenses. Para esse efeito, os especialistas forenses recorrem à avaliação do padrão de calcificação dentária e/ou ao estudo do esqueleto. Neste último, o comprimento das diáfises de ossos longos é um dos métodos mais utilizados, sendo utilizadas equações de regressão de obras pouco atuais ou baseadas em dados ecográficos, cujas medições diferem das efetuadas diretamente no osso. Este trabalho tem como objetivo principal a obtenção de equações de regressão para a população Portuguesa, com base na medição das diáfises de fémur, tíbia e úmero, utilizando radiografias postmortem. A amostra é constituída por 80 fetos de IG conhecida. Tratando-se de um estudo retrospectivo, os casos foram selecionados com base nas informações clínicas e anatomopatológicas, excluindo-se aqueles cujo normal crescimento se encontrava efetiva ou potencialmente comprometido. Os resultados confirmaram uma forte correlação entre o comprimento das diáfises estudadas e a IG, apresentando o fémur a correlação mais forte (r=0.967; p <0,01). Assim, foi possível obter uma equação de regressão para cada um dos ossos estudados. Concluindo, os objetivos do estudo foram atingidos com a obtenção das equações de regressão para os ossos estudados. Pretende-se, futuramente, alargar a amostra para validar e consolidar os resultados obtidos neste estudo.
Resumo:
O principal objectivo da animação de personagens virtuais é o de contar uma história através da utilização de personagens virtuais emocionalmente expressivos. Os personagens têm personalidades distintas, e transmitem as suas emoções e processos de pensamento através dos seus comportamentos (comunicação não verbal). As suas acções muitas das vezes constituem a geração de movimentos corporais complexos. Existem diversas questões a considerar quando se anima uma entidade complexa, tais como, a posição das zonas móveis e as suas velocidades. Os personagens virtuais são um exemplo de entidades complexas e estão entre os elementos mais utilizados em animação computacional. O foco desta dissertação consistiu na criação de uma proposta de sistema de animação de personagens virtuais, cujos movimentos e expressões faciais são capazes de transmitir emoções e estados de espírito. Os movimentos primários, ou seja os movimentos que definem o comportamento dos personagens, são provenientes da captura de movimentos humanos (Motion Capture). As animações secundárias, tais como as expressões faciais, são criadas em Autodesk Maya recorrendo à técnica BlendShapes. Os dados obtidos pela captura de movimentos, são organizados numa biblioteca de comportamentos através de um grafo de movimentos, conhecido por Move Tree. Esta estrutura permite o controlo em tempo real dos personagens através da gestão do estado dos personagens. O sistema possibilita também a transição eficaz entre movimentos semelhantes e entre diferentes velocidades de locomoção, minimizando o efeito de arrastamento de pés conhecido como footskate. Torna-se assim possível definir um trajecto que o personagem poderá seguir com movimentos suaves. Estão também disponíveis os resultados obtidos nas sessões de avaliação realizadas, que visaram a determinação da qualidade das transições entre animações. Propõem-se ainda o melhoramento do sistema através da implementação da construção automática do grafo de movimentos.
Resumo:
Seismic data is difficult to analyze and classical mathematical tools reveal strong limitations in exposing hidden relationships between earthquakes. In this paper, we study earthquake phenomena in the perspective of complex systems. Global seismic data, covering the period from 1962 up to 2011 is analyzed. The events, characterized by their magnitude, geographic location and time of occurrence, are divided into groups, either according to the Flinn-Engdahl (F-E) seismic regions of Earth or using a rectangular grid based in latitude and longitude coordinates. Two methods of analysis are considered and compared in this study. In a first method, the distributions of magnitudes are approximated by Gutenberg-Richter (G-R) distributions and the parameters used to reveal the relationships among regions. In the second method, the mutual information is calculated and adopted as a measure of similarity between regions. In both cases, using clustering analysis, visualization maps are generated, providing an intuitive and useful representation of the complex relationships that are present among seismic data. Such relationships might not be perceived on classical geographic maps. Therefore, the generated charts are a valid alternative to other visualization tools, for understanding the global behavior of earthquakes.
Resumo:
Research on the problem of feature selection for clustering continues to develop. This is a challenging task, mainly due to the absence of class labels to guide the search for relevant features. Categorical feature selection for clustering has rarely been addressed in the literature, with most of the proposed approaches having focused on numerical data. In this work, we propose an approach to simultaneously cluster categorical data and select a subset of relevant features. Our approach is based on a modification of a finite mixture model (of multinomial distributions), where a set of latent variables indicate the relevance of each feature. To estimate the model parameters, we implement a variant of the expectation-maximization algorithm that simultaneously selects the subset of relevant features, using a minimum message length criterion. The proposed approach compares favourably with two baseline methods: a filter based on an entropy measure and a wrapper based on mutual information. The results obtained on synthetic data illustrate the ability of the proposed expectation-maximization method to recover ground truth. An application to real data, referred to official statistics, shows its usefulness.
Resumo:
Research on cluster analysis for categorical data continues to develop, new clustering algorithms being proposed. However, in this context, the determination of the number of clusters is rarely addressed. We propose a new approach in which clustering and the estimation of the number of clusters is done simultaneously for categorical data. We assume that the data originate from a finite mixture of multinomial distributions and use a minimum message length criterion (MML) to select the number of clusters (Wallace and Bolton, 1986). For this purpose, we implement an EM-type algorithm (Silvestre et al., 2008) based on the (Figueiredo and Jain, 2002) approach. The novelty of the approach rests on the integration of the model estimation and selection of the number of clusters in a single algorithm, rather than selecting this number based on a set of pre-estimated candidate models. The performance of our approach is compared with the use of Bayesian Information Criterion (BIC) (Schwarz, 1978) and Integrated Completed Likelihood (ICL) (Biernacki et al., 2000) using synthetic data. The obtained results illustrate the capacity of the proposed algorithm to attain the true number of cluster while outperforming BIC and ICL since it is faster, which is especially relevant when dealing with large data sets.
Resumo:
Cluster analysis for categorical data has been an active area of research. A well-known problem in this area is the determination of the number of clusters, which is unknown and must be inferred from the data. In order to estimate the number of clusters, one often resorts to information criteria, such as BIC (Bayesian information criterion), MML (minimum message length, proposed by Wallace and Boulton, 1968), and ICL (integrated classification likelihood). In this work, we adopt the approach developed by Figueiredo and Jain (2002) for clustering continuous data. They use an MML criterion to select the number of clusters and a variant of the EM algorithm to estimate the model parameters. This EM variant seamlessly integrates model estimation and selection in a single algorithm. For clustering categorical data, we assume a finite mixture of multinomial distributions and implement a new EM algorithm, following a previous version (Silvestre et al., 2008). Results obtained with synthetic datasets are encouraging. The main advantage of the proposed approach, when compared to the above referred criteria, is the speed of execution, which is especially relevant when dealing with large data sets.
Resumo:
Consider the problem of disseminating data from an arbitrary source node to all other nodes in a distributed computer system, like Wireless Sensor Networks (WSNs). We assume that wireless broadcast is used and nodes do not know the topology. We propose new protocols which disseminate data faster and use fewer broadcasts than the simple broadcast protocol.
Resumo:
OBJECTIVE: To review the use of accelerometry as an objective measure of physical activity in adults and elderly people. METHODS: A systematic review of studies on the use of accelerometty as an objective measure to assess physical activity in adults were examined in PubMed Central, Web of Knowledge, EBSCO and Medline databases from March 29 to April 15, 2010. The following keywords were used: "accelerometry," "accelerometer," "physical activity," "PA," "patterns," "levels," "adults," "older adults," and "elderly," either alone or in combination using "AND" or "OR." The reference lists of the articles retrieved were examined to capture any other potentially relevant article. Of 899 studies initially identified, only 18 were fully reviewed, and their outcome measures abstracted and analyzed. RESULTS: Eleven studies were conducted in North America (United States), five in Europe, one in Africa (Cameroon) and one in Australia. Very few enrolled older people, and only one study reported the season or time of year when data was collected. The articles selected had different methods, analyses, and results, which prevented comparison between studies. CONCLUSIONS: There is a need to standardize study methods for data reporting to allow comparisons of results across studies and monitor changes in populations. These data can help design more adequate strategies for monitoring and promotion of physical activity.
Resumo:
Nowadays, due to the incredible grow of the mobile devices market, when we want to implement a client-server applications we must consider mobile devices limitations. In this paper we discuss which can be the more reliable and fast way to exchange information between a server and an Android mobile application. This is an important issue because with a responsive application the user experience is more enjoyable. In this paper we present a study that test and evaluate two data transfer protocols, socket and HTTP, and three data serialization formats (XML, JSON and Protocol Buffers) using different environments and mobile devices to realize which is the most practical and fast to use.
Resumo:
Dissertação de Natureza Científica para obtenção do grau de Mestre em Engenharia Civil na Área de Especialização de Edificações
Resumo:
The goal of the this paper is to show that the DGPS data Internet service we designed and developed provides campus-wide real time access to Differential GPS (DGPS) data and, thus, supports precise outdoor navigation. First we describe the developed distributed system in terms of architecture (a three tier client/server application), services provided (real time DGPS data transportation from remote DGPS sources and campus wide data dissemination) and transmission modes implemented (raw and frame mode over TCP and UDP). Then we present and discuss the results obtained and, finally, we draw some conclusions.
Resumo:
The principal topic of this work is the application of data mining techniques, in particular of machine learning, to the discovery of knowledge in a protein database. In the first chapter a general background is presented. Namely, in section 1.1 we overview the methodology of a Data Mining project and its main algorithms. In section 1.2 an introduction to the proteins and its supporting file formats is outlined. This chapter is concluded with section 1.3 which defines that main problem we pretend to address with this work: determine if an amino acid is exposed or buried in a protein, in a discrete way (i.e.: not continuous), for five exposition levels: 2%, 10%, 20%, 25% and 30%. In the second chapter, following closely the CRISP-DM methodology, whole the process of construction the database that supported this work is presented. Namely, it is described the process of loading data from the Protein Data Bank, DSSP and SCOP. Then an initial data exploration is performed and a simple prediction model (baseline) of the relative solvent accessibility of an amino acid is introduced. It is also introduced the Data Mining Table Creator, a program developed to produce the data mining tables required for this problem. In the third chapter the results obtained are analyzed with statistical significance tests. Initially the several used classifiers (Neural Networks, C5.0, CART and Chaid) are compared and it is concluded that C5.0 is the most suitable for the problem at stake. It is also compared the influence of parameters like the amino acid information level, the amino acid window size and the SCOP class type in the accuracy of the predictive models. The fourth chapter starts with a brief revision of the literature about amino acid relative solvent accessibility. Then, we overview the main results achieved and finally discuss about possible future work. The fifth and last chapter consists of appendices. Appendix A has the schema of the database that supported this thesis. Appendix B has a set of tables with additional information. Appendix C describes the software provided in the DVD accompanying this thesis that allows the reconstruction of the present work.
Resumo:
Sensor/actuator networks promised to extend automated monitoring and control into industrial processes. Avionic system is one of the prominent technologies that can highly gain from dense sensor/actuator deployments. An aircraft with smart sensing skin would fulfill the vision of affordability and environmental friendliness properties by reducing the fuel consumption. Achieving these properties is possible by providing an approximate representation of the air flow across the body of the aircraft and suppressing the detected aerodynamic drags. To the best of our knowledge, getting an accurate representation of the physical entity is one of the most significant challenges that still exists with dense sensor/actuator network. This paper offers an efficient way to acquire sensor readings from very large sensor/actuator network that are located in a small area (dense network). It presents LIA algorithm, a Linear Interpolation Algorithm that provides two important contributions. First, it demonstrates the effectiveness of employing a transformation matrix to mimic the environmental behavior. Second, it renders a smart solution for updating the previously defined matrix through a procedure called learning phase. Simulation results reveal that the average relative error in LIA algorithm can be reduced by as much as 60% by exploiting transformation matrix.