976 resultados para Technical Report, Computer Science
Resumo:
TROST. S. G., R. R. PATE, J. F. SALLIS, P. S. FREEDSON, W. C. TAYLOR, M. DOWDA, and J. SIRARD. Age and gender differences in objectively measured physical activity in youth. Med. Sci. Sports Ererc., Vol. 34, No. 2, pp. 350-355, 2002. Purpose: The purpose of this study was to evaluate age and gender differences in objectively measured physical activity (PA) in a population-based sample of students in grades 1-12. Methods: Participants (185 male, 190 female) wore a CSA 7164 accelerometer for 7 consecutive days. To examine age-related trends. students were grouped as follows: grades 1-3 (N = 90), grades 4-6 (N = 91), grades 7-9 (N = 96). and grades 10-12 (N = 92). Bouts of PA and minutes spent in moderate-to-vigorous PA (MVPA) and vigorous PA (VPA) were examined. Results: Daily MVPA and VPA exhibited a significant inverse relationship with grade level, with the largest differences occurring between grades 1d-3 and 4-6. Boys were more active than girls; however, for overall PA, the magnitudes of the gender differences were modest. Participation in continuous 20-min bouts of PA was low to nonexistent. Conclusion: Our results support the notion that PA declines rapidly during childhood and adolescence and that accelerometers are feasible alternatives to self-report methods in moderately sized population-level surveillance studies.
Resumo:
An increasing amount of research is being developed in the area where technology and humans meet. The success or failure of technologies and the question whether technology helps humans to fulfill their goals or whether it hinders them is in most cases not a technical one. User Perception and Influencing Factors of Technology in Everyday Life addresses issues of human and technology interaction. The research in this work is interdisciplinary, ranging from more technical subjects such as computer science, engineering, and information systems, to non-technical descriptions of technology and human interaction from the point of view of sociology or philosophy. This book is perfect for academics, researchers, and professionals alike as it presents a set of theories that allow us to understand the interaction of technology and humans and to put it to practical use.
Resumo:
The way humans interact with technology is undergoing a tremendous change. It is hard to imagine the lives we live today without the benefits of technology that we take for granted. Applying research in computer science, engineering, and information systems to non-technical descriptions of technology, such as human interaction, has shaped and continues to shape our lives. Human Interaction with Technology for Working, Communicating, and Learning: Advancements provides a framework for conceptual, theoretical, and applied research in regards to the relationship between technology and humans. This book is unique in the sense that it does not only cover technology, but also science, research, and the relationship between these fields and individuals' experience. This book is a must have for anyone interested in this research area, as it provides a voice for all users and a look into our future.
Resumo:
When exploring a virtual environment, realism depends mainly on two factors: realistic images and real-time feedback (motions, behaviour etc.). In this context, photo realism and physical validity of computer generated images required by emerging applications, such as advanced e-commerce, still impose major challenges in the area of rendering research whereas the complexity of lighting phenomena further requires powerful and predictable computing if time constraints must be attained. In this technical report we address the state-of-the-art on rendering, trying to put the focus on approaches, techniques and technologies that might enable real-time interactive web-based clientserver rendering systems. The focus is on the end-systems and not the networking technologies used to interconnect client(s) and server(s).
Resumo:
15th International Conference on Mixed Design of Integrated Circuits and Systems, pp. 177 – 180, Poznan, Polónia
Resumo:
Multi-agent architectures are well suited for complex inherently distributed problem solving domains. From the many challenging aspects that arise within this framework, a crucial one emerges: how to incorporate dynamic and conflicting agent beliefs? While the belief revision activity in a single agent scenario is concentrated on incorporating new information while preserving consistency, in a multi-agent system it also has to deal with possible conflicts between the agents perspectives. To provide an adequate framework, each agent, built as a combination of an assumption based belief revision system and a cooperation layer, was enriched with additional features: a distributed search control mechanism allowing dynamic context management, and a set of different distributed consistency methodologies. As a result, a Distributed Belief Revision Testbed (DiBeRT) was developed. This paper is a preliminary report presenting some of DiBeRT contributions: a concise representation of external beliefs; a simple and innovative methodology to achieve distributed context management; and a reduced inter-agent data exchange format.
Resumo:
To meet the increasing demands of the complex inter-organizational processes and the demand for continuous innovation and internationalization, it is evident that new forms of organisation are being adopted, fostering more intensive collaboration processes and sharing of resources, in what can be called collaborative networks (Camarinha-Matos, 2006:03). Information and knowledge are crucial resources in collaborative networks, being their management fundamental processes to optimize. Knowledge organisation and collaboration systems are thus important instruments for the success of collaborative networks of organisations having been researched in the last decade in the areas of computer science, information science, management sciences, terminology and linguistics. Nevertheless, research in this area didn’t give much attention to multilingual contexts of collaboration, which pose specific and challenging problems. It is then clear that access to and representation of knowledge will happen more and more on a multilingual setting which implies the overcoming of difficulties inherent to the presence of multiple languages, through the use of processes like localization of ontologies. Although localization, like other processes that involve multilingualism, is a rather well-developed practice and its methodologies and tools fruitfully employed by the language industry in the development and adaptation of multilingual content, it has not yet been sufficiently explored as an element of support to the development of knowledge representations - in particular ontologies - expressed in more than one language. Multilingual knowledge representation is then an open research area calling for cross-contributions from knowledge engineering, terminology, ontology engineering, cognitive sciences, computational linguistics, natural language processing, and management sciences. This workshop joined researchers interested in multilingual knowledge representation, in a multidisciplinary environment to debate the possibilities of cross-fertilization between knowledge engineering, terminology, ontology engineering, cognitive sciences, computational linguistics, natural language processing, and management sciences applied to contexts where multilingualism continuously creates new and demanding challenges to current knowledge representation methods and techniques. In this workshop six papers dealing with different approaches to multilingual knowledge representation are presented, most of them describing tools, approaches and results obtained in the development of ongoing projects. In the first case, Andrés Domínguez Burgos, Koen Kerremansa and Rita Temmerman present a software module that is part of a workbench for terminological and ontological mining, Termontospider, a wiki crawler that aims at optimally traverse Wikipedia in search of domainspecific texts for extracting terminological and ontological information. The crawler is part of a tool suite for automatically developing multilingual termontological databases, i.e. ontologicallyunderpinned multilingual terminological databases. In this paper the authors describe the basic principles behind the crawler and summarized the research setting in which the tool is currently tested. In the second paper, Fumiko Kano presents a work comparing four feature-based similarity measures derived from cognitive sciences. The purpose of the comparative analysis presented by the author is to verify the potentially most effective model that can be applied for mapping independent ontologies in a culturally influenced domain. For that, datasets based on standardized pre-defined feature dimensions and values, which are obtainable from the UNESCO Institute for Statistics (UIS) have been used for the comparative analysis of the similarity measures. The purpose of the comparison is to verify the similarity measures based on the objectively developed datasets. According to the author the results demonstrate that the Bayesian Model of Generalization provides for the most effective cognitive model for identifying the most similar corresponding concepts existing for a targeted socio-cultural community. In another presentation, Thierry Declerck, Hans-Ulrich Krieger and Dagmar Gromann present an ongoing work and propose an approach to automatic extraction of information from multilingual financial Web resources, to provide candidate terms for building ontology elements or instances of ontology concepts. The authors present a complementary approach to the direct localization/translation of ontology labels, by acquiring terminologies through the access and harvesting of multilingual Web presences of structured information providers in the field of finance, leading to both the detection of candidate terms in various multilingual sources in the financial domain that can be used not only as labels of ontology classes and properties but also for the possible generation of (multilingual) domain ontologies themselves. In the next paper, Manuel Silva, António Lucas Soares and Rute Costa claim that despite the availability of tools, resources and techniques aimed at the construction of ontological artifacts, developing a shared conceptualization of a given reality still raises questions about the principles and methods that support the initial phases of conceptualization. These questions become, according to the authors, more complex when the conceptualization occurs in a multilingual setting. To tackle these issues the authors present a collaborative platform – conceptME - where terminological and knowledge representation processes support domain experts throughout a conceptualization framework, allowing the inclusion of multilingual data as a way to promote knowledge sharing and enhance conceptualization and support a multilingual ontology specification. In another presentation Frieda Steurs and Hendrik J. Kockaert present us TermWise, a large project dealing with legal terminology and phraseology for the Belgian public services, i.e. the translation office of the ministry of justice, a project which aims at developing an advanced tool including expert knowledge in the algorithms that extract specialized language from textual data (legal documents) and whose outcome is a knowledge database including Dutch/French equivalents for legal concepts, enriched with the phraseology related to the terms under discussion. Finally, Deborah Grbac, Luca Losito, Andrea Sada and Paolo Sirito report on the preliminary results of a pilot project currently ongoing at UCSC Central Library, where they propose to adapt to subject librarians, employed in large and multilingual Academic Institutions, the model used by translators working within European Union Institutions. The authors are using User Experience (UX) Analysis in order to provide subject librarians with a visual support, by means of “ontology tables” depicting conceptual linking and connections of words with concepts presented according to their semantic and linguistic meaning. The organizers hope that the selection of papers presented here will be of interest to a broad audience, and will be a starting point for further discussion and cooperation.
Resumo:
O projeto, financiado pelo Programa Aprendizagem ao Longo da Vida, decorreu entre 1 de agosto de 2011 e 31 de julho de 2013 e foi coordenado pelo Instituto de Administração Pública de Praga, tendo como parceiros a Escola de Economia e Direito de Berlim, a Escola Nacional de Administração Pública da Polónia e o INA, de Portugal. A coordenação portuguesa do estudo esteve a cargo da Prof. Doutora Helena Rato e da Dra. Matilde Gago, com a colaboração do Prof. Doutor César Madureira e da Dra. Margarida Quintela, ex investigadores do INA, atualmente na DGAEP.
Resumo:
This report describes a study based on a research and evaluation of platforms (JavaScript frameworks and libraries carried out under the discipline of Tese/Dissertação/Estágio in the Master Degree Computer Science - area of expertise in Arquitecturas Sistemas e Redes, in Instituto Superior de Engenharia do Porto (ISEP). The JavaScript language is increasingly present in the development of software and particularly software for the web environment. For this reason there is a need to investigate the issue, as the platforms that are emerging in the market have a very low level of maturity compared to existing, it is crucial to identify the main differences so that we can save time, labor and other costs in research and experimentation developers. This project aims to inform a range of platforms and existing JavaScript libraries on the market. For further investigation, a study of use cases used at this point in the market was held. To assist the work were used qualitative research methods through the existing content of research related to the area, and quantitative research methods. It was thus conducted a survey sent via email to fifty regular users of the language, platforms, and JavaScript libraries, linking the functional and non-functional characteristics of the platforms and existing use cases in the market, in order to realize what is the use of these tools in his professional life. This study allowed developers to access information that will be compared and evaluated in order to achieve a more precise evaluation of the various analyzed platforms. It was also made an evaluation of adaptation of the various platforms to the various use cases.
Resumo:
Based on the report for the unit “Sociology of New Information Technologies” of the Master on Computer Sciences at FCT/University Nova Lisbon in 2015-16. The responsible of this curricular unit is Prof. António Moniz
Resumo:
Relatório de estágio de mestrado em Ensino de Informática
Resumo:
Archeology and related areas have a special interest on cultural heritage sites since they provide valuable information about past civilizations. However, the ancient buildings present in these sites are commonly found in an advanced state of degradation which difficult the professional/expert analysis. Virtual reconstructions of such buildings aim to provide a digital insight of how these historical places could have been in ancient times. Moreover, the visualization of such models has been explored by some Augmented Reality (AR) systems capable of providing support to experts. Their compelling and appealing environments have also been applied to promote the social and cultural participation of general public. The existing AR solutions regarding this thematic rarely explore the potential of realism, due to the following lacks: the exploration of mixed environments is usually only supported for indoors or outdoors, not both in the same system; the adaptation of the illumination conditions to the reconstructed structures is rarely addressed causing a decrease of credibility. MixAR [1] is a system concerned with those challenges, aiming to provide the visualization of virtual buildings augmented upon real ruins, allowing soft transitions among its interiors and exteriors and using relighting techniques for a faithful interior illumination, while the user freely moves in a given cultural heritage site, carrying a mobile unit. Regarding the focus of this paper, we intend to report the current state of MixAR mobile unit prototype, which allows visualizing virtual buildings – properly aligned with real-world structures – based on user's location, during outdoor navigation. In order to evaluate the prototype performance, a set of tests were made using virtual models with different complexities.
Resumo:
Children are an especially vulnerable population, particularly in respect to drug administration. It is estimated that neonatal and pediatric patients are at least three times more vulnerable to damage due to adverse events and medication errors than adults are. With the development of this framework, it is intended the provision of a Clinical Decision Support System based on a prototype already tested in a real environment. The framework will include features such as preparation of Total Parenteral Nutrition prescriptions, table pediatric and neonatal emergency drugs, medical scales of morbidity and mortality, anthropometry percentiles (weight, length/height, head circumference and BMI), utilities for supporting medical decision on the treatment of neonatal jaundice and anemia and support for technical procedures and other calculators and widespread use tools. The solution in development means an extension of INTCare project. The main goal is to provide an approach to get the functionality at all times of clinical practice and outside the hospital environment for dissemination, education and simulation of hypothetical situations. The aim is also to develop an area for the study and analysis of information and extraction of knowledge from the data collected by the use of the system. This paper presents the architecture, their requirements and functionalities and a SWOT analysis of the solution proposed.