837 resultados para Operating system kernels
Resumo:
Mémoire numérisé par la Direction des bibliothèques de l'Université de Montréal.
Resumo:
Wireless sensor networks (WSNs) differ from conventional distributed systems in many aspects. The resource limitation of sensor nodes, the ad-hoc communication and topology of the network, coupled with an unpredictable deployment environment are difficult non-functional constraints that must be carefully taken into account when developing software systems for a WSN. Thus, more research needs to be done on designing, implementing and maintaining software for WSNs. This thesis aims to contribute to research being done in this area by presenting an approach to WSN application development that will improve the reusability, flexibility, and maintainability of the software. Firstly, we present a programming model and software architecture aimed at describing WSN applications, independently of the underlying operating system and hardware. The proposed architecture is described and realized using the Model-Driven Architecture (MDA) standard in order to achieve satisfactory levels of encapsulation and abstraction when programming sensor nodes. Besides, we study different non-functional constrains of WSN application and propose two approaches to optimize the application to satisfy these constrains. A real prototype framework was built to demonstrate the developed solutions in the thesis. The framework implemented the programming model and the multi-layered software architecture as components. A graphical interface, code generation components and supporting tools were also included to help developers design, implement, optimize, and test the WSN software. Finally, we evaluate and critically assess the proposed concepts. Two case studies are provided to support the evaluation. The first case study, a framework evaluation, is designed to assess the ease at which novice and intermediate users can develop correct and power efficient WSN applications, the portability level achieved by developing applications at a high-level of abstraction, and the estimated overhead due to usage of the framework in terms of the footprint and executable code size of the application. In the second case study, we discuss the design, implementation and optimization of a real-world application named TempSense, where a sensor network is used to monitor the temperature within an area.
Resumo:
The protein folding problem has been one of the most challenging subjects in biological physics due to its complexity. Energy landscape theory based on statistical mechanics provides a thermodynamic interpretation of the protein folding process. We have been working to answer fundamental questions about protein-protein and protein-water interactions, which are very important for describing the energy landscape surface of proteins correctly. At first, we present a new method for computing protein-protein interaction potentials of solvated proteins directly from SAXS data. An ensemble of proteins was modeled by Metropolis Monte Carlo and Molecular Dynamics simulations, and the global X-ray scattering of the whole model ensemble was computed at each snapshot of the simulation. The interaction potential model was optimized and iterated by a Levenberg-Marquardt algorithm. Secondly, we report that terahertz spectroscopy directly probes hydration dynamics around proteins and determines the size of the dynamical hydration shell. We also present the sequence and pH-dependence of the hydration shell and the effect of the hydrophobicity. On the other hand, kinetic terahertz absorption (KITA) spectroscopy is introduced to study the refolding kinetics of ubiquitin and its mutants. KITA results are compared to small angle X-ray scattering, tryptophan fluorescence, and circular dichroism results. We propose that KITA monitors the rearrangement of hydrogen bonding during secondary structure formation. Finally, we present development of the automated single molecule operating system (ASMOS) for a high throughput single molecule detector, which levitates a single protein molecule in a 10 µm diameter droplet by the laser guidance. I also have performed supporting calculations and simulations with my own program codes.
Resumo:
Only recently, during the past five years, consumer electronics has been evolving rapidly. Many products have started to include “smart home” capabilities, enabling communication and interoperability of various smart devices. Even more devices and sensors can be remote controlled and monitored through cloud services. While the smart home systems have become very affordable to average consumer compared to the early solutions decades ago, there are still many issues and things that need to be fixed or improved upon: energy efficiency, connectivity with other devices and applications, security and privacy concerns, reliability, and response time. This paper focuses on designing Internet of Things (IoT) node and platform architectures that take these issues into account, notes other currently used solutions, and selects technologies in order to provide better solution. The node architecture aims for energy efficiency and modularity, while the platform architecture goals are in scalability, portability, maintainability, performance, and modularity. Moreover, the platform architecture attempts to improve user experience by providing higher reliability and lower response time compared to the alternative platforms. The architectures were developed iteratively using a development process involving research, planning, design, implementation, testing, and analysis. Additionally, they were documented using Kruchten’s 4+1 view model, which is used to describe the use cases and different views of the architectures. The node architecture consisted of energy efficient hardware, FC3180 microprocessor and CC2520 RF transceiver, modular operating system, Contiki, and a communication protocol, AllJoyn, used for providing better interoperability with other IoT devices and applications. The platform architecture provided reliable low response time control, monitoring, and initial setup capabilities by utilizing web technologies on various devices such as smart phones, tablets, and computers. Furthermore, an optional cloud service was provided in order to control devices and monitor sensors remotely by utilizing scalable high performance technologies in the backend enabling low response time and high reliability.
Resumo:
Data leakage is a serious issue and can result in the loss of sensitive data, compromising user accounts and details, potentially affecting millions of internet users. This paper contributes to research in online security and reducing personal footprint by evaluating the levels of privacy provided by the Firefox browser. The aim of identifying conditions that would minimize data leakage and maximize data privacy is addressed by assessing and comparing data leakage in the four possible browsing modes: normal and private modes using a browser installed on the host PC or using a portable browser from a connected USB device respectively. To provide a firm foundation for analysis, a series of carefully designed, pre-planned browsing sessions were repeated in each of the various modes of Firefox. This included low RAM environments to determine any effects low RAM may have on browser data leakage. The results show that considerable data leakage may occur within Firefox. In normal mode, all of the browsing information is stored within the Mozilla profile folder in Firefox-specific SQLite databases and sessionstore.js. While passwords were not stored as plain text, other confidential information such as credit card numbers could be recovered from the Form history under certain conditions. There is no difference when using a portable browser in normal mode, except that the Mozilla profile folder is located on the USB device rather than the host's hard disk. By comparison, private browsing reduces data leakage. Our findings confirm that no information is written to the Firefox-related locations on the hard disk or USB device during private browsing, implying that no deletion would be necessary and no remnants of data would be forensically recoverable from unallocated space. However, two aspects of data leakage occurred equally in all four browsing modes. Firstly, all of the browsing history was stored in the live RAM and was therefore accessible while the browser remained open. Secondly, in low RAM situations, the operating system caches out RAM to pagefile.sys on the host's hard disk. Irrespective of the browsing mode used, this may include Firefox history elements which can then remain forensically recoverable for considerable time.
Resumo:
Trabalho Final de Mestrado para obtenção do Grau de Mestre em Engenharia de Redes de Comunicação e Multimédia
Resumo:
In this work a system of autonomous agents engaged in cyclic pursuit (under constant bearing (CB) strategy) is considered, for which one informed agent (the leader) also senses and responds to a stationary beacon. Building on the framework proposed in a previous work on beacon-referenced cyclic pursuit, necessary and suffi- cient conditions for the existence of circling equilibria in a system with one informed agent are derived, with discussion of stability and performance. In a physical testbed, the leader (robot) is equipped with a sound sensing apparatus composed of a real time embedded system, estimating direction of arrival of sound by an Interaural Level and Phase Difference Algorithm, using empirically determined phase and level signatures, and breaking front-back ambiguity with appropriate sensor placement. Furthermore a simple framework for implementing and evaluating the performance of control laws with the Robot Operating System (ROS) is proposed, demonstrated, and discussed.
Resumo:
Following a drop in estrogen in the period of menopause some women begin to lose bone mass more than 1% per year reaching the end of five years with loss greater than 25%. In this regard, factors such as older age, low calcium intake and premature menopause favor the onset of osteoporosis. Preventive methods such as nutritional counseling to a proper diet and the support of technology through applications that assess dietary intake are essential. Thus, this study aimed to develop an application for Android® platform focused on the evaluation of nutritional and organic conditions involved in bone health and risks for developing osteoporosis in postmenopausal women. To achieve this goal we proceeded to a study of 72 women aged 46-79 years, from the physical exercise for bone health of the Laboratory for Research in Biochemistry and Densitometry the Federal Technological University of Paraná program. Data were collected in the second half of 2014 through tests Bone Densitometry and Body Composition, Blood Tests, Anthropometric data and Nutrition Assessment. The study included women with a current diagnosis of osteopenia or osteoporosis primary, aged more than 45 years postmenopausal. For the assessment of bone mineral density and body composition used the device Absorptiometry Dual Energy X-ray (DXA) brand Hologic Discovery TM Model A. For anthropometric assessment was included to body mass, height, abdominal circumference, Waist circumference and hip circumference. The instrument for assessing food consumption was used Recall 24 hours a day (24HR). The estimated intake of energy and nutrients was carried from the tabulation of the food eaten in the Software Diet Pro 4®. In a sub sample of 30 women with osteopenia / osteoporosis serum calcium and alkaline phosphatase tests were performed. The results demonstrated a group of women (n = 30) average calcium intake of 570mg / day (± 340). The analysis showed a mean serum calcium within the normal range (10,20mg / dl ± 0.32) and average values and slightly increased alkaline phosphatase (105.40 U / L ± 23.70). Furthermore, there was a significant correlation between the consumption of protein and the optimal daily intake of calcium (0.375 p-value 0.05). Based on these findings, we developed an application early stage in Android® platform operating system Google®, being called OsteoNutri. We chose to use Java Eclipse® where it was executed Android® version of the project; choice of application icons and setting the visual editor for building the application layouts. The DroidDraw® was used for development of the three application GUIs. For practical tests we used a cell compatible with the version that was created (4.4 or higher). The prototype was developed in conjunction with the Group and Instrumentation Applications Development (GDAI) of the Federal Technological University of Paraná. So this application can be considered an important tool in dietary control, allowing closer control consumption of calcium and dietary proteins.
Resumo:
En este artículo se presenta a DeBuPa (Detección Búsqueda Pateo) un humanoide de tamaño pequeño (38 cm de alto) construido con las piezas del kit Bioloid. Del kit se ha excluido la tarjeta CM-510 para sustituirla por la tarjeta controladora Arbotix, que será la que controle los 16 motores Dynamixel Ax-12+ (para mover al robot) y 2 servomotores analógicos (para mover la cámara). Además se ha agregado un mini computador Raspberry Pi, con su cámara, para que el robot pueda detectar y seguir la pelota de forma autónoma. Todos estos componentes deben ser coordinados para que se logre cumplir la tarea de detectar, seguir y patear la pelota. Por ello se hace necesaria la comunicación entre la Arbotix y la Raspberry Pi. La herramienta empleada para ello es el framework ROS (Robot Operating System). En la Raspberry Pi se usa el lenguaje C++ y se ejecuta un solo programa encargado de captar la imagen de la cámara, filtrar y procesar para encontrar la pelota, tomar la decisión de la acción a ejecutar y hacer la petición a la Arbotix para que dé la orden a los motores de ejecutar el movimiento. Para captar la imagen de la cámara se ha utilizado la librería RasPiCam CV. Para filtrar y procesar la imagen se ha usado las librerías de OpenCV. La Arbotix, además de controlar los motores, se encarga de monitorizar que el robot se encuentre balanceado, para ello usa el sensor Gyro de Robotis. Si detecta un desbalance de un cierto tamaño puede saber si se ha caído y levantarse.
Resumo:
This document presents GEmSysC, an unified cryptographic API for embedded systems. Software layers implementing this API can be built over existing libraries, allowing embedded software to access cryptographic functions in a consistent way that does not depend on the underlying library. The API complies to good practices for API design and good practices for embedded software development and took its inspiration from other cryptographic libraries and standards. The main inspiration for creating GEmSysC was the CMSIS-RTOS standard, which defines an unified API for embedded software in an implementation-independent way, but targets operating systems instead of cryptographic functions. GEmSysC is made of a generic core and attachable modules, one for each cryptographic algorithm. This document contains the specification of the core of GEmSysC and three of its modules: AES, RSA and SHA-256. GEmSysC was built targeting embedded systems, but this does not restrict its use only in such systems – after all, embedded systems are just very limited computing devices. As a proof of concept, two implementations of GEmSysC were made. One of them was built over wolfSSL, which is an open source library for embedded systems. The other was built over OpenSSL, which is open source and a de facto standard. Unlike wolfSSL, OpenSSL does not specifically target embedded systems. The implementation built over wolfSSL was evaluated in a Cortex- M3 processor with no operating system while the implementation built over OpenSSL was evaluated on a personal computer with Windows 10 operating system. This document displays test results showing GEmSysC to be simpler than other libraries in some aspects. These results have shown that both implementations incur in little overhead in computation time compared to the cryptographic libraries themselves. The overhead of the implementation has been measured for each cryptographic algorithm and is between around 0% and 0.17% for the implementation over wolfSSL and between 0.03% and 1.40% for the one over OpenSSL. This document also presents the memory costs for each implementation.
Resumo:
É crescente a utilização dos dispositivos móveis com ecrãs maiores e melhores, mais memória, maiores capacidades multimédia e métodos mais refinados para introduzir dados. Dispositivos que integram comunicações, acesso à internet e diversos tipos de sensores possibilitarão, seguramente, abordagens inovadoras e criativas em atividades pedagógicas, em contraste com as utilizações atuais nos computadores pessoais. A análise das aplicações que atualmente integram os módulos do Moodle nos dispositivos móveis mostra que existe ainda um longo caminho a percorrer. As aplicações existentes têm, quase na sua totalidade, como objetivo adaptar o interface aos dispositivos móveis, o que é apenas o primeiro passo no sentido de aproveitar todas as potencialidades destes dispositivos. É, pois, possível imaginar um futuro próximo, onde as potencialidades dos dispositivos móveis darão origem a aplicações com um enorme potencial de aprendizagem, que advém do facto de os estudantes encontrarem conexões entre as suas vidas e a sua educação, através da realização de atividades em contexto no dispositivo móvel, sempre omnipresente. Com este trabalho de investigação e desenvolvimento pretende-se: a) avaliar o estado da arte do mobile learning, na área dos Learning Management System (LMS); b) refletir sobre as funcionalidades que deve oferecer uma aplicação para dispositivos móveis, com enfoque no sistema operativo Android, que permita a gestão e atualização dos fóruns e ficheiros do Moodle; c) conceber e produzir a referida aplicação, de acordo com as especificações consideradas relevantes; d) avaliar o seu impacto educativo e funcional. É demonstrado neste estudo que o recurso a dispositivos móveis potencia a aprendizagem baseada em LMS (Learning Management System), identificando-se as vantagens da sua utilização. São também apresentadas as funcionalidades da aplicação Mais(f), desenvolvida no âmbito da investigação, a avaliação da mesma pelos participantes no estudo, bem como as perspectivas futuras de utilização da aplicação Mais(f).
Resumo:
Dissertação (mestrado)—Universidade de Brasília, Instituto de Geociências, 2016.
Resumo:
Cada vez mais, se torna necessário, apresentar, transpor e adequar as Novas Tecnologias ao ambiente de aprendizagem na sala de aula, interligando não só as Áreas Curriculares Não Disciplinares (ACND) como as Áreas Curriculares Disciplinares (ACD) a um trabalho colaborativo. Por conseguinte, o objetivo deste projeto de investigação é aplicar um projeto de aprendizagem colaborativa denominado “Triângulo das Histórias”, utilizando como instrumento de partilha um blogue na internet, e avaliar o tipo de aceitação que a sua aplicação gerou. A aplicação do projeto em contexto educativo visará uma interação entre a comunidade educativa, possibilitando assim, a experimentação de comunidades virtuais de comunicação e de colaboração escolar. Para complementar este estudo pretende-se também perceber os hábitos de utilização da Internet dos alunos envolvidos no projeto. A amostra utilizada para este estudo foi um conjunto de dezoito alunos dos 1.º e 2.º Ciclos do Ensino Básico, com idades compreendidas entre os 6 e os 12 anos de idade, dos regimes público, cooperativo e particular, repartidos por três escolas distintas, com meios socioeconómicos também diferentes. Para a recolha dos dados foi realizado um inquérito aos alunos envolvidos no projeto e realizadas observações sobre as suas interações e comportamentos. Os dados extraídos dos inquéritos foram apresentados sob a forma de gráfico. A implementação deste projeto permitiu perceber que os alunos foram recetivos a esta nova abordagem de realização de trabalho de grupo colaborativo recorrendo às comunidades virtuais. A criação de uma história associando três escolas distintas permitiu que os alunos, para além de terem de elaborar um texto de forma coletiva, tivessem acesso a novas ferramentas tanto da internet como do sistema operativo, tornando-se, desta forma, um bom complemento de aprendizagem.
Resumo:
Dissertação de mestrado, Engenharia Electrónica e Telecomunicações, Faculdade de Ciências e Tecnologia, Universidade do Algarve, 2011
Resumo:
Current industry proposals for Hardware Transactional Memory (HTM) focus on best-effort solutions (BE-HTM) where hardware limits are imposed on transactions. These designs may show a significant performance degradation due to high contention scenarios and different hardware and operating system limitations that abort transactions, e.g. cache overflows, hardware and software exceptions, etc. To deal with these events and to ensure forward progress, BE-HTM systems usually provide a software fallback path to execute a lock-based version of the code. In this paper, we propose a hardware implementation of an irrevocability mechanism as an alternative to the software fallback path to gain insight into the hardware improvements that could enhance the execution of such a fallback. Our mechanism anticipates the abort that causes the transaction serialization, and stalls other transactions in the system so that transactional work loss is mini- mized. In addition, we evaluate the main software fallback path approaches and propose the use of ticket locks that hold precise information of the number of transactions waiting to enter the fallback. Thus, the separation of transactional and fallback execution can be achieved in a precise manner. The evaluation is carried out using the Simics/GEMS simulator and the complete range of STAMP transactional suite benchmarks. We obtain significant performance benefits of around twice the speedup and an abort reduction of 50% over the software fallback path for a number of benchmarks.