980 resultados para Memory systems
Resumo:
Dissertação de Mestrado, Geologia do Ambiente e Sociedade, 15 de Fevereiro de 2016, Universidade dos Açores.
Resumo:
Actualmente, os smartphones e outros dispositivos móveis têm vindo a ser dotados com cada vez maior poder computacional, sendo capazes de executar um vasto conjunto de aplicações desde simples programas de para tirar notas até sofisticados programas de navegação. Porém, mesmo com a evolução do seu hardware, os actuais dispositivos móveis ainda não possuem as mesmas capacidades que os computadores de mesa ou portáteis. Uma possível solução para este problema é distribuir a aplicação, executando partes dela no dispositivo local e o resto em outros dispositivos ligados à rede. Adicionalmente, alguns tipos de aplicações como aplicações multimédia, jogos electrónicos ou aplicações de ambiente imersivos possuem requisitos em termos de Qualidade de Serviço, particularmente de tempo real. Ao longo desta tese é proposto um sistema de execução de código remota para sistemas distribuídos com restrições de tempo-real. A arquitectura proposta adapta-se a sistemas que necessitem de executar periodicamente e em paralelo mesmo conjunto de funções com garantias de tempo real, mesmo desconhecendo os tempos de execução das referidas funções. A plataforma proposta foi desenvolvida para sistemas móveis capazes de executar o Sistema Operativo Android.
Resumo:
As emoções dos indivíduos e o contexto social do grupo onde estes estão inseridos têm influência no seu desempenho no que se refere ao desenvolvimento de várias tarefas, incluindo as que são realizadas via electrónica. O processo de geração de ideias em grupo mediado por computador tem vantagens consideráveis em relação ao processo de geração de ideias em grupo tradicional, nomeadamente no que se refere ao aumento da sinergia entre os elementos do grupo, à existência da memória de grupo e à possibilidade dos elementos estarem dispersos no espaço e no tempo. Com isto em mente, o presente trabalho pretende analisar a importância do estado de espírito do participante e a influência que os vários aspectos sociais têm no participante, para assim ser possível tomar determinadas acções com o objectivo de potenciar o desempenho dos utilizadores ao longo da reunião de geração de ideias. Neste trabalho é analisada a influência que o estado de espírito dos participantes e o contexto social das reuniões podem ter no sucesso de uma reunião de geração de ideias electrónica. Considerando a influência de estes factores, é proposto um modelo que inclui essas variáveis no processo de geração de ideias em grupo mediado por computador. Com isto pretende-se demonstrar que a inclusão do modelo proposto numa ferramenta de apoio à geração de ideias em grupo permite melhorar o desempenho individual e consequentemente o desempenho do grupo, bem como a interacção entre todos os elementos. Assim, este trabalho pretende gerar sugestões com o objectivo de manter os participantes atentos e motivados para as tarefas que têm de realizar, nomeadamente a tarefa de geração de ideias. Com o objectivo de aplicar o modelo proposto é também apresentado neste trabalho uma nova ferramenta de geração de ideias em computador que considera o contexto emocional e social da reunião, o S-IGTAI (Social Idea Generation Tool for Ambient Intelligence). Através das interacções entre os participantes e a ferramenta S-IGTAI, é recolhida informação que será o input do modelo proposto, sendo que o output serão as sugestões enviadas para o facilitador. Estas sugestões têm o propósito que o facilitador realize recomendações aos participantes no sentido de manter os seus estados de espírito num nível positivo e eliminar a influência negativa dos vários aspectos sociais, potenciando dessa forma o desempenho de todos os participantes. Com a finalidade de validar o modelo proposto e a nova ferramenta (S-IGTAI) é apresentado um caso de estudo neste documento que permite realizar a avaliação do trabalho desenvolvido.
Resumo:
Actualmente verifica-se que a complexidade dos sistemas informáticos tem vindo a aumentar, fazendo parte das nossas ferramentas diárias de trabalho a utilização de sistemas informáticos e a utilização de serviços online. Neste âmbito, a internet obtém um papel de destaque junto das universidades, ao permitir que alunos e professores possam interagir mais facilmente. A internet e a educação baseada na Web vêm oferecer acesso remoto a qualquer informação independentemente da localização ou da hora. Como consequência, qualquer pessoa com uma ligação à internet, ao poder adquirir informações sobre um determinado tema junto dos maiores peritos, obtém vantagens significativas. Os laboratórios remotos são uma solução muito valorizada no que toca a interligar tecnologia e recursos humanos em ambientes que podem estar afastados no tempo ou no espaço. A criação deste tipo de laboratórios e a sua utilidade real só é possível porque as tecnologias de comunicação emergentes têm contribuído de uma forma muito relevante para melhorar a sua disponibilização à distância. A necessidade de criação de laboratórios remotos torna-se imprescindível para pesquisas relacionadas com engenharia que envolvam a utilização de recursos escassos ou de grandes dimensões. Apoiado neste conceito, desenvolveu-se um laboratório remoto para os alunos de engenharia que precisam de testar circuitos digitais numa carta de desenvolvimento de hardware configurável, permitindo a utilização deste recurso de uma forma mais eficiente. O trabalho consistiu na criação de um laboratório remoto de baixo custo, com base em linguagens de programação open source, sendo utilizado como unidade de processamento um router da ASUS com o firmware OpenWrt. Este firmware é uma distribuição Linux para sistemas embutidos. Este laboratório remoto permite o teste dos circuitos digitais numa carta de desenvolvimento de hardware configurável em tempo real, utilizando a interface JTAG. O laboratório desenvolvido tem a particularidade de ter como unidade de processamento um router. A utilização do router como servidor é uma solução muito pouco usual na implementação de laboratórios remotos. Este router, quando comparado com um computador normal, apresenta uma capacidade de processamento e memória muito inferior, embora os testes efectuados provassem que apresenta um desempenho muito adequado às expectativas.
Resumo:
This thesis presents the Fuzzy Monte Carlo Model for Transmission Power Systems Reliability based studies (FMC-TRel) methodology, which is based on statistical failure and repair data of the transmission power system components and uses fuzzyprobabilistic modeling for system component outage parameters. Using statistical records allows developing the fuzzy membership functions of system component outage parameters. The proposed hybrid method of fuzzy set and Monte Carlo simulation based on the fuzzy-probabilistic models allows catching both randomness and fuzziness of component outage parameters. A network contingency analysis to identify any overloading or voltage violation in the network is performed once obtained the system states. This is followed by a remedial action algorithm, based on Optimal Power Flow, to reschedule generations and alleviate constraint violations and, at the same time, to avoid any load curtailment, if possible, or, otherwise, to minimize the total load curtailment, for the states identified by the contingency analysis. For the system states that cause load curtailment, an optimization approach is applied to reduce the probability of occurrence of these states while minimizing the costs to achieve that reduction. This methodology is of most importance for supporting the transmission system operator decision making, namely in the identification of critical components and in the planning of future investments in the transmission power system. A case study based on Reliability Test System (RTS) 1996 IEEE 24 Bus is presented to illustrate with detail the application of the proposed methodology.
Resumo:
The main goal of this research study was the removal of Cu(II), Ni(II) and Zn(II) from aqueous solutions using peanut hulls. This work was mainly focused on the following aspects: chemical characterization of the biosorbent, kinetic studies, study of the pH influence in mono-component systems, equilibrium isotherms and column studies, both in mono and tri-component systems, and with a real industrial effluent from the electroplating industry. The chemical characterization of peanut hulls showed a high cellulose (44.8%) and lignin (36.1%) content, which favours biosorption of metal cations. The kinetic studies performed indicate that most of the sorption occurs in the first 30 min for all systems. In general, a pseudo-second order kinetics was followed, both in mono and tri-component systems. The equilibrium isotherms were better described by Freundlich model in all systems. Peanut hulls showed higher affinity for copper than for nickel and zinc when they are both present. The pH value between 5 and 6 was the most favourable for all systems. The sorbent capacity in column was 0.028 and 0.025 mmol g-1 for copper, respectively in mono and tri-component systems. A decrease of capacity for copper (50%) was observed when dealing with the real effluent. The Yoon-Nelson, Thomas and Yan’s models were fitted to the experimental data, being the latter the best fit.
Resumo:
Radiotherapy (RT) is one of the most important approaches in the treatment of cancer and its performance can be improved in three different ways: through the optimization of the dose distribution, by the use of different irradiation techniques or through the study of radiobiological initiatives. The first is purely physical because is related to the physical dose distributiuon. The others are purely radiobiological because they increase the differential effect between the tumour and the health tissues. The Treatment Planning Systems (TPS) are used in RT to create dose distributions with the purpose to maximize the tumoral control and minimize the complications in the healthy tissues. The inverse planning uses dose optimization techniques that satisfy the criteria specified by the user, regarding the target and the organs at risk (OAR’s). The dose optimization is possible through the analysis of dose-volume histograms (DVH) and with the use of computed tomography, magnetic resonance and other digital image techniques.
Resumo:
A flow injection analysis (FIA) system having a chlormequat selective electrode is proposed. Several electrodes with poly(vinyl chloride) based membranes were constructed for this purpose. Comparative characterization suggestedthe use of membrane with chlormequat tetraphenylborate and dibutylphthalate. On a single-line FIA set-up, operating with 1x10-2 mol L-1 ionic strength and 6.3 pH, calibration curves presented slopes of 53.6±0.4mV decade-1 within 5.0x10-6 and1.0x10-3 mol L-1, andsquaredcorrelation coefficients >0.9953. The detection limit was 2.2x10-6 mol L-1 and the repeatability equal to ±0.68mV (0.7%). A dual-channel FIA manifold was therefore constructed, enabling automatic attainment of previous ionic strength andpH conditions and thus eliminating sample preparation steps. Slopes of 45.5±0.2mV decade -1 along a concentration range of 8.0x10-6 to 1.0x10-3 mol L-1 with a repeatability ±0.4mV (0.69%) were obtained. Analyses of real samples were performed, and recovery gave results ranging from 96.6 to 101.1%.
Resumo:
Infiltration galleries are among the oldest known means used for small public water fountains. Owing to its ancestral origin they are usually associated with high quality water. Thirty-one compounds, including pesticides and estrogens from different chemical families, were analysed in waters from infiltration galleries collected in Alto Douro Demarcated Wine region (North of Portugal). A total of twelve compounds were detected in the water samples. Nine of these compounds are described as presenting evidence or potential evidence of interfering with the hormone system of humans and wildlife. Although concentrations of the target analytes were relatively low, many of them below their limit of quantification, four compounds were above quantification limit and two of them even above the legal limit of 0.1 lg/L: dimethoate (30.38 ng/L), folpet (64.35 ng/L), terbuthylazine-desethyl (22.28 to 292.36 ng/L) and terbuthylazine (22.49 to 369.33 ng/L).
Resumo:
Doutoramento em Gestão
Resumo:
Pesticide exposure during brain development could represent an important risk factor for the onset of neurodegenerative diseases. Previous studies investigated the effect of permethrin (PERM) administered at 34 mg/kg, a dose close to the no observable adverse effect level (NOAEL) from post natal day (PND) 6 to PND 21 in rats. Despite the PERM dose did not elicited overt signs of toxicity (i.e. normal body weight gain curve), it was able to induce striatal neurodegeneration (dopamine and Nurr1 reduction, and lipid peroxidation increase). The present study was designed to characterize the cognitive deficits in the current animal model. When during late adulthood PERM treated rats were tested for spatial working memory performances in a T-maze-rewarded alternation task they took longer to choose for the correct arm in comparison to age matched controls. No differences between groups were found in anxiety-like state, locomotor activity, feeding behavior and spatial orientation task. Our findings showing a selective effect of PERM treatment on the T-maze task point to an involvement of frontal cortico-striatal circuitry rather than to a role for the hippocampus. The predominant disturbances concern the dopamine (DA) depletion in the striatum and, the serotonin (5-HT) and noradrenaline (NE) unbalance together with a hypometabolic state in the medial prefrontal cortex area. In the hippocampus, an increase of NE and a decrease of DA were observed in PERM treated rats as compared to controls. The concentration of the most representative marker for pyrethroid exposure (3-phenoxybenzoic acid) measured in the urine of rodents 12 h after the last treatment was 41.50 µ/L and it was completely eliminated after 96 h.
Resumo:
With the advent of Web 2.0, new kinds of tools became available, which are not seen as novel anymore but are widely used. For instance, according to Eurostat data, in 2010 32% of individuals aged 16 to 74 used the Internet to post messages to social media sites or instant messaging tools, ranging from 17% in Romania to 46% in Sweden (Eurostat, 2012). Web 2.0 applications have been used in technology-enhanced learning environments. Learning 2.0 is a concept that has been used to describe the use of social media for learning. Many Learning 2.0 initiatives have been launched by educational and training institutions in Europe. Web 2.0 applications have also been used for informal learning. Web 2.0 tools can be used in classrooms, virtual or not, not only to engage students but also to support collaborative activities. Many of these tools allow users to use tags to organize resources and facilitate their retrieval at a later date or time. The aim of this chapter is to describe how tagging has been used in systems that support formal or informal learning and to summarize the functionalities that are common to these systems. In addition, common and unusual tagging applications that have been used in some Learning Objects Repositories are analysed.
Resumo:
Competitive electricity markets have arisen as a result of power-sector restructuration and power-system deregulation. The players participating in competitive electricity markets must define strategies and make decisions using all the available information and business opportunities.
Resumo:
Embedded systems are increasingly complex and dynamic, imposing progressively higher developing time and costs. Tuning a particular system for deployment is thus becoming more demanding. Furthermore when considering systems which have to adapt themselves to evolving requirements and changing service requests. In this perspective, run-time monitoring of the system behaviour becomes an important requirement, allowing to dynamically capturing the actual scheduling progress and resource utilization. For this to succeed, operating systems need to expose their internal behaviour and state, making it available to external applications, and a runtime monitoring mechanism must be available. However, such mechanism can impose a burden in the system itself if not wisely used. In this paper we explore this problem and propose a framework, which is intended to provide this run-time mechanism whilst achieving code separation, run-time efficiency and flexibility for the final developer.
Resumo:
A dynamic scheduler that supports the coexistence of guaranteed and non-guaranteed bandwidth servers is proposed. Overloads are handled by an efficient reclaiming of residual capacities originated by early completions as well as by allowing reserved capacity stealing of non-guaranteed bandwidth servers. The proposed dynamic budget accounting mechanism ensures that at a particular time the currently executing server is using a residual capacity, its own capacity or is stealing some reserved capacity, eliminating the need of additional server states or unbounded queues. The server to which the budget accounting is going to be performed is dynamically determined at the time instant when a capacity is needed. This paper describes and evaluates the proposed scheduling algorithm, showing that it can efficiently reduce the mean tardiness of periodic jobs. The achieved results become even more significant when tasks’ computation times have a large variance.