940 resultados para Encryption schemes


Relevância:

10.00% 10.00%

Publicador:

Resumo:

An individual experiences double coverage when he bene ts from more than one health insurance plan at the same time. This paper examines the impact of such supplementary insurance on the demand for health care services. Its novelty is that within the context of count data modelling and without imposing restrictive parametric assumptions, the analysis is carried out for di¤erent points of the conditional distribution, not only for its mean location. Results indicate that moral hazard is present across the whole outcome distribution for both public and private second layers of health insurance coverage but with greater magnitude in the latter group. By looking at di¤erent points we unveil that stronger double coverage e¤ects are smaller for high levels of usage. We use data for Portugal, taking advantage of particular features of the public and private protection schemes on top of the statutory National Health Service. By exploring the last Portuguese Health Survey, we were able to evaluate their impacts on the consumption of doctor visi

Relevância:

10.00% 10.00%

Publicador:

Resumo:

O presente artigo intenta sistematizar as formas ou regimes do emprego público em países da União Europeia e da OCDE. A metodologia compreende a análise dos tradicionais sistemas de carreira (career-based system) e sistema de emprego (position-based system) no emprego público. Desenvolvem-se, ainda, breves reflexões nas mudanças operadas naqueles dois regimes que migraram para um terceiro modelo, vulgarmente designado por modelo híbrido de emprego público, mais flexível e mais correlativo às circunstâncias do século XXI.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

A biomassa é uma das fontes de energia renovável com maior potencial em Portugal, sendo a capacidade de produção de pellets de biomassa atualmente instalada superior a 1 milhão de toneladas/ano. Contudo, a maioria desta produção destina-se à exportação ou à utilização em centrais térmicas a biomassa, cujo crescimento tem sido significativo nos últimos anos, prevendo-se que a capacidade instalada em 2020 seja de aproximadamente 250 MW. O mercado português de caldeiras a pellets é bastante diversificado. O estudo que realizamos permitiu concluir que cerca de 90% das caldeiras existentes no mercado português têm potências inferiores a 60 kW, possuindo na sua maioria grelha fixa (81%), com sistema de ignição eléctrica (92%) e alimentação superior do biocombustível sólido (94%). O objetivo do presente trabalho foi o desenvolvimento de um modelo para simulação de uma caldeira a pellets de biomassa, que para além de permitir otimizar o projeto e operação deste tipo de equipamento, permitisse avaliar as inovações tecnológicas nesta área. Para tal recorreu-se o BiomassGasificationFoam, um código recentemente publicado, e escrito para utilização com o OpenFOAM, uma ferramenta computacional de acesso livre, que permite a simulação dos processos de pirólise, gasificação e combustão de biomassa. Este código, que foi inicialmente desenvolvido para descrever o processo de gasificação na análise termogravimétrica de biomassa, foi por nós adaptado para considerar as reações de combustão em fase gasosa dos gases libertados durante a pirólise da biomassa (recorrendo para tal ao solver reactingFoam), e ter a possibilidade de realizar a ignição da biomassa, o que foi conseguido através de uma adaptação do código de ignição do XiFoam. O esquema de ignição da biomassa não se revelou adequado, pois verificou-se que a combustão parava sempre que a ignição era inativada, independentemente do tempo que ela estivesse ativa. Como alternativa, usaram-se outros dois esquemas para a combustão da biomassa: uma corrente de ar quente, e uma resistência de aquecimento. Ambos os esquemas funcionaram, mas nunca foi possível fazer com que a combustão fosse autossustentável. A análise dos resultados obtidos permitiu concluir que a extensão das reações de pirólise e de gasificação, que são ambas endotérmicas, é muito pequena, pelo que a quantidade de gases libertados é igualmente muito pequena, não sendo suficiente para libertar a energia necessária à combustão completa da biomassa de uma maneira sustentável. Para tentar ultrapassar esta dificuldade foram testadas várias alternativas, , que incluíram o uso de diferentes composições de biomassa, diferentes cinéticas, calores de reação, parâmetros de transferência de calor, velocidades do ar de alimentação, esquemas de resolução numérica do sistema de equações diferenciais, e diferentes parâmetros dos esquemas de resolução utilizados. Todas estas tentativas se revelaram infrutíferas. Este estudo permitiu concluir que o solver BiomassGasificationFoam, que foi desenvolvido para descrever o processo de gasificação de biomassa em meio inerte, e em que a biomassa é aquecida através de calor fornecido pelas paredes do reator, aparentemente não é adequado à descrição do processo de combustão da biomassa, em que a combustão deve ser autossustentável, e em que as reações de combustão em fase gasosa são importantes. Assim, é necessário um estudo mais aprofundado que permita adaptar este código à simulação do processo de combustão de sólidos porosos em leito fixo.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

This article presents a work-in-progress version of a Dublin Core Application Profile (DCAP) developed to serve the Social and Solidarity Economy (SSE). Studies revealed that this community is interested in implementing both internal interoperability between their Web platforms to build a global SSE e-marketplace, and external interoperability among their Web platforms and external ones. The Dublin Core Application Profile for Social and Solidarity Economy (DCAP-SSE) serves this purpose. SSE organisations are submerged in the market economy but they have specificities not taken into account in this economy. The DCAP-SSE integrates terms from well-known metadata schemas, Resource Description Framework (RDF) vocabularies or ontologies, in order to enhance interoperability and take advantage of the benefits of the Linked Open Data ecosystem. It also integrates terms from the new essglobal RDF vocabulary which was created with the goal to respond to the SSE-specific needs. The DCAP-SSE also integrates five new Vocabulary Encoding Schemes to be used with DCAP-SSE properties. The DCAP development was based on a method for the development of application profiles (Me4MAP). We believe that this article has an educational value since it presents the idea that it is important to base DCAP developments on a method. This article shows the main results of applying such a method.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Ergonomic interventions such as increased scheduled breaks or job rotation have been proposed to reduce upper limb muscle fatigue in repetitive low-load work. This review was performed to summarize and analyze the studies investigating the effect of job rotation and work-rest schemes, as well as, work pace, cycle time and duty cycle, on upper limb muscle fatigue. The effects of these work organization factors on subjective fatigue or discomfort were also analyzed. This review was based on relevant articles published in PubMed, Scopus and Web of Science. The studies included in this review were performed in humans and assessed muscle fatigue in upper limbs. 14 articles were included in the systematic review. Few studies were performed in a real work environment and the most common methods used to assess muscle fatigue were surface electromyography (EMG). No consistent results were found related to the effects of job rotation on muscle activity and subjective measurements of fatigue. Rest breaks had some positive effects, particularly in perceived discomfort. The increase in work pace reveals a higher muscular load in specific muscles. The duration of experiments and characteristics of participants appear to be the factors that most have influenced the results. Future research should be focused on the improvement of the experimental protocols and instrumentation, in order to the outcomes represent adequately the actual working conditions. Relevance to industry: Introducing more physical workload variation in low-load repetitive work is considered an effective ergonomic intervention against muscle fatigue and musculoskeletal disorders in industry. Results will be useful to identify the need of future research, which will eventually lead to the adoption of best industrial work practices according to the workers capabilities.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

In recent years, vehicular cloud computing (VCC) has emerged as a new technology which is being used in wide range of applications in the area of multimedia-based healthcare applications. In VCC, vehicles act as the intelligent machines which can be used to collect and transfer the healthcare data to the local, or global sites for storage, and computation purposes, as vehicles are having comparatively limited storage and computation power for handling the multimedia files. However, due to the dynamic changes in topology, and lack of centralized monitoring points, this information can be altered, or misused. These security breaches can result in disastrous consequences such as-loss of life or financial frauds. Therefore, to address these issues, a learning automata-assisted distributive intrusion detection system is designed based on clustering. Although there exist a number of applications where the proposed scheme can be applied but, we have taken multimedia-based healthcare application for illustration of the proposed scheme. In the proposed scheme, learning automata (LA) are assumed to be stationed on the vehicles which take clustering decisions intelligently and select one of the members of the group as a cluster-head. The cluster-heads then assist in efficient storage and dissemination of information through a cloud-based infrastructure. To secure the proposed scheme from malicious activities, standard cryptographic technique is used in which the auotmaton learns from the environment and takes adaptive decisions for identification of any malicious activity in the network. A reward and penalty is given by the stochastic environment where an automaton performs its actions so that it updates its action probability vector after getting the reinforcement signal from the environment. The proposed scheme was evaluated using extensive simulations on ns-2 with SUMO. The results obtained indicate that the proposed scheme yields an improvement of 10 % in detection rate of malicious nodes when compared with the existing schemes.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Maintaining a high level of data security with a low impact on system performance is more challenging in wireless multimedia applications. Protocols that are used for wireless local area network (WLAN) security are known to significantly degrade performance. In this paper, we propose an enhanced security system for a WLAN. Our new design aims to decrease the processing delay and increase both the speed and throughput of the system, thereby making it more efficient for multimedia applications. Our design is based on the idea of offloading computationally intensive encryption and authentication services to the end systems’ CPUs. The security operations are performed by the hosts’ central processor (which is usually a powerful processor) before delivering the data to a wireless card (which usually has a low-performance processor). By adopting this design, we show that both the delay and the jitter are significantly reduced. At the access point, we improve the performance of network processing hardware for real-time cryptographic processing by using a specialized processor implemented with field-programmable gate array technology. Furthermore, we use enhanced techniques to implement the Counter (CTR) Mode with Cipher Block Chaining Message Authentication Code Protocol (CCMP) and the CTR protocol. Our experiments show that it requires timing in the range of 20–40 μs to perform data encryption and authentication on different end-host CPUs (e.g., Intel Core i5, i7, and AMD 6-Core) as compared with 10–50 ms when performed using the wireless card. Furthermore, when compared with the standard WiFi protected access II (WPA2), results show that our proposed security system improved the speed to up to 3.7 times.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

In this manuscript we tackle the problem of semidistributed user selection with distributed linear precoding for sum rate maximization in multiuser multicell systems. A set of adjacent base stations (BS) form a cluster in order to perform coordinated transmission to cell-edge users, and coordination is carried out through a central processing unit (CU). However, the message exchange between BSs and the CU is limited to scheduling control signaling and no user data or channel state information (CSI) exchange is allowed. In the considered multicell coordinated approach, each BS has its own set of cell-edge users and transmits only to one intended user while interference to non-intended users at other BSs is suppressed by signal steering (precoding). We use two distributed linear precoding schemes, Distributed Zero Forcing (DZF) and Distributed Virtual Signalto-Interference-plus-Noise Ratio (DVSINR). Considering multiple users per cell and the backhaul limitations, the BSs rely on local CSI to solve the user selection problem. First we investigate how the signal-to-noise-ratio (SNR) regime and the number of antennas at the BSs impact the effective channel gain (the magnitude of the channels after precoding) and its relationship with multiuser diversity. Considering that user selection must be based on the type of implemented precoding, we develop metrics of compatibility (estimations of the effective channel gains) that can be computed from local CSI at each BS and reported to the CU for scheduling decisions. Based on such metrics, we design user selection algorithms that can find a set of users that potentially maximizes the sum rate. Numerical results show the effectiveness of the proposed metrics and algorithms for different configurations of users and antennas at the base stations.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

4th International Conference on Future Generation Communication Technologies (FGCT 2015), Luton, United Kingdom.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Com as dificuldades económicas que se fazem sentir nos últimos anos, tanto em Portugal como no resto do mundo, em qualquer área de negócio, a qualidade tornou-se um dos fatores de sucesso. Com um mercado cada vez com mais concorrência, os clientes procuram frequentemente a melhor qualidade do produto ou do serviço da empresa. Para isso as empresas cada vez mais apostam em fornecer um produto de qualidade através de um sistema da qualidade bem delineado, bem gerido e certificado. O presente trabalho que foi desenvolvido nesta dissertação de Mestrado consistiu na conceção e acompanhamento da Implementação do Sistema de Gestão da Qualidade (SGQ) segundo a norma NP EN ISO 9001:2008 realizada numa indústria de componentes de calçado, designada NEWCOMP, Componentes para Calçado, LDA. (NCP), localizada em S. João da Madeira. Será uma ferramenta bastante útil e estratégica para a empresa a certificação, com inúmeras vantagens e benefícios para a sua imagem. As principais razões que levaram a empresa a interessar-se pela certificação do SGQ foi, principalmente, a melhoria da qualidade dos seus produtos, satisfação dos clientes e alcançar novos mercados. A documentação elaborada do SGQ inclui a política e objetivos da qualidade, esquemas dos principais processos, manual de funções entre outros documentos considerados necessários, para utilização de todas as partes envolvidas.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

A evolução tecnológica das últimas décadas na área das Tecnologias da Informação e Comunicação (TIC) contribuiu para a proliferação de fontes de informação e de sistemas de partilha de recursos. As diversas redes sociais são um exemplo paradigmático de sistemas de partilha tanto de informação como de recursos (e.g. audiovisuais). Essa abundância crescente de recursos e fontes aumenta a importância de sistemas capazes de recomendar em tempo útil recursos personalizados, tendo por base o perfil e o contexto do utilizador. O objetivo deste projeto é partilhar e recomendar locais, artigos e vídeos em função do contexto do utilizador assim como proporcionar uma experiência mais rica de reprodução dos vídeos partilhados, simulando as condições de gravação dos vídeos. Este sistema teve como inspiração dois projetos anteriormente desenvolvidos de partilha e recomendação de locais, artigos e vídeos turísticos em função da localização do utilizador. O sistema desenvolvido consiste numa aplicação distribuída composta por um módulo cliente Android, que inclui a interface com o utilizador e o consumo direto de serviços externos de suporte, e um módulo servidor que controla o acesso à base de dados central e inclui o serviço de recomendação baseado no contexto do utilizador. A comunicação entre os módulos cliente e servidor utiliza um protocolo do nível de aplicação dedicado. As recomendações geradas pelo sistema têm por base o perfil de utilizador, informação contextual (posição do utilizador, data e hora atual e velocidade atual do utilizador) e podem ser geradas a pedido do utilizador ou automaticamente, caso sejam encontrados pontos de interesse de grande relevância para o utilizador. Os pontos de interesse recomendados são apresentados com recurso ao Google Maps, incluindo o período de funcionamento, artigos complementares e a reprodução imersiva dos vídeos relacionados. Essa imersão tem em consideração as condições meteorológicas, temporais e espaciais aquando da gravação do vídeo.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The paper revisits the convolution operator and addresses its generalization in the perspective of fractional calculus. Two examples demonstrate the feasibility of the concept using analytical expressions and the inverse Fourier transform, for real and complex orders. Two approximate calculation schemes in the time domain are also tested.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The Azores archipelago is a zone with a vast cultural heritage, presenting a building stock mainly constructed in traditional stone masonry. It is known that this type of construction exhibits poor behaviour under seismic excitations; however it is extensively used in seismic prone areas, such as this case. The 9th of July of 1998 earthquake was the last seismic event in the islands, leaving many traditional stone constructions severely damaged or totally destroyed. This scenario led to an effort by the local government of improving the seismic resistance of these constructions, with the application of several reinforcement techniques. This work aims to study some of the most used reinforcement schemes after the 1998 earthquake, and to assess their effectiveness in the mitigation of the construction’s seismic vulnerability. A brief evaluation of the cost versus benefit of these retrofitting techniques is also made, seeking to identify those that are most suitable for each building typology. Thus, it was sought to analyze the case of real structures with different geometrical and physical characteristics, by establishing a comparison between the seismic performance of reinforced and non-reinforced structures. The first section contains the analysis of a total of six reinforcement scenarios for each building chosen. Using the recorded 1998 earthquake accelerograms, a linear time-history analysis was performed for each reinforcement scenario. A comparison was then established between the maximum displacements, inter-storey drift and maximum stress obtained, in order to evaluate the global seismic response of each reinforced structure. In the second part of the work, the examination of the performance obtained in the previous section, in relation to the cost of implementing each reinforcement technique, allowed to draw conclusions concerning the viability of implementing each reinforcement method, based on the book value of the buildings in study.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The purpose of this work is to present an algorithm to solve nonlinear constrained optimization problems, using the filter method with the inexact restoration (IR) approach. In the IR approach two independent phases are performed in each iteration—the feasibility and the optimality phases. The first one directs the iterative process into the feasible region, i.e. finds one point with less constraints violation. The optimality phase starts from this point and its goal is to optimize the objective function into the satisfied constraints space. To evaluate the solution approximations in each iteration a scheme based on the filter method is used in both phases of the algorithm. This method replaces the merit functions that are based on penalty schemes, avoiding the related difficulties such as the penalty parameter estimation and the non-differentiability of some of them. The filter method is implemented in the context of the line search globalization technique. A set of more than two hundred AMPL test problems is solved. The algorithm developed is compared with LOQO and NPSOL software packages.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

SUMMARY Combination Antiretroviral Therapy (cART) aims to inhibit viral replication, delay immunodeficiency progression and improve survival in AIDS patients. The objective of this study was to compare two different schemes of cART, based on plasma viral load (VL) and CD4+ T lymphocyte count, during 48 weeks of treatment. For this purpose, 472 medical charts of a Specialized Outpatient Service were reviewed from 1998 to 2005. Out of these, 58 AIDS patients who had received a triple drug scheme as the initial treatment were included in the study and two groups were formed: Group 1 (G1): 47 individuals treated with two nucleoside reverse-transcriptase inhibitors (NRTI) and one non-nucleoside reverse-transcriptase inhibitor; Group 2 (G2): 11 patients treated with two NRTI and one protease inhibitor. In G1 and G2, 53.2% and 81.8% respectively were patients with an AIDS-defining disease. The T CD4+ lymphocyte count increased progressively up until the 24th week of treatment in all patients, while VL became undetectable in 68.1% of G1 and in 63.6% of G2. The study concluded that the evolutions of laboratory tests were similar in the two treatment groups and that both presented a favorable clinical evolution.