10 resultados para drawbacks

em Repositório Institucional da Universidade de Aveiro - Portugal


Relevância:

10.00% 10.00%

Publicador:

Resumo:

Este trabalho surge do interesse em substituir os nós de rede óptica baseados maioritariamente em electrónica por nós de rede baseados em tecnologia óptica. Espera-se que a tecnologia óptica permita maiores débitos binários na rede, maior transparência e maior eficiência através de novos paradigmas de comutação. Segundo esta visão, utilizou-se o MZI-SOA, um dispositivo semicondutor integrado hibridamente, para realizar funcionalidades de processamento óptico de sinal necessárias em nós de redes ópticas de nova geração. Nas novas redes ópticas são utilizados formatos de modulação avançados, com gestão da fase, pelo que foi estudado experimentalmente e por simulação o impacto da utilização destes formatos no desempenho do MZI-SOA na conversão de comprimento de onda e formato, em várias condições de operação. Foram derivadas regras de utilização para funcionamento óptimo. Foi também estudado o impacto da forma dos pulsos do sinal no desempenho do dispositivo. De seguida, o MZI-SOA foi utilizado para realizar funcionalidades temporais ao nível do bit e do pacote. Foi investigada a operação de um conversor de multiplexagem por divisão no comprimento de onda para multiplexagem por divisão temporal óptica, experimentalmente e por simulação, e de um compressor e descompressor de pacotes, por simulação. Para este último, foi investigada a operação com o MZI-SOA baseado em amplificadores ópticos de semicondutor com geometria de poço quântico e ponto quântico. Foi também realizado experimentalmente um ermutador de intervalos temporais que explora o MZI-SOA como conversor de comprimento de onda e usa um banco de linhas de atraso ópticas para introduzir no sinal um atraso seleccionável. Por fim, foi estudado analiticamente, experimentalmente e por simulação o impacto de diafonia em redes ópticas em diversas situações. Extendeu-se um modelo analítico de cálculo de desempenho para contemplar sinais distorcidos e afectados por diafonia. Estudou-se o caso de sinais muito filtrados e afectados por diafonia e mostrou-se que, para determinar correctamente as penalidades que ocorrem, ambos os efeitos devem ser considerados simultaneamente e não em separado. Foi estudada a escalabilidade limitada por diafonia de um comutador de intervalos temporais baseado em MZI-SOA a operar como comutador espacial. Mostrou-se também que sinais afectados fortemente por não-linearidades podem causar penalidades de diafonia mais elevadas do que sinais não afectados por não-linearidades. Neste trabalho foi demonstrado que o MZI-SOA permite construir vários e pertinentes circuitos ópticos, funcionando como bloco fundamental de construção, tendo sido o seu desempenho analisado, desde o nível de componente até ao nível de sistema. Tendo em conta as vantagens e desvantagens do MZI-SOA e os desenvolvimentos recentes de outras tecnologias, foram sugeridos tópicos de investigação com o intuito de evoluir para as redes ópticas de nova geração.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

O presente trabalho pretende auxiliar o processo de consolidação do conceito de sustentabilidade no seio das organizações. Partindo de ferramentas de gestão e avaliação já existentes, esta tese sugere a sua integração numa única metodologia, ultrapassando desse modo as limitações e potenciando as suas capacidades enquanto ferramentas isoladas. O modelo proposto para o Sistema de Gestão da Sustentabilidade (SGS) integra assim: o conceito de melhoria contínua característico dos sistemas de gestão normalizados; a capacidade de tradução da perspetiva estratégica da gestão para o plano operacional, característica do Business Scorecard (BSC); e, por fim, a avaliação emergética é ainda utilizada como uma ferramenta de avaliação da sustentabilidade de sistemas. Um objetivo secundário desta tese prende-se com o desenvolvimento de um procedimento para a realização da análise emergética de um sistema. Depois de analisada a literatura referente à utilização da análise emergética, identificou-se como necessária a definição de um procedimento normalizado, adotando um conjunto de tarefas e um formato de apresentação de resultados que permita disseminar o conceito e tornar a ferramenta mais “utilizável”. Por outro lado, procurou-se dotar o procedimento com um conjunto de indicações que permitem ultrapassar limitações e inconvenientes apontados pelos críticos mas também utilizadores do método, nomeadamente: problemas de dupla contagem, cálculo da incerteza da análise e critérios de qualidade da informação utilizada. O modelo dos sistemas de gestão normalizados apresenta um papel central na metodologia proposta. O conceito de “melhoria contínua” afigura-se como fundamental num sistema que pretende implementar o conceito “desenvolvimento sustentável” e avaliar o seu desempenho à luz do mesmo. Assim, o ciclo Plan-Do-check-Act (PDCA) deve ser utilizado para implementar o SGS de acordo com uma Política para a Sustentabilidade que a organização deve desenvolver. Definida a Política, o modelo baseia-se então no ciclo PDCA: fase de planeamento; fase de implementação; fase de verificação; e fase de revisão. É na fase de planeamento do SGS que se sugere a introdução das outras duas ferramentas: a análise emergética (AEm) e o BSC. A fase de planeamento do modelo de SGS proposto neste trabalho foi aplicada à Universidade de Aveiro (UA), incluindo a definição de uma Política para a Sustentabilidade e o planeamento estratégico e operacional. A avaliação emergética à UA foi realizada recorrendo ao procedimento desenvolvido nesta tese e permitiu caracterizar e avaliar os fluxos de recursos que a “alimentam” sob uma só unidade, atribuindo deste modo graus de importância aos diferentes recursos utilizados. A informação representa 96% do total de recursos utilizados na UA, quando avaliados sob o ponto de vista emergética. Para além da informação, os fluxos financeiros representam a maior fatia do orçamento emergético da UA, grande parte dos quais serve para sustentar os serviços prestados pelo corpo docente da UA. Analisando valores históricos de 3 indicadores de desempenho emergético, observa-se que a UA não regista uma evolução positiva em nenhum dos indicadores: a emergia utilizada nos edifícios tem-se mantido mais ou menos constante; a retribuição emergética da UA para a sociedade, avaliada sobre a forma de diplomados, tem diminuído; e a relação emergética entre professores e alunos tem também diminuído, facto que pode refletir-se na qualidade dos “produtos” da UA. Da aplicação do SGS à UA regista-se: a adequabilidade do ciclo PDCA à implementação de um SGS; a capacidade da AEm “obrigar” a organização a adotar uma abordagem sistémica da sua atividade, resultando numa visão mais aprofundada da sua relação com o contexto ambiental, económico e social em que se insere; a importância da visão estratégica e da sua tradução em termos operacionais na fase de planeamento de um SGS; e, por fim, a capacidade de adaptação e dupla funcionalidade (implementação e avaliação) do modelo de SGS proposto. A metodologia de SGS proposta nesta tese, sendo direcionada para todo o tipo de organizações, não se desvirtua quando aplicada ao contexto específico das instituições de ensino superior e permite implementar e avaliar o conceito “desenvolvimento sustentável” nas quatro dimensões da universidade (Educação, Investigação, Operação; Relação com as partes interessadas).

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The evolution and maturation of Cloud Computing created an opportunity for the emergence of new Cloud applications. High-performance Computing, a complex problem solving class, arises as a new business consumer by taking advantage of the Cloud premises and leaving the expensive datacenter management and difficult grid development. Standing on an advanced maturing phase, today’s Cloud discarded many of its drawbacks, becoming more and more efficient and widespread. Performance enhancements, prices drops due to massification and customizable services on demand triggered an emphasized attention from other markets. HPC, regardless of being a very well established field, traditionally has a narrow frontier concerning its deployment and runs on dedicated datacenters or large grid computing. The problem with common placement is mainly the initial cost and the inability to fully use resources which not all research labs can afford. The main objective of this work was to investigate new technical solutions to allow the deployment of HPC applications on the Cloud, with particular emphasis on the private on-premise resources – the lower end of the chain which reduces costs. The work includes many experiments and analysis to identify obstacles and technology limitations. The feasibility of the objective was tested with new modeling, architecture and several applications migration. The final application integrates a simplified incorporation of both public and private Cloud resources, as well as HPC applications scheduling, deployment and management. It uses a well-defined user role strategy, based on federated authentication and a seamless procedure to daily usage with balanced low cost and performance.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The number of connected devices collecting and distributing real-world information through various systems, is expected to soar in the coming years. As the number of such connected devices grows, it becomes increasingly difficult to store and share all these new sources of information. Several context representation schemes try to standardize this information, but none of them have been widely adopted. In previous work we addressed this challenge, however our solution had some drawbacks: poor semantic extraction and scalability. In this paper we discuss ways to efficiently deal with representation schemes' diversity and propose a novel d-dimension organization model. Our evaluation shows that d-dimension model improves scalability and semantic extraction.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Marine Recirculating Aquaculture Systems (RAS) produce great volume of wastewater, which may be reutilized/recirculated or reutilized after undergoing different treatment/remediation methods, or partly discharged into neighbour water-bodies (DWW). Phosphates, in particular, are usually accumulated at high concentrations in DWW, both because its monitoring is not compulsory for fish production since it is not a limiting parameter, and also because there is no specific treatment so far developed to remove them, especially in what concerns saltwater effluents. As such, this work addresses two main scientific questions. One of them regards the understanding of the actual (bio)remediation methods applied to effluents produced in marine RAS, by identifying their advantages, drawbacks and gaps concerning their exploitation in saltwater effluents. The second one is the development of a new, innovative and efficient method for the treatment of saltwater effluents that potentially fulfil the gaps identified in the conventional treatments. Thereby, the aims of this thesis are: (i) to revise the conventional treatments targeting major contaminants in marine RAS effluents, with a particular focus on the bioremediation approaches already conducted for phosphates; (ii) to characterize and evaluate the potential of oyster-shell waste collected in Ria de Aveiro as a bioremediation agent of phosphates spiked into artificial saltwater, over different influencing factors (e.g., oyster-shell pre-treatment through calcination, particle size, adsorbent concentration). Despite the use of oyster-shells for phosphorous (P) removal has already been applied in freshwater, its biosorptive potential for P in saltwater was never evaluated, as far as I am aware. The results herein generated showed that NOS is mainly composed by carbonates, which are almost completely converted into lime (CaO) after calcination (COS). Such pre-treatment allowed obtaining a more reactive material for P removal, since higher removal percentages and adsorption capacity was observed for COS. Smaller particle size fractions for both NOS and COS samples also increased P removal. Kinetic models showed that NOS adsorption followed, simultaneously, Elovich and Intraparticle Difusion kinetic models, suggesting that P removal is both a diffusional and chemically rate-controlled process. The percentage of P removal by COS was not controlled by Intraparticle Diffusion and the Elovich model was the kinetic model that best fitted phosphate removal. This work demonstrated that waste oyster-shells, either NOS or COS, could be used as an effective biosorbent for P removal from seawater. Thereby, this biomaterial can sustain a cost-effective and eco-friendly bioremediation strategy with potential application in marine RAS.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Head and Neck Cancers (HNC) are a group of tumours located in the upper aero-digestive tract. Head and Neck Squamous Cell Carcinoma (HNSCC) represent about 90% of all HNC cases. It has been considered the sixth most malignant tumour worldwide and, despite clinical and technological advances, the five-year survival rate has not improved much in the last years. Nowadays, HNSCC is well established as a heterogeneous disease and that its development is due to accumulation of genetic events. Apart from the majority of the patients being diagnosed in an advanced stage, HNSCC is also a disease with poor therapeutic outcome. One of the therapeutic approaches is radiotherapy. However, this approach has different drawbacks like the radioresistance acquired by some tumour cells, leading to a worse prognosis. A major knowledge in radiation biology is imperative to improve this type of treatment and avoid late toxicities, maintaining patient quality of life in the subsequent years after treatment. Then, identification of genetic markers associated to radiotherapy response in patients and possible alterations in cells after radiotherapy are essential steps towards an improved diagnosis, higher survival rate and a better life quality. Not much is known about the radiation effects on cells, so, the principal aim of this study was to contribute to a more extensive knowledge about radiation treatment in HNSCC. For this, two commercial cell lines, HSC-3 and BICR-10, were used and characterized resorting to karyotyping, aCGH and MS-MLPA. These cell lines were submitted to different doses of irradiation and the resulting genetic and methylation alterations were evaluated. Our results showed a great difference in radiation response between the two cell lines, allowing the conclusion that HSC-3 was much more radiosensitive than BICR-10. Bearing this in mind, analysis of cell death, cell cycle and DNA damages was performed to try to elucidate the motifs behind this difference. The characterization of both cell lines allowed the confirmation that HSC-3 was derived from a metastatic tumour and the hypothesis that BICR-10 was derived from a dysplasia. Furthermore, this pilot study enabled the suggestion of some genetic and epigenetic alterations that cells suffer after radiation treatment. Additionally, it also allowed the association of some genetic characteristics that could be related to the differences in radiation response observable in this two cell lines. Taken together all of our results contribute to a better understanding of radiation effects on HNSCC allowing one further step towards the prediction of patients’ outcome, better choice of treatment approaches and ultimately a better quality of life.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Call Level Interfaces (CLI) play a key role in business tiers of relational and on some NoSQL database applications whenever a fine tune control between application tiers and the host databases is a key requirement. Unfortunately, in spite of this significant advantage, CLI are low level API, this way not addressing high level architectural requirements. Among the examples we emphasize two situations: a) the need to decouple or not to decouple the development process of business tiers from the development process of application tiers and b) the need to automatically adapt business tiers to new business and/or security needs at runtime. To tackle these CLI drawbacks, and simultaneously keep their advantages, this paper proposes an architecture relying on CLI from which multi-purpose business tiers components are built, herein referred to as Adaptable Business Tier Components (ABTC). Beyond the reference architecture, this paper presents a proof of concept based on Java and Java Database Connectivity (an example of CLI).

Relevância:

10.00% 10.00%

Publicador:

Resumo:

In database applications, access control security layers are mostly developed from tools provided by vendors of database management systems and deployed in the same servers containing the data to be protected. This solution conveys several drawbacks. Among them we emphasize: 1) if policies are complex, their enforcement can lead to performance decay of database servers; 2) when modifications in the established policies implies modifications in the business logic (usually deployed at the client-side), there is no other possibility than modify the business logic in advance and, finally, 3) malicious users can issue CRUD expressions systematically against the DBMS expecting to identify any security gap. In order to overcome these drawbacks, in this paper we propose an access control stack characterized by: most of the mechanisms are deployed at the client-side; whenever security policies evolve, the security mechanisms are automatically updated at runtime and, finally, client-side applications do not handle CRUD expressions directly. We also present an implementation of the proposed stack to prove its feasibility. This paper presents a new approach to enforce access control in database applications, this way expecting to contribute positively to the state of the art in the field.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Call Level Interfaces (CLI) are low level API that play a key role in database applications whenever a fine tune control between application tiers and the host databases is a key requirement. Unfortunately, in spite of this significant advantage, CLI were not designed to address organizational requirements and contextual runtime requirements. Among the examples we emphasize the need to decouple or not to decouple the development process of business tiers from the development process of application tiers and also the need to automatically adapt to new business and/or security needs at runtime. To tackle these CLI drawbacks, and simultaneously keep their advantages, this paper proposes an architecture relying on CLI from which multi-purpose business tiers components are built, herein referred to as Adaptable Business Tier Components (ABTC). This paper presents the reference architecture for those components and a proof of concept based on Java and Java Database Connectivity (an example of CLI).

Relevância:

10.00% 10.00%

Publicador:

Resumo:

In database applications, access control security layers are mostly developed from tools provided by vendors of database management systems and deployed in the same servers containing the data to be protected. This solution conveys several drawbacks. Among them we emphasize: (1) if policies are complex, their enforcement can lead to performance decay of database servers; (2) when modifications in the established policies implies modifications in the business logic (usually deployed at the client-side), there is no other possibility than modify the business logic in advance and, finally, 3) malicious users can issue CRUD expressions systematically against the DBMS expecting to identify any security gap. In order to overcome these drawbacks, in this paper we propose an access control stack characterized by: most of the mechanisms are deployed at the client-side; whenever security policies evolve, the security mechanisms are automatically updated at runtime and, finally, client-side applications do not handle CRUD expressions directly. We also present an implementation of the proposed stack to prove its feasibility. This paper presents a new approach to enforce access control in database applications, this way expecting to contribute positively to the state of the art in the field.