989 resultados para multi-constraint assignment


Relevância:

20.00% 20.00%

Publicador:

Resumo:

O Acidente Vascular Encefálico é uma das principais causas de morte, tornando-se cada vez mais iminente processos de reabilitação que minimizem as sequelas, nomeadamente as limitações do membro superior que dificultam o envolvimento em atividades da vida diária. O Constraint-Induced Movement Therapy, surge como uma abordagem que incrementa o uso do membro superior mais afetado. A presente investigação trata-se de um estudo de casos múltiplos. Pretende-se verificar se existem melhorias na funcionalidade do membro superior mais afetado, analisar em que atividades da vida diária são visíveis melhorias funcionais e compreender se o maior envolvimento nas atividades diárias está diretamente relacionado com a melhoria na capacidade funcional. Pretende-se ainda que os valores obtidos no Wolf Motor Function Test sejam um contributo para a sua validação para a população portuguesa. Utilizou-se um questionário para recolha de dados pessoais e clínicos (amplitudes de movimento, dor e espasticidade); o Wolf Motor Function Test e o Action Research Arm Test para verificar a funcionalidade do membro superior mais afetado; e a Motor Activity Log que avalia o envolvimento em atividades da vida diária. O grupo é constituído por 3 utentes que sofreram um primeiro Acidente Vascular Encefálico até 9 meses de evolução, internados na Santa Casa da Misericórdia de Monção e que cumpriam os critérios de inclusão. O programa foi implementado três horas/dia, durante 10 dias, mantendo a restrição no membro superior menos afetado durante 90% do dia acordado. Como se trata de um estudo de casos múltiplos, analisou-se cada participante individualmente e verificou-se a diferença entre os resultados finais e iniciais para cada uma das variáveis. Os resultados obtidos revelam ganhos na amplitude de movimento, velocidade de execução e capacidade funcional do membro superior mais afetado, nomeadamente nas funções de preensão e pinça da mão, bem como se testemunhou minimização do fenómeno learned nonuse. Verificaram-se ganhos funcionais em todos os participantes nas atividades da vida diária apesar de serem diferentes de participante para participante. Dois participantes afirmaram que voltariam a participar no programa.Conclui-se, assim que a técnica resulta em ganhos funcionais nestes utentes, indicando um caminho alternativo a outras abordagens de reabilitação.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Este trabalho visa apresentar um enquadramento da realidade económica e industrial do sector transformador de granitos ornamentais em Portugal e fazer uma análise do processo de serragem, com engenhos multi-lâminas e granalha de aço, na medida em que este é o método de seccionamento de blocos de granito mais utilizado pelas grandes indústrias do sector. Tendo em conta a importância económica desta operação produtiva na indústria em causa, foi definido como fito deste projecto a análise estatística dos custos de produção; a definição de fórmulas de cálculo que permitam prever o custo médio de serragem; e o estudo de soluções economicamente viáveis e ambientalmente sustentáveis para o problema das lamas resultantes do expurgo dos engenhos. Para a persecução deste projecto foi realizada uma recolha de dados implementando rotinas de controlo e registo dos mesmos, em quadros de produção normalizados e de fácil preenchimento, pelos operadores destes equipamentos. Esta recolha de dados permitiu isolar, quantificar e formular os factores de rentabilização do processo de serragem selecionando, dentro da amostra de estudo obtida, um conjunto de serragens com características similares e com valores próximos dos valores da média estatística. Apartir dos dados destas serragens foram geradas curvas de tendência polinomial com as quais se analisaram as variações provocadas no custo médio de serragem, pelas variações do factor em estudo. A formulação dos factores de rentabilização e os dados estatísticos obtidos permitiram depois o desenvolvimento de fórmulas de cálculo do custo médio de serragem que establecem o custo de produção diferenciado em função das espessuras com, ou sem, a incorporação dos factores de rentabilização. Como consequência do projecto realizado obteve-se um conjunto de conclusões util, para o sector industrial em causa, que evidencia a importancia da Ocupação dos engenhos e rentabilização de um espaço confinado, da Resistência oferecida à serragem pelos granitos, e da Diferença de altura entre os blocos de uma mesma carga, nos custos de transformação.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

A new high performance architecture for the computation of all the DCT operations adopted in the H.264/AVC and HEVC standards is proposed in this paper. Contrasting to other dedicated transform cores, the presented multi-standard transform architecture is supported on a completely configurable, scalable and unified structure, that is able to compute not only the forward and the inverse 8×8 and 4×4 integer DCTs and the 4×4 and 2×2 Hadamard transforms defined in the H.264/AVC standard, but also the 4×4, 8×8, 16×16 and 32×32 integer transforms adopted in HEVC. Experimental results obtained using a Xilinx Virtex-7 FPGA demonstrated the superior performance and hardware efficiency levels provided by the proposed structure, which outperforms its more prominent related designs by at least 1.8 times. When integrated in a multi-core embedded system, this architecture allows the computation, in real-time, of all the transforms mentioned above for resolutions as high as the 8k Ultra High Definition Television (UHDTV) (7680×4320 @ 30fps).

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This chapter aims to demonstrate how PAOL - Unit for Innovation in Education, a project from ISCAP - School of Accounting and Administration of Oporto ....

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Solvent extraction is considered as a multi-criteria optimization problem, since several chemical species with similar extraction kinetic properties are frequently present in the aqueous phase and the selective extraction is not practicable. This optimization, applied to mixer–settler units, considers the best parameters and operating conditions, as well as the best structure or process flow-sheet. Global process optimization is performed for a specific flow-sheet and a comparison of Pareto curves for different flow-sheets is made. The positive weight sum approach linked to the sequential quadratic programming method is used to obtain the Pareto set. In all investigated structures, recovery increases with hold-up, residence time and agitation speed, while the purity has an opposite behaviour. For the same treatment capacity, counter-current arrangements are shown to promote recovery without significant impairment in purity. Recycling the aqueous phase is shown to be irrelevant, but organic recycling with as many stages as economically feasible clearly improves the design criteria and reduces the most efficient organic flow-rate.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Multi-objective particle swarm optimization (MOPSO) is a search algorithm based on social behavior. Most of the existing multi-objective particle swarm optimization schemes are based on Pareto optimality and aim to obtain a representative non-dominated Pareto front for a given problem. Several approaches have been proposed to study the convergence and performance of the algorithm, particularly by accessing the final results. In the present paper, a different approach is proposed, by using Shannon entropy to analyzethe MOPSO dynamics along the algorithm execution. The results indicate that Shannon entropy can be used as an indicator of diversity and convergence for MOPSO problems.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This study focused on the development of a sensitive enzymatic biosensor for the determination of pirimicarb pesticide based on the immobilization of laccase on composite carbon paste electrodes. Multi- walled carbon nanotubes(MWCNTs)paste electrode modified by dispersion of laccase(3%,w/w) within the optimum composite matrix(60:40%,w/w,MWCNTs and paraffin binder)showed the best performance, with excellent electron transfer kinetic and catalytic effects related to the redox process of the substrate4- aminophenol. No metal or anti-interference membrane was added. Based on the inhibition of laccase activity, pirimicarb can be determined in the range 9.90 ×10- 7 to 1.15 ×10- 5 molL 1 using 4- aminophenol as substrate at the optimum pH of 5.0, with acceptable repeatability and reproducibility (relative standard deviations lower than 5%).The limit of detection obtained was 1.8 × 10-7 molL 1 (0.04 mgkg 1 on a fresh weight vegetable basis).The high activity and catalytic properties of the laccase- based biosensor are retained during ca. one month. The optimized electroanalytical protocol coupled to the QuEChERS methodology were applied to tomato and lettuce samples spiked at three levels; recoveries ranging from 91.0±0.1% to 101.0 ± 0.3% were attained. No significant effects in the pirimicarb electro- analysis were observed by the presence of pro-vitamin A, vitamins B1 and C,and glucose in the vegetable extracts. The proposed biosensor- based pesticide residue methodology fulfills all requisites to be used in implementation of food safety programs.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The relation between the information/knowledge expression and the physical expression can be involved as one of items for an ambient intelligent computing [2],[3]. Moreover, because there are so many contexts around user/spaces during a user movement, all appplcation which are using AmI for users are based on the relation between user devices and environments. In these situations, it is possible that the AmI may output the wrong result from unreliable contexts by attackers. Recently, establishing a server have been utilizes, so finding secure contexts and make contexts of higher security level for save communication have been given importance. Attackers try to put their devices on the expected path of all users in order to obtain users informationillegally or they may try to broadcast their SPAMS to users. This paper is an extensionof [11] which studies the Security Grade Assignment Model (SGAM) to set Cyber-Society Organization (CSO).

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Fieldbus communication networks aim to interconnect sensors, actuators and controllers within process control applications. Therefore, they constitute the foundation upon which real-time distributed computer-controlled systems can be implemented. P-NET is a fieldbus communication standard, which uses a virtual token-passing medium-access-control mechanism. In this paper pre-run-time schedulability conditions for supporting real-time traffic with P-NET networks are established. Essentially, formulae to evaluate the upper bound of the end-to-end communication delay in P-NET messages are provided. Using this upper bound, a feasibility test is then provided to check the timing requirements for accessing remote process variables. This paper also shows how P-NET network segmentation can significantly reduce the end-to-end communication delays for messages with stringent timing requirements.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Trabalho Final de Mestrado para obtenção do grau de Mestre em Engenharia de Redes de Comunicação e Multimédia

Relevância:

20.00% 20.00%

Publicador:

Resumo:

It is generally challenging to determine end-to-end delays of applications for maximizing the aggregate system utility subject to timing constraints. Many practical approaches suggest the use of intermediate deadline of tasks in order to control and upper-bound their end-to-end delays. This paper proposes a unified framework for different time-sensitive, global optimization problems, and solves them in a distributed manner using Lagrangian duality. The framework uses global viewpoints to assign intermediate deadlines, taking resource contention among tasks into consideration. For soft real-time tasks, the proposed framework effectively addresses the deadline assignment problem while maximizing the aggregate quality of service. For hard real-time tasks, we show that existing heuristic solutions to the deadline assignment problem can be incorporated into the proposed framework, enriching their mathematical interpretation.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This paper presents an architecture (Multi-μ) being implemented to study and develop software based fault tolerant mechanisms for Real-Time Systems, using the Ada language (Ada 95) and Commercial Off-The-Shelf (COTS) components. Several issues regarding fault tolerance are presented and mechanisms to achieve fault tolerance by software active replication in Ada 95 are discussed. The Multi-μ architecture, based on a specifically proposed Fault Tolerance Manager (FTManager), is then described. Finally, some considerations are made about the work being done and essential future developments.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The basic motivation of this work was the integration of biophysical models within the interval constraints framework for decision support. Comparing the major features of biophysical models with the expressive power of the existing interval constraints framework, it was clear that the most important inadequacy was related with the representation of differential equations. System dynamics is often modelled through differential equations but there was no way of expressing a differential equation as a constraint and integrate it within the constraints framework. Consequently, the goal of this work is focussed on the integration of ordinary differential equations within the interval constraints framework, which for this purpose is extended with the new formalism of Constraint Satisfaction Differential Problems. Such framework allows the specification of ordinary differential equations, together with related information, by means of constraints, and provides efficient propagation techniques for pruning the domains of their variables. This enabled the integration of all such information in a single constraint whose variables may subsequently be used in other constraints of the model. The specific method used for pruning its variable domains can then be combined with the pruning methods associated with the other constraints in an overall propagation algorithm for reducing the bounds of all model variables. The application of the constraint propagation algorithm for pruning the variable domains, that is, the enforcement of local-consistency, turned out to be insufficient to support decision in practical problems that include differential equations. The domain pruning achieved is not, in general, sufficient to allow safe decisions and the main reason derives from the non-linearity of the differential equations. Consequently, a complementary goal of this work proposes a new strong consistency criterion, Global Hull-consistency, particularly suited to decision support with differential models, by presenting an adequate trade-of between domain pruning and computational effort. Several alternative algorithms are proposed for enforcing Global Hull-consistency and, due to their complexity, an effort was made to provide implementations able to supply any-time pruning results. Since the consistency criterion is dependent on the existence of canonical solutions, it is proposed a local search approach that can be integrated with constraint propagation in continuous domains and, in particular, with the enforcing algorithms for anticipating the finding of canonical solutions. The last goal of this work is the validation of the approach as an important contribution for the integration of biophysical models within decision support. Consequently, a prototype application that integrated all the proposed extensions to the interval constraints framework is developed and used for solving problems in different biophysical domains.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Modern multicore processors for the embedded market are often heterogeneous in nature. One feature often available are multiple sleep states with varying transition cost for entering and leaving said sleep states. This research effort explores the energy efficient task-mapping on such a heterogeneous multicore platform to reduce overall energy consumption of the system. This is performed in the context of a partitioned scheduling approach and a very realistic power model, which improves over some of the simplifying assumptions often made in the state-of-the-art. The developed heuristic consists of two phases, in the first phase, tasks are allocated to minimise their active energy consumption, while the second phase trades off a higher active energy consumption for an increased ability to exploit savings through more efficient sleep states. Extensive simulations demonstrate the effectiveness of the approach.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Consider the problem of assigning real-time tasks on a heterogeneous multiprocessor platform comprising two different types of processors — such a platform is referred to as two-type platform. We present two linearithmic timecomplexity algorithms, SA and SA-P, each providing the follow- ing guarantee. For a given two-type platform and a given task set, if there exists a feasible task-to-processor-type assignment such that tasks can be scheduled to meet deadlines by allowing them to migrate only between processors of the same type, then (i) using SA, it is guaranteed to find such a feasible task-to- processor-type assignment where the same restriction on task migration applies but given a platform in which processors are 1+α/2 times faster and (ii) SA-P succeeds in finding 2 a feasible task-to-processor assignment where tasks are not allowed to migrate between processors but given a platform in which processors are 1+α/times faster, where 0<α≤1. The parameter α is a property of the task set — it is the maximum utilization of any task which is less than or equal to 1.