1000 resultados para Multi-modelos
Resumo:
Revisão bibliográfica que descreve as taxas de cesárea em diferentes países e os modelos de atenção ao parto de acordo com o uso de tecnologias assistenciais. Foram analisados 60 estudos publicados entre 1999 e 2010, obtidos nas bases de dados da Coordenação de Aperfeiçoamento de Pessoal de Nível Superior e ProQuest. O modelo de assistência obstétrica praticado no país baseia-se na relação médico-paciente, ao grau de utilização de tecnologias e à realização do parto cesáreo.
Resumo:
Nas últimas décadas, o estudo das séries temporais de contagem univariadas tem sido objecto de interesse crescente na literatura. Uma das classes de modelos mais populares é a dos modelos auto-regressivos e médias móveis de valor inteiro não-negativo, INARMA, obtida através da substituição da multiplicação por um operador aleatório, chamado thinning, nos modelos ARMA convencionais. Os modelos INAR para séries de contagem univariadas têm sido amplamente estudados na literatura no que diz respeito quer às suas propriedades probabilísticas quer à inferência estatística. Por outro lado, a classe alargada de modelos de médias móveis para séries de valor inteiro, INMA, em que as operações thinning em instantes diferentes são dependentes entre si, não tem sido objecto de tanta atenção. De facto os modelos INMA não são Markovianos pelo que a inferência estatística apresenta dificuldades adicionais. Actualmente o interesse na análise de séries temporais de contagem centra-se em modelos e métodos para séries multivariadas. Neste trabalho, consideram-se os modelos INMA bivariados propostos por Torres et al. (2012). Nesta classe alargada de modelos bivariados INMA, a estrutura de dependência entre as duas séries temporais é introduzia pela dependência entre os dois processos de chegada enquanto que a dependência em cada série é definida pela dependência de operações thinning em instantes diferentes. Consideram-se estimadores baseados em momentos: método dos momentos (MM), método dos momentos generalizados (GMM) e método dos momentos eficiente (EMM), assim como estimadores baseados na função geradora de probabilidades.
Resumo:
São apresentados os modelos cognitivos da tonalidade desenvolvidos por Shepard (1964; 1982a), Longuet- Higgins (1987), Lerdahl (1988; 2001), e Krumhansl (1983; 1990), com particular incidência no modelo de Lerdahl. Um breve exame comparativo evidencia as potencialidades deste último como instrumento de investigação no âmbito dos estudos de performance bem como na área da percepção. Palavras-chave: Percepção; Tonalidade; Lerdahl; Longuet-Higgins; Krumhansl; Shepard.
Resumo:
Multi-objective particle swarm optimization (MOPSO) is a search algorithm based on social behavior. Most of the existing multi-objective particle swarm optimization schemes are based on Pareto optimality and aim to obtain a representative non-dominated Pareto front for a given problem. Several approaches have been proposed to study the convergence and performance of the algorithm, particularly by accessing the final results. In the present paper, a different approach is proposed, by using Shannon entropy to analyzethe MOPSO dynamics along the algorithm execution. The results indicate that Shannon entropy can be used as an indicator of diversity and convergence for MOPSO problems.
Resumo:
Atualmente o sistema produtivo do tipo job shop é muito comum nas PMEs (Pequenas e Médias Empresas). Estas empresas trabalham por encomenda. Produzem grande variedade de modelos, e em pequenas quantidades. Os prazos de entrega são um fator de elevada importância, pois os clientes exigem um produto de qualidade no tempo certo. O presente trabalho, pretende criar uma ferramenta de programação da produção para a secção da costura, usando dados reais da empresa, que tem uma implantação do tipo job shop com máquinas multi-operação (Multi-Purpose -Machines Job Shop). No final, são reunidas as principais conclusões e perspetivados futuros desenvolvimentos. Os resultados obtidos comprovam que o algoritmo desenvolvido, com base no algoritmo de Giffler & Thompson, consegue obter com grande precisão e de forma rápida o escalonamento / balanceamento da secção da costura. Com a ferramenta criada, a empresa otimiza a programação da secção da costura e fornece informação importante á gestão da produção, possibilitando uma melhoria do planeamento da empresa.
Resumo:
Dissertação de Natureza Científica elabora da no âmbito do protocolo de cooperação entre o ISEL e o LNEC para obtenção do grau de Mestre em Engenharia Civil
Resumo:
This study focused on the development of a sensitive enzymatic biosensor for the determination of pirimicarb pesticide based on the immobilization of laccase on composite carbon paste electrodes. Multi- walled carbon nanotubes(MWCNTs)paste electrode modified by dispersion of laccase(3%,w/w) within the optimum composite matrix(60:40%,w/w,MWCNTs and paraffin binder)showed the best performance, with excellent electron transfer kinetic and catalytic effects related to the redox process of the substrate4- aminophenol. No metal or anti-interference membrane was added. Based on the inhibition of laccase activity, pirimicarb can be determined in the range 9.90 ×10- 7 to 1.15 ×10- 5 molL 1 using 4- aminophenol as substrate at the optimum pH of 5.0, with acceptable repeatability and reproducibility (relative standard deviations lower than 5%).The limit of detection obtained was 1.8 × 10-7 molL 1 (0.04 mgkg 1 on a fresh weight vegetable basis).The high activity and catalytic properties of the laccase- based biosensor are retained during ca. one month. The optimized electroanalytical protocol coupled to the QuEChERS methodology were applied to tomato and lettuce samples spiked at three levels; recoveries ranging from 91.0±0.1% to 101.0 ± 0.3% were attained. No significant effects in the pirimicarb electro- analysis were observed by the presence of pro-vitamin A, vitamins B1 and C,and glucose in the vegetable extracts. The proposed biosensor- based pesticide residue methodology fulfills all requisites to be used in implementation of food safety programs.
Resumo:
Mestrado em Controlo de Gestão e dos Negócios
Resumo:
Fieldbus communication networks aim to interconnect sensors, actuators and controllers within process control applications. Therefore, they constitute the foundation upon which real-time distributed computer-controlled systems can be implemented. P-NET is a fieldbus communication standard, which uses a virtual token-passing medium-access-control mechanism. In this paper pre-run-time schedulability conditions for supporting real-time traffic with P-NET networks are established. Essentially, formulae to evaluate the upper bound of the end-to-end communication delay in P-NET messages are provided. Using this upper bound, a feasibility test is then provided to check the timing requirements for accessing remote process variables. This paper also shows how P-NET network segmentation can significantly reduce the end-to-end communication delays for messages with stringent timing requirements.
Resumo:
Trabalho Final de Mestrado para obtenção do grau de Mestre em Engenharia de Redes de Comunicação e Multimédia
Resumo:
Trabalho Final de Mestrado para obtenção do grau de Mestre em Engenharia Civil na Área de Especialização de Edificações
Resumo:
Dissertação de mestrado em Ciências de Educação: área de Educação e Desenvolvimento
Resumo:
This paper presents an architecture (Multi-μ) being implemented to study and develop software based fault tolerant mechanisms for Real-Time Systems, using the Ada language (Ada 95) and Commercial Off-The-Shelf (COTS) components. Several issues regarding fault tolerance are presented and mechanisms to achieve fault tolerance by software active replication in Ada 95 are discussed. The Multi-μ architecture, based on a specifically proposed Fault Tolerance Manager (FTManager), is then described. Finally, some considerations are made about the work being done and essential future developments.
Resumo:
Due to usage conditions, hazardous environments or intentional causes, physical and virtual systems are subject to faults in their components, which may affect their overall behaviour. In a ‘black-box’ agent modelled by a set of propositional logic rules, in which just a subset of components is externally visible, such faults may only be recognised by examining some output function of the agent. A (fault-free) model of the agent’s system provides the expected output given some input. If the real output differs from that predicted output, then the system is faulty. However, some faults may only become apparent in the system output when appropriate inputs are given. A number of problems regarding both testing and diagnosis thus arise, such as testing a fault, testing the whole system, finding possible faults and differentiating them to locate the correct one. The corresponding optimisation problems of finding solutions that require minimum resources are also very relevant in industry, as is minimal diagnosis. In this dissertation we use a well established set of benchmark circuits to address such diagnostic related problems and propose and develop models with different logics that we formalise and generalise as much as possible. We also prove that all techniques generalise to agents and to multiple faults. The developed multi-valued logics extend the usual Boolean logic (suitable for faultfree models) by encoding values with some dependency (usually on faults). Such logics thus allow modelling an arbitrary number of diagnostic theories. Each problem is subsequently solved with CLP solvers that we implement and discuss, together with a new efficient search technique that we present. We compare our results with other approaches such as SAT (that require substantial duplication of circuits), showing the effectiveness of constraints over multi-valued logics, and also the adequacy of a general set constraint solver (with special inferences over set functions such as cardinality) on other problems. In addition, for an optimisation problem, we integrate local search with a constructive approach (branch-and-bound) using a variety of logics to improve an existing efficient tool based on SAT and ILP.
Resumo:
Modern multicore processors for the embedded market are often heterogeneous in nature. One feature often available are multiple sleep states with varying transition cost for entering and leaving said sleep states. This research effort explores the energy efficient task-mapping on such a heterogeneous multicore platform to reduce overall energy consumption of the system. This is performed in the context of a partitioned scheduling approach and a very realistic power model, which improves over some of the simplifying assumptions often made in the state-of-the-art. The developed heuristic consists of two phases, in the first phase, tasks are allocated to minimise their active energy consumption, while the second phase trades off a higher active energy consumption for an increased ability to exploit savings through more efficient sleep states. Extensive simulations demonstrate the effectiveness of the approach.