952 resultados para Distributed Lag Non-linear Models


Relevância:

100.00% 100.00%

Publicador:

Resumo:

In health related research it is common to have multiple outcomes of interest in a single study. These outcomes are often analysed separately, ignoring the correlation between them. One would expect that a multivariate approach would be a more efficient alternative to individual analyses of each outcome. Surprisingly, this is not always the case. In this article we discuss different settings of linear models and compare the multivariate and univariate approaches. We show that for linear regression models, the estimates of the regression parameters associated with covariates that are shared across the outcomes are the same for the multivariate and univariate models while for outcome-specific covariates the multivariate model performs better in terms of efficiency.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

In this work, kriging with covariates is used to model and map the spatial distribution of salinity measurements gathered by an autonomous underwater vehicle in a sea outfall monitoring campaign aiming to distinguish the effluent plume from the receiving waters and characterize its spatial variability in the vicinity of the discharge. Four different geostatistical linear models for salinity were assumed, where the distance to diffuser, the west-east positioning, and the south-north positioning were used as covariates. Sample variograms were fitted by the Mat`ern models using weighted least squares and maximum likelihood estimation methods as a way to detect eventual discrepancies. Typically, the maximum likelihood method estimated very low ranges which have limited the kriging process. So, at least for these data sets, weighted least squares showed to be the most appropriate estimation method for variogram fitting. The kriged maps show clearly the spatial variation of salinity, and it is possible to identify the effluent plume in the area studied. The results obtained show some guidelines for sewage monitoring if a geostatistical analysis of the data is in mind. It is important to treat properly the existence of anomalous values and to adopt a sampling strategy that includes transects parallel and perpendicular to the effluent dispersion.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

A crescente necessidade de reduzir a dependência energética e a emissão de gases de efeito de estufa levou à adoção de uma série de políticas a nível europeu com vista a aumentar a eficiência energética e nível de controlo de equipamentos, reduzir o consumo e aumentar a percentagem de energia produzida a partir de fontes renováveis. Estas medidas levaram ao desenvolvimento de duas situações críticas para o setor elétrico: a substituição das cargas lineares tradicionais, pouco eficientes, por cargas não-lineares mais eficientes e o aparecimento da produção distribuída de energia a partir de fontes renováveis. Embora apresentem vantagens bem documentadas, ambas as situações podem afetar negativamente a qualidade de energia elétrica na rede de distribuição, principalmente na rede de baixa tensão onde é feita a ligação com a maior parte dos clientes e onde se encontram as cargas não-lineares e a ligação às fontes de energia descentralizadas. Isto significa que a monitorização da qualidade de energia tem, atualmente, uma importância acrescida devido aos custos relacionados com perdas inerentes à falta de qualidade de energia elétrica na rede e à necessidade de verificar que determinados parâmetros relacionados com a qualidade de energia elétrica se encontram dentro dos limites previstos nas normas e nos contratos com clientes de forma a evitar disputas ou reclamações. Neste sentido, a rede de distribuição tem vindo a sofrer alterações a nível das subestações e dos postos de transformação que visam aumentar a visibilidade da qualidade de energia na rede em tempo real. No entanto, estas medidas só permitem monitorizar a qualidade de energia até aos postos de transformação de média para baixa tensão, não revelando o estado real da qualidade de energia nos pontos de entrega ao cliente. A monitorização nestes pontos é feita periodicamente e não em tempo real, ficando aquém do necessário para assegurar a deteção correta de problemas de qualidade de energia no lado do consumidor. De facto, a metodologia de monitorização utilizada atualmente envolve o envio de técnicos ao local onde surgiu uma reclamação ou a um ponto de medição previsto para instalar um analisador de energia que permanece na instalação durante um determinado período de tempo. Este tipo de monitorização à posteriori impossibilita desde logo a deteção do problema de qualidade de energia que levou à reclamação, caso não se trate de um problema contínuo. Na melhor situação, o aparelho poderá detetar uma réplica do evento, mas a larga percentagem anomalias ficam fora deste processo por serem extemporâneas. De facto, para detetar o evento que deu origem ao problema é necessário monitorizar permanentemente a qualidade de energia. No entanto este método de monitorização implica a instalação permanente de equipamentos e não é viável do ponto de vista das empresas de distribuição de energia já que os equipamentos têm custos demasiado elevados e implicam a necessidade de espaços maiores nos pontos de entrega para conter os equipamentos e o contador elétrico. Uma alternativa possível que pode tornar viável a monitorização permanente da qualidade de energia consiste na introdução de uma funcionalidade de monitorização nos contadores de energia de determinados pontos da rede de distribuição. Os contadores são obrigatórios em todas as instalações ligadas à rede, para efeitos de faturação. Tradicionalmente estes contadores são eletromecânicos e recentemente começaram a ser substituídos por contadores inteligentes (smart meters), de natureza eletrónica, que para além de fazer a contagem de energia permitem a recolha de informação sobre outros parâmetros e aplicação de uma serie de funcionalidades pelo operador de rede de distribuição devido às suas capacidades de comunicação. A reutilização deste equipamento com finalidade de analisar a qualidade da energia junto dos pontos de entrega surge assim como uma forma privilegiada dado que se trata essencialmente de explorar algumas das suas características adicionais. Este trabalho tem como objetivo analisar a possibilidade descrita de monitorizar a qualidade de energia elétrica de forma permanente no ponto de entrega ao cliente através da utilização do contador elétrico do mesmo e elaborar um conjunto de requisitos para o contador tendo em conta a normalização aplicável, as características dos equipamentos utilizados atualmente pelo operador de rede e as necessidades do sistema elétrico relativamente à monitorização de qualidade de energia.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Neste documento, são investigados vários métodos usados na inteligência artificial, com o objetivo de obter previsões precisas da evolução dos mercados financeiros. O uso de ferramentas lineares como os modelos AR, MA, ARMA e GARCH têm muitas limitações, pois torna-se muito difícil adaptá-los às não linearidades dos fenómenos que ocorrem nos mercados. Pelas razões anteriormente referidas, os algoritmos como as redes neuronais dinâmicas (TDNN, NARX e ESN), mostram uma maior capacidade de adaptação a estas não linearidades, pois não fazem qualquer pressuposto sobre as distribuições de probabilidade que caracterizam estes mercados. O facto destas redes neuronais serem dinâmicas, faz com que estas exibam um desempenho superior em relação às redes neuronais estáticas, ou outros algoritmos que não possuem qualquer tipo de memória. Apesar das vantagens reveladas pelas redes neuronais, estas são um sistema do tipo black box, o que torna muito difícil extrair informação dos pesos da rede. Isto significa que estes algoritmos devem ser usados com precaução, pois podem tornar-se instáveis.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Dissertação para obtenção do Grau de Doutor em Engenharia Civil

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Dissertação para obtenção do Grau de Mestre em Engenharia Electrotécnica e de Computadores

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Rupture of aortic aneurysms (AA) is a major cause of death in the Western world. Currently, clinical decision upon surgical intervention is based on the diameter of the aneurysm. However, this method is not fully adequate. Noninvasive assessment of the elastic properties of the arterial wall can be a better predictor for AA growth and rupture risk. The purpose of this study is to estimate mechanical properties of the aortic wall using in vitro inflation testing and 2D ultrasound (US) elastography, and investigate the performance of the proposed methodology for physiological conditions. Two different inflation experiments were performed on twelve porcine aortas: 1) a static experiment for a large pressure range (0 – 140 mmHg); 2) a dynamic experiment closely mimicking the in vivo hemodynamics at physiological pressures (70 – 130 mmHg). 2D raw radiofrequency (RF) US datasets were acquired for one longitudinal and two cross-sectional imaging planes, for both experiments. The RF-data were manually segmented and a 2D vessel wall displacement tracking algorithm was applied to obtain the aortic diameter–time behavior. The shear modulus G was estimated assuming a Neo-Hookean material model. In addition, an incremental study based on the static data was performed to: 1) investigate the changes in G for increasing mean arterial pressure (MAP), for a certain pressure difference (30, 40, 50 and 60 mmHg); 2) compare the results with those from the dynamic experiment, for the same pressure range. The resulting shear modulus G was 94 ± 16 kPa for the static experiment, which is in agreement with literature. A linear dependency on MAP was found for G, yet the effect of the pressure difference was negligible. The dynamic data revealed a G of 250 ± 20 kPa. For the same pressure range, the incremental shear modulus (Ginc) was 240 ± 39 kPa, which is in agreement with the former. In general, for all experiments, no significant differences in the values of G were found between different image planes. This study shows that 2D US elastography of aortas during inflation testing is feasible under controlled and physiological circumstances. In future studies, the in vivo, dynamic experiment should be repeated for a range of MAPs and pathological vessels should be examined. Furthermore, the use of more complex material models needs to be considered to describe the non-linear behavior of the vascular tissue.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The main purpose of the present dissertation is the simulation of the response of fibre grout strengthened RC panels when subjected to blast effects using the Applied Element Method, in order to validate and verify its applicability. Therefore, four experimental models, three of which were strengthened with a cement-based grout, each reinforced by one type of steel reinforcement, were tested against blast effects. After the calibration of the experimental set-up, it was possible to obtain and compare the response to the blast effects of the model without strengthening (reference model), and a fibre grout strengthened RC panel (strengthened model). Afterwards, a numerical model of the reference model was created in the commercial software Extreme Loading for Structures, which is based on the Applied Element Method, and calibrated to the obtained experimental results, namely to the residual displacement obtained by the experimental monitoring system. With the calibration verified, it is possible to assume that the numerical model correctly predicts the response of fibre grout RC panels when subjected to blast effects. In order to verify this assumption, the strengthened model was modelled and subjected to the blast effects of the corresponding experimental set-up. The comparison between the residual and maximum displacements and the bottom surface’s cracking obtained in the experimental and the numerical tests yields a difference of 4 % for the maximum displacements of the reference model, and a difference of 4 and 10 % for the residual and maximum displacements of the strengthened model, respectively. Additionally, the cracking on the bottom surface of the models was similar in both methods. Therefore, one can conclude that the Applied ElementMethod can correctly predict and simulate the response of fibre grout strengthened RC panels when subjected to blast effects.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The design of anchorage blisters of internal continuity post-tensioning tendons of bridges built by the cantilever method, presents some peculiarities, not only because they are intermediate anchorages but also because these anchorages are located in blisters, so the prestressing force has to be transferred from the blister the bottom slab and web of the girder. The high density of steel reinforcement in anchorage blisters is the most common reason for problems with concrete cast in situ, resulting in zones with low concrete compacity, leading to concrete crushing failures under the anchor plates. A solution may involve improving the concrete compression and tensile strength. To meet these requirements a high-performance fibre reinforced self-compacting mix- ture (HPFRC) was used in anchorage corner blisters of post-tensioning tendons, reducing the concrete cross-section and decreasing the reinforcement needed. To assess the ultimate capacity and the adequate serviceability of the local anchorage zone after reducing the minimum concrete cross-section and the confining reinforcement, specified by the anchorage device supplier for the particular tendon, load transfer tests were performed. To investigate the behaviour of anchorage blisters regarding the transmission of stresses to the web and the bottom slab of the girder, and the feasibility of using high performance concrete only in the blister, two half scale models of the inferior corner of a box girder existing bridge were studied: a reference specimen of ordinary reinforced concrete and a HPFRC blister specimen. The design of the reinforcement was based in the tensile forces obtained on strut-and-tie models. An experimental program was carried out to assess the models used in design and to study the feasibility of using high performance concrete only in the blister, either with casting in situ, or with precast solutions. A non-linear finite element analysis of the tested specimens was also performed and the results compared.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Relatório de estágio de mestrado em Ensino de Matemática no 3.º Ciclo do Ensino Básico e no Ensino Secundário

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The structural analysis involves the definition of the model and selection of the analysis type. The model should represent the stiffness, the mass and the loads of the structure. The structures can be represented using simplified models, such as the lumped mass models, and advanced models resorting the Finite Element Method (FEM) and Discrete Element Method (DEM). Depending on the characteristics of the structure, different types of analysis can be used such as limit analysis, linear and non-linear static analysis and linear and non-linear dynamic analysis. Unreinforced masonry structures present low tensile strength and the linear analyses seem to not be adequate for assessing their structural behaviour. On the other hand, the static and dynamic non-linear analyses are complex, since they involve large time computational requirements and advanced knowledge of the practitioner. The non-linear analysis requires advanced knowledge on the material properties, analysis tools and interpretation of results. The limit analysis with macro-blocks can be assumed as a more practical method in the estimation of maximum load capacity of structure. Furthermore, the limit analysis require a reduced number of parameters, which is an advantage for the assessment of ancient and historical masonry structures, due to the difficult in obtaining reliable data.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

For any vacuum initial data set, we define a local, non-negative scalar quantity which vanishes at every point of the data hypersurface if and only if the data are Kerr initial data. Our scalar quantity only depends on the quantities used to construct the vacuum initial data set which are the Riemannian metric defined on the initial data hypersurface and a symmetric tensor which plays the role of the second fundamental form of the embedded initial data hypersurface. The dependency is algorithmic in the sense that given the initial data one can compute the scalar quantity by algebraic and differential manipulations, being thus suitable for an implementation in a numerical code. The scalar could also be useful in studies of the non-linear stability of the Kerr solution because it serves to measure the deviation of a vacuum initial data set from the Kerr initial data in a local and algorithmic way.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Dissertação de mestrado integrado em Civil Engineering

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Dissertação de mestrado em Optometria Avançada

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The relationship between body size and geographic range was analyzed for 70 species of terrestrial Carnivora ("fissipeds") of the New World, after the control of phylogenetic patterns in the data using phylogenetic eigenvector regression. The analysis from EcoSim software showed that the variables are related as a triangular envelope. Phylogenetic patterns in data were detected by means of phylogenetic correlograms, and 200 simulations of the phenotypic evolution were also performed over the phylogeny. For body size, the simulations suggested a non-linear relationship for the evolution of this character along the phylogeny. For geographic range size, the correlogram showed no phylogenetic patterns. A phylogenetic eigenvector regression was performed on original data and on data simulated under Ornstein-Uhlenbeck process. Since both characters did not evolve under a simple Brownian motion process, the Type I errors should be around 10%, compatible with other methods to analyze correlated evolution. The significant correlation of the original data (r = 0.38; P < 0.05), as well as the triangular envelope, then indicate ecological and adaptive processes connecting the two variables, such as those proposed in minimum viable population models.