926 resultados para Model-Data Integration and Data Assimilation


Relevância:

100.00% 100.00%

Publicador:

Resumo:

Background and objective: Dynamic indices represented by systolic pressure variation and pulse pressure variation have been demonstrated to be more accurate than filling pressures in predicting fluid responsiveness. However, the literature is scarce concerning the impact of different ventilatory modes on these indices. We hypothesized that systolic pressure variation or pulse pressure variation could be affected differently by volume-controlled ventilation and pressure-controlled ventilation in an experimental model, during normovolaemia and hypovolaemia. Method: Thirty-two anaesthetized rabbits were randomly allocated into four groups according to ventilatory modality and volaemic status where G1-ConPCV was the pressure-controlled ventilation control group, G2-HemPCV was associated with haemorrhage, G3-ConVCV was the volume-controlled ventilation control group and G4-HemVCV was associated with haemorrhage. In the haemorrhage groups, blood was removed in two stages: 15% of the estimated blood volume withdrawal at M1, and, 30 min later, an additional 15% at M2. Data were submitted to analysis of variance for repeated measures; a value of P < 0.05 was considered to be statistically significant. Results: At MO (baseline), no significant differences were observed among groups. At M1, dynamic parameters differed significantly among the control and hypovolaemic groups (P < 0.05) but not between ventilation modes. However, when 30% of the estimated blood volume was removed (M2), dynamic parameters became significantly higher in animals under volume-controlled ventilation when compared with those under pressure-controlled ventilation. Conclusions: Under normovolaemia and moderate haemorrhage, dynamic parameters were not influenced by either ventilatory modalities. However, in the second stage of haemorrhage (30%), animals in volume-controlled ventilation presented higher values of systolic pressure variation and pulse pressure variation when compared with those submitted to pressure-controlled ventilation.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The species abundance distribution (SAD) has been a central focus of community ecology for over fifty years, and is currently the subject of widespread renewed interest. The gambin model has recently been proposed as a model that provides a superior fit to commonly preferred SAD models. It has also been argued that the model's single parameter (α) presents a potentially informative ecological diversity metric, because it summarises the shape of the SAD in a single number. Despite this potential, few empirical tests of the model have been undertaken, perhaps because the necessary methods and software for fitting the model have not existed. Here, we derive a maximum likelihood method to fit the model, and use it to undertake a comprehensive comparative analysis of the fit of the gambin model. The functions and computational code to fit the model are incorporated in a newly developed free-to-download R package (gambin). We test the gambin model using a variety of datasets and compare the fit of the gambin model to fits obtained using the Poisson lognormal, logseries and zero-sum multinomial distributions. We found that gambin almost universally provided a better fit to the data and that the fit was consistent for a variety of sample grain sizes. We demonstrate how α can be used to differentiate intelligibly between community structures of Azorean arthropods sampled in different land use types. We conclude that gambin presents a flexible model capable of fitting a wide variety of observed SAD data, while providing a useful index of SAD form in its single fitted parameter. As such, gambin has wide potential applicability in the study of SADs, and ecology more generally.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

O aumento de tecnologias disponíveis na Web favoreceu o aparecimento de diversas formas de informação, recursos e serviços. Este aumento aliado à constante necessidade de formação e evolução das pessoas, quer a nível pessoal como profissional, incentivou o desenvolvimento área de sistemas de hipermédia adaptativa educacional - SHAE. Estes sistemas têm a capacidade de adaptar o ensino consoante o modelo do aluno, características pessoais, necessidades, entre outros aspetos. Os SHAE permitiram introduzir mudanças relativamente à forma de ensino, passando do ensino tradicional que se restringia apenas ao uso de livros escolares até à utilização de ferramentas informáticas que através do acesso à internet disponibilizam material didático, privilegiando o ensino individualizado. Os SHAE geram grande volume de dados, informação contida no modelo do aluno e todos os dados relativos ao processo de aprendizagem de cada aluno. Facilmente estes dados são ignorados e não se procede a uma análise cuidada que permita melhorar o conhecimento do comportamento dos alunos durante o processo de ensino, alterando a forma de aprendizagem de acordo com o aluno e favorecendo a melhoria dos resultados obtidos. O objetivo deste trabalho foi selecionar e aplicar algumas técnicas de Data Mining a um SHAE, PCMAT - Mathematics Collaborative Educational System. A aplicação destas técnicas deram origem a modelos de dados que transformaram os dados em informações úteis e compreensíveis, essenciais para a geração de novos perfis de alunos, padrões de comportamento de alunos, regras de adaptação e pedagógicas. Neste trabalho foram criados alguns modelos de dados recorrendo à técnica de Data Mining de classificação, abordando diferentes algoritmos. Os resultados obtidos permitirão definir novas regras de adaptação e padrões de comportamento dos alunos, poderá melhorar o processo de aprendizagem disponível num SHAE.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Dissertation to obtain PhD in Industrial Engineering

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Hospitals are nowadays collecting vast amounts of data related with patient records. All this data hold valuable knowledge that can be used to improve hospital decision making. Data mining techniques aim precisely at the extraction of useful knowledge from raw data. This work describes an implementation of a medical data mining project approach based on the CRISP-DM methodology. Recent real-world data, from 2000 to 2013, were collected from a Portuguese hospital and related with inpatient hospitalization. The goal was to predict generic hospital Length Of Stay based on indicators that are commonly available at the hospitalization process (e.g., gender, age, episode type, medical specialty). At the data preparation stage, the data were cleaned and variables were selected and transformed, leading to 14 inputs. Next, at the modeling stage, a regression approach was adopted, where six learning methods were compared: Average Prediction, Multiple Regression, Decision Tree, Artificial Neural Network ensemble, Support Vector Machine and Random Forest. The best learning model was obtained by the Random Forest method, which presents a high quality coefficient of determination value (0.81). This model was then opened by using a sensitivity analysis procedure that revealed three influential input attributes: the hospital episode type, the physical service where the patient is hospitalized and the associated medical specialty. Such extracted knowledge confirmed that the obtained predictive model is credible and with potential value for supporting decisions of hospital managers.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

DNA microarrays are one of the most used technologies for gene expression measurement. However, there are several distinct microarray platforms, from different manufacturers, each with its own measurement protocol, resulting in data that can hardly be compared or directly integrated. Data integration from multiple sources aims to improve the assertiveness of statistical tests, reducing the data dimensionality problem. The integration of heterogeneous DNA microarray platforms comprehends a set of tasks that range from the re-annotation of the features used on gene expression, to data normalization and batch effect elimination. In this work, a complete methodology for gene expression data integration and application is proposed, which comprehends a transcript-based re-annotation process and several methods for batch effect attenuation. The integrated data will be used to select the best feature set and learning algorithm for a brain tumor classification case study. The integration will consider data from heterogeneous Agilent and Affymetrix platforms, collected from public gene expression databases, such as The Cancer Genome Atlas and Gene Expression Omnibus.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The nitrogen dioxide is a primary pollutant, regarded for the estimation of the air quality index, whose excessive presence may cause significant environmental and health problems. In the current work, we suggest characterizing the evolution of NO2 levels, by using geostatisti- cal approaches that deal with both the space and time coordinates. To develop our proposal, a first exploratory analysis was carried out on daily values of the target variable, daily measured in Portugal from 2004 to 2012, which led to identify three influential covariates (type of site, environment and month of measurement). In a second step, appropriate geostatistical tools were applied to model the trend and the space-time variability, thus enabling us to use the kriging techniques for prediction, without requiring data from a dense monitoring network. This method- ology has valuable applications, as it can provide accurate assessment of the nitrogen dioxide concentrations at sites where either data have been lost or there is no monitoring station nearby.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Currently, the quality of the Indonesian national road network is inadequate due to several constraints, including overcapacity and overloaded trucks. The high deterioration rate of the road infrastructure in developing countries along with major budgetary restrictions and high growth in traffic have led to an emerging need for improving the performance of the highway maintenance system. However, the high number of intervening factors and their complex effects require advanced tools to successfully solve this problem. The high learning capabilities of Data Mining (DM) are a powerful solution to this problem. In the past, these tools have been successfully applied to solve complex and multi-dimensional problems in various scientific fields. Therefore, it is expected that DM can be used to analyze the large amount of data regarding the pavement and traffic, identify the relationship between variables, and provide information regarding the prediction of the data. In this paper, we present a new approach to predict the International Roughness Index (IRI) of pavement based on DM techniques. DM was used to analyze the initial IRI data, including age, Equivalent Single Axle Load (ESAL), crack, potholes, rutting, and long cracks. This model was developed and verified using data from an Integrated Indonesia Road Management System (IIRMS) that was measured with the National Association of Australian State Road Authorities (NAASRA) roughness meter. The results of the proposed approach are compared with the IIRMS analytical model adapted to the IRI, and the advantages of the new approach are highlighted. We show that the novel data-driven model is able to learn (with high accuracy) the complex relationships between the IRI and the contributing factors of overloaded trucks

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Dissertação de mestrado integrado em Engenharia e Gestão de Sistemas de Informação

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Dissertação de mestrado em Estatística

Relevância:

100.00% 100.00%

Publicador:

Resumo:

HIV latency is a major obstacle to curing infection. Current strategies to eradicate HIV aim at increasing transcription of the latent provirus. In the present study we observed that latently infected CD4+ T cells from HIV-infected individuals failed to produce viral particles upon ex vivo exposure to SAHA (vorinostat), despite effective inhibition of histone deacetylases. To identify steps that were not susceptible to the action of SAHA or other latency reverting agents, we used a primary CD4+ T cell model, joint host and viral RNA sequencing, and a viral-encoded reporter. This model served to investigate the characteristics of latently infected cells, the dynamics of HIV latency, and the process of reactivation induced by various stimuli. During latency, we observed persistence of viral transcripts but only limited viral translation. Similarly, the reactivating agents SAHA and disulfiram successfully increased viral transcription, but failed to effectively enhance viral translation, mirroring the ex vivo data. This study highlights the importance of post-transcriptional blocks as one mechanism leading to HIV latency that needs to be relieved in order to purge the viral reservoir.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Converging evidence favors an abnormal susceptibility to oxidative stress in schizophrenia. Decreased levels of glutathione (GSH), the major cellular antioxidant and redox regulator, was observed in cerebrospinal-fluid and prefrontal cortex of patients. Importantly, abnormal GSH synthesis of genetic origin was observed: Two case-control studies showed an association with a GAG trinucleotide repeat (TNR) polymorphism in the GSH key synthesizing enzyme glutamate-cysteine-ligase (GCL) catalytic subunit (GCLC) gene. The most common TNR genotype 7/7 was more frequent in controls, whereas the rarest TNR genotype 8/8 was three times more frequent in patients. The disease associated genotypes (35% of patients) correlated with decreased GCLC protein, GCL activity and GSH content. Similar GSH system anomalies were observed in early psychosis patients. Such redox dysregulation combined with environmental stressors at specific developmental stages could underlie structural and functional connectivity anomalies. In pharmacological and knock-out (KO) models, GSH deficit induces anomalies analogous to those reported in patients. (a) morphology: spine density and GABA-parvalbumine immunoreactivity (PV-I) were decreased in anterior cingulate cortex. KO mice showed delayed cortical PV-I at PD10. This effect is exacerbated in mice with increased DA from PD5-10. KO mice exhibit cortical impairment in myelin and perineuronal net known to modulate PV connectivity. (b) physiology: In cultured neurons, NMDA response are depressed by D2 activation. In hippocampus, NMDA-dependent synaptic plasticity is impaired and kainate induced g-oscillations are reduced in parallel to PV-I. (c) cognition: low GSH models show increased sensitivity to stress, hyperactivity, abnormal object recognition, olfactory integration and social behavior. In a clinical study, GSH precursor N-acetyl cysteine (NAC) as add on therapy, improves the negative symptoms and decreases the side effects of antipsychotics. In an auditory oddball paradigm, NAC improves the mismatched negativity, an evoked potential related to pre-attention and to NMDA receptors function. In summary, clinical and experimental evidence converge to demonstrate that a genetically induced dysregulation of GSH synthesis combined with environmental insults in early development represent a major risk factor contributing to the development of schizophrenia Conclusion Based on these data, we proposed a model for PSIP1 promoter activity involving a complex interplay between yet undefined regulatory elements to modulate gene expression.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Geochemical and petrographical studies of lavas and ignimbrites from the Quaternary Nisyros-Yali volcanic system in the easternmost part of the Hellenic arc (Greece) reveal insight into magma generating processes. A compositional gap between 61 and 68 wt.% SiO2 is recognized that coincides with the stratigraphic distinction between pre-caldera and postcaldera volcanic units. Trace element systematics support the subdivision of Nisyros and Yali volcanic units into two distinct suites of rocks. The variation of Nd and Hf present day isotope data and the fact that they are distinct from the isotope compositions of MORB rule out an origin by pure differentiation and require assimilation of a crustal component. Lead isotope ratios of Nisyros and Yali volcanic rocks support mixing of mantle material with a lower crust equivalent. However, Sr-87/Sr-86 ratios of 0.7036-0.7048 are incompatible with a simple binary mixing scenario and give low depleted mantle extraction ages (< 0.1 Ga), in contrast with Pb model ages of 0.3 Ga and Hf and Nd model ages of ca. 0.8 Ga. The budget of fluid-mobile elements Sr and Pb is likely to be dominated by abundant hydrous fluids characterised by mantle-like Sr isotope ratios. Late stage fluids probably were enriched in CO2, needed to explain the high Th concentrations. The occurrence of hydrated minerals (e.g., amphibole) in the first post-caldera unit with the lowermost Sr-87/Sr-86 ratio of 0.7036 +/- 2 can be interpreted as the result of the increased water activity in the source. The presence of two different plagioclase phenocryst generations in the first lava subsequent to the caldera-causing event is indicative for a longer storage time of this magma at a shallower level. A model capable of explaining these observations involves three evolutionary stages. First stage, assimilation of lower crustal material by a primitive magma of mantle origin (as modelled by Nd-Hf isotope systematics). This stage ended by an interruption in replenishment that led to an increase of crystallization and, hence, an increase in viscosity, suppressing eruption. During this time gap, differentiation by fractional crystallization led to enrichment of incompatible species, especially aqueous fluids, to silica depolymerisation and to a decrease in viscosity, finally enabling eruption again in the third stage. (c) 2005 Elsevier B.V. All rights reserved.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Quantitative or algorithmic trading is the automatization of investments decisions obeying a fixed or dynamic sets of rules to determine trading orders. It has increasingly made its way up to 70% of the trading volume of one of the biggest financial markets such as the New York Stock Exchange (NYSE). However, there is not a signi cant amount of academic literature devoted to it due to the private nature of investment banks and hedge funds. This projects aims to review the literature and discuss the models available in a subject that publications are scarce and infrequently. We review the basic and fundamental mathematical concepts needed for modeling financial markets such as: stochastic processes, stochastic integration and basic models for prices and spreads dynamics necessary for building quantitative strategies. We also contrast these models with real market data with minutely sampling frequency from the Dow Jones Industrial Average (DJIA). Quantitative strategies try to exploit two types of behavior: trend following or mean reversion. The former is grouped in the so-called technical models and the later in the so-called pairs trading. Technical models have been discarded by financial theoreticians but we show that they can be properly cast into a well defined scientific predictor if the signal generated by them pass the test of being a Markov time. That is, we can tell if the signal has occurred or not by examining the information up to the current time; or more technically, if the event is F_t-measurable. On the other hand the concept of pairs trading or market neutral strategy is fairly simple. However it can be cast in a variety of mathematical models ranging from a method based on a simple euclidean distance, in a co-integration framework or involving stochastic differential equations such as the well-known Ornstein-Uhlenbeck mean reversal ODE and its variations. A model for forecasting any economic or financial magnitude could be properly defined with scientific rigor but it could also lack of any economical value and be considered useless from a practical point of view. This is why this project could not be complete without a backtesting of the mentioned strategies. Conducting a useful and realistic backtesting is by no means a trivial exercise since the \laws" that govern financial markets are constantly evolving in time. This is the reason because we make emphasis in the calibration process of the strategies' parameters to adapt the given market conditions. We find out that the parameters from technical models are more volatile than their counterpart form market neutral strategies and calibration must be done in a high-frequency sampling manner to constantly track the currently market situation. As a whole, the goal of this project is to provide an overview of a quantitative approach to investment reviewing basic strategies and illustrating them by means of a back-testing with real financial market data. The sources of the data used in this project are Bloomberg for intraday time series and Yahoo! for daily prices. All numeric computations and graphics used and shown in this project were implemented in MATLAB^R scratch from scratch as a part of this thesis. No other mathematical or statistical software was used.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

BACKGROUND: Solexa/Illumina short-read ultra-high throughput DNA sequencing technology produces millions of short tags (up to 36 bases) by parallel sequencing-by-synthesis of DNA colonies. The processing and statistical analysis of such high-throughput data poses new challenges; currently a fair proportion of the tags are routinely discarded due to an inability to match them to a reference sequence, thereby reducing the effective throughput of the technology. RESULTS: We propose a novel base calling algorithm using model-based clustering and probability theory to identify ambiguous bases and code them with IUPAC symbols. We also select optimal sub-tags using a score based on information content to remove uncertain bases towards the ends of the reads. CONCLUSION: We show that the method improves genome coverage and number of usable tags as compared with Solexa's data processing pipeline by an average of 15%. An R package is provided which allows fast and accurate base calling of Solexa's fluorescence intensity files and the production of informative diagnostic plots.