891 resultados para Bayesian risk prediction models
Resumo:
Alpine tree-line ecotones are characterized by marked changes at small spatial scales that may result in a variety of physiognomies. A set of alternative individual-based models was tested with data from four contrasting Pinus uncinata ecotones in the central Spanish Pyrenees to reveal the minimal subset of processes required for tree-line formation. A Bayesian approach combined with Markov chain Monte Carlo methods was employed to obtain the posterior distribution of model parameters, allowing the use of model selection procedures. The main features of real tree lines emerged only in models considering nonlinear responses in individual rates of growth or mortality with respect to the altitudinal gradient. Variation in tree-line physiognomy reflected mainly changes in the relative importance of these nonlinear responses, while other processes, such as dispersal limitation and facilitation, played a secondary role. Different nonlinear responses also determined the presence or absence of krummholz, in agreement with recent findings highlighting a different response of diffuse and abrupt or krummholz tree lines to climate change. The method presented here can be widely applied in individual-based simulation models and will turn model selection and evaluation in this type of models into a more transparent, effective, and efficient exercise.
Resumo:
PURPOSE: According to estimations around 230 people die as a result of radon exposure in Switzerland. This public health concern makes reliable indoor radon prediction and mapping methods necessary in order to improve risk communication to the public. The aim of this study was to develop an automated method to classify lithological units according to their radon characteristics and to develop mapping and predictive tools in order to improve local radon prediction. METHOD: About 240 000 indoor radon concentration (IRC) measurements in about 150 000 buildings were available for our analysis. The automated classification of lithological units was based on k-medoids clustering via pair-wise Kolmogorov distances between IRC distributions of lithological units. For IRC mapping and prediction we used random forests and Bayesian additive regression trees (BART). RESULTS: The automated classification groups lithological units well in terms of their IRC characteristics. Especially the IRC differences in metamorphic rocks like gneiss are well revealed by this method. The maps produced by random forests soundly represent the regional difference of IRCs in Switzerland and improve the spatial detail compared to existing approaches. We could explain 33% of the variations in IRC data with random forests. Additionally, the influence of a variable evaluated by random forests shows that building characteristics are less important predictors for IRCs than spatial/geological influences. BART could explain 29% of IRC variability and produced maps that indicate the prediction uncertainty. CONCLUSION: Ensemble regression trees are a powerful tool to model and understand the multidimensional influences on IRCs. Automatic clustering of lithological units complements this method by facilitating the interpretation of radon properties of rock types. This study provides an important element for radon risk communication. Future approaches should consider taking into account further variables like soil gas radon measurements as well as more detailed geological information.
Resumo:
Notre consommation en eau souterraine, en particulier comme eau potable ou pour l'irrigation, a considérablement augmenté au cours des années. De nombreux problèmes font alors leur apparition, allant de la prospection de nouvelles ressources à la remédiation des aquifères pollués. Indépendamment du problème hydrogéologique considéré, le principal défi reste la caractérisation des propriétés du sous-sol. Une approche stochastique est alors nécessaire afin de représenter cette incertitude en considérant de multiples scénarios géologiques et en générant un grand nombre de réalisations géostatistiques. Nous rencontrons alors la principale limitation de ces approches qui est le coût de calcul dû à la simulation des processus d'écoulements complexes pour chacune de ces réalisations. Dans la première partie de la thèse, ce problème est investigué dans le contexte de propagation de l'incertitude, oú un ensemble de réalisations est identifié comme représentant les propriétés du sous-sol. Afin de propager cette incertitude à la quantité d'intérêt tout en limitant le coût de calcul, les méthodes actuelles font appel à des modèles d'écoulement approximés. Cela permet l'identification d'un sous-ensemble de réalisations représentant la variabilité de l'ensemble initial. Le modèle complexe d'écoulement est alors évalué uniquement pour ce sousensemble, et, sur la base de ces réponses complexes, l'inférence est faite. Notre objectif est d'améliorer la performance de cette approche en utilisant toute l'information à disposition. Pour cela, le sous-ensemble de réponses approximées et exactes est utilisé afin de construire un modèle d'erreur, qui sert ensuite à corriger le reste des réponses approximées et prédire la réponse du modèle complexe. Cette méthode permet de maximiser l'utilisation de l'information à disposition sans augmentation perceptible du temps de calcul. La propagation de l'incertitude est alors plus précise et plus robuste. La stratégie explorée dans le premier chapitre consiste à apprendre d'un sous-ensemble de réalisations la relation entre les modèles d'écoulement approximé et complexe. Dans la seconde partie de la thèse, cette méthodologie est formalisée mathématiquement en introduisant un modèle de régression entre les réponses fonctionnelles. Comme ce problème est mal posé, il est nécessaire d'en réduire la dimensionnalité. Dans cette optique, l'innovation du travail présenté provient de l'utilisation de l'analyse en composantes principales fonctionnelles (ACPF), qui non seulement effectue la réduction de dimensionnalités tout en maximisant l'information retenue, mais permet aussi de diagnostiquer la qualité du modèle d'erreur dans cet espace fonctionnel. La méthodologie proposée est appliquée à un problème de pollution par une phase liquide nonaqueuse et les résultats obtenus montrent que le modèle d'erreur permet une forte réduction du temps de calcul tout en estimant correctement l'incertitude. De plus, pour chaque réponse approximée, une prédiction de la réponse complexe est fournie par le modèle d'erreur. Le concept de modèle d'erreur fonctionnel est donc pertinent pour la propagation de l'incertitude, mais aussi pour les problèmes d'inférence bayésienne. Les méthodes de Monte Carlo par chaîne de Markov (MCMC) sont les algorithmes les plus communément utilisés afin de générer des réalisations géostatistiques en accord avec les observations. Cependant, ces méthodes souffrent d'un taux d'acceptation très bas pour les problèmes de grande dimensionnalité, résultant en un grand nombre de simulations d'écoulement gaspillées. Une approche en deux temps, le "MCMC en deux étapes", a été introduite afin d'éviter les simulations du modèle complexe inutiles par une évaluation préliminaire de la réalisation. Dans la troisième partie de la thèse, le modèle d'écoulement approximé couplé à un modèle d'erreur sert d'évaluation préliminaire pour le "MCMC en deux étapes". Nous démontrons une augmentation du taux d'acceptation par un facteur de 1.5 à 3 en comparaison avec une implémentation classique de MCMC. Une question reste sans réponse : comment choisir la taille de l'ensemble d'entrainement et comment identifier les réalisations permettant d'optimiser la construction du modèle d'erreur. Cela requiert une stratégie itérative afin que, à chaque nouvelle simulation d'écoulement, le modèle d'erreur soit amélioré en incorporant les nouvelles informations. Ceci est développé dans la quatrième partie de la thèse, oú cette méthodologie est appliquée à un problème d'intrusion saline dans un aquifère côtier. -- Our consumption of groundwater, in particular as drinking water and for irrigation, has considerably increased over the years and groundwater is becoming an increasingly scarce and endangered resource. Nofadays, we are facing many problems ranging from water prospection to sustainable management and remediation of polluted aquifers. Independently of the hydrogeological problem, the main challenge remains dealing with the incomplete knofledge of the underground properties. Stochastic approaches have been developed to represent this uncertainty by considering multiple geological scenarios and generating a large number of realizations. The main limitation of this approach is the computational cost associated with performing complex of simulations in each realization. In the first part of the thesis, we explore this issue in the context of uncertainty propagation, where an ensemble of geostatistical realizations is identified as representative of the subsurface uncertainty. To propagate this lack of knofledge to the quantity of interest (e.g., the concentration of pollutant in extracted water), it is necessary to evaluate the of response of each realization. Due to computational constraints, state-of-the-art methods make use of approximate of simulation, to identify a subset of realizations that represents the variability of the ensemble. The complex and computationally heavy of model is then run for this subset based on which inference is made. Our objective is to increase the performance of this approach by using all of the available information and not solely the subset of exact responses. Two error models are proposed to correct the approximate responses follofing a machine learning approach. For the subset identified by a classical approach (here the distance kernel method) both the approximate and the exact responses are knofn. This information is used to construct an error model and correct the ensemble of approximate responses to predict the "expected" responses of the exact model. The proposed methodology makes use of all the available information without perceptible additional computational costs and leads to an increase in accuracy and robustness of the uncertainty propagation. The strategy explored in the first chapter consists in learning from a subset of realizations the relationship between proxy and exact curves. In the second part of this thesis, the strategy is formalized in a rigorous mathematical framework by defining a regression model between functions. As this problem is ill-posed, it is necessary to reduce its dimensionality. The novelty of the work comes from the use of functional principal component analysis (FPCA), which not only performs the dimensionality reduction while maximizing the retained information, but also allofs a diagnostic of the quality of the error model in the functional space. The proposed methodology is applied to a pollution problem by a non-aqueous phase-liquid. The error model allofs a strong reduction of the computational cost while providing a good estimate of the uncertainty. The individual correction of the proxy response by the error model leads to an excellent prediction of the exact response, opening the door to many applications. The concept of functional error model is useful not only in the context of uncertainty propagation, but also, and maybe even more so, to perform Bayesian inference. Monte Carlo Markov Chain (MCMC) algorithms are the most common choice to ensure that the generated realizations are sampled in accordance with the observations. Hofever, this approach suffers from lof acceptance rate in high dimensional problems, resulting in a large number of wasted of simulations. This led to the introduction of two-stage MCMC, where the computational cost is decreased by avoiding unnecessary simulation of the exact of thanks to a preliminary evaluation of the proposal. In the third part of the thesis, a proxy is coupled to an error model to provide an approximate response for the two-stage MCMC set-up. We demonstrate an increase in acceptance rate by a factor three with respect to one-stage MCMC results. An open question remains: hof do we choose the size of the learning set and identify the realizations to optimize the construction of the error model. This requires devising an iterative strategy to construct the error model, such that, as new of simulations are performed, the error model is iteratively improved by incorporating the new information. This is discussed in the fourth part of the thesis, in which we apply this methodology to a problem of saline intrusion in a coastal aquifer.
Resumo:
BACKGROUND AND AIMS: Parental history (PH) and genetic risk scores (GRSs) are separately associated with coronary heart disease (CHD), but evidence regarding their combined effects is lacking. We aimed to evaluate the joint associations and predictive ability of PH and GRSs for incident CHD. METHODS: Data for 4283 Caucasians were obtained from the population-based CoLaus Study, over median follow-up time of 5.6 years. CHD was defined as incident myocardial infarction, angina, percutaneous coronary revascularization or bypass grafting. Single nucleotide polymorphisms for CHD identified by genome-wide association studies were used to construct unweighted and weighted versions of three GRSs, comprising of 38, 53 and 153 SNPs respectively. RESULTS: PH was associated with higher values of all weighted GRSs. After adjustment for age, sex, smoking, diabetes, systolic blood pressure, low and high density lipoprotein cholesterol, PH was significantly associated with CHD [HR 2.61, 95% CI (1.47-4.66)] and further adjustment for GRSs did not change this estimate. Similarly, one standard deviation change of the weighted 153-SNPs GRS was significantly associated with CHD [HR 1.50, 95% CI (1.26-1.80)] and remained so, after further adjustment for PH. The weighted, 153-SNPs GRS, but not PH, modestly improved discrimination [(C-index improvement, 0.016), p = 0.048] and reclassification [(NRI improvement, 8.6%), p = 0.027] beyond cardiovascular risk factors. After including both the GRS and PH, model performance improved further [(C-index improvement, 0.022), p = 0.006]. CONCLUSION: After adjustment for cardiovascular risk factors, PH and a weighted, polygenic GRS were jointly associated with CHD and provided additive information for coronary events prediction.
Resumo:
Objective: We used demographic and clinical data to design practical classification models for prediction of neurocognitive impairment (NCI) in people with HIV infection. Methods: The study population comprised 331 HIV-infected patients with available demographic, clinical, and neurocognitive data collected using a comprehensive battery of neuropsychological tests. Classification and regression trees (CART) were developed to btain detailed and reliable models to predict NCI. Following a practical clinical approach, NCI was considered the main variable for study outcomes, and analyses were performed separately in treatment-naïve and treatment-experienced patients. Results: The study sample comprised 52 treatment-naïve and 279 experienced patients. In the first group, the variables identified as better predictors of NCI were CD4 cell count and age (correct classification [CC]: 79.6%, 3 final nodes). In treatment-experienced patients, the variables most closely related to NCI were years of education, nadir CD4 cell count, central nervous system penetration-effectiveness score, age, employment status, and confounding comorbidities (CC: 82.1%, 7 final nodes). In patients with an undetectable viral load and no comorbidities, we obtained a fairly accurate model in which the main variables were nadir CD4 cell count, current CD4 cell count, time on current treatment, and past highest viral load (CC: 88%, 6 final nodes). Conclusion: Practical classification models to predict NCI in HIV infection can be obtained using demographic and clinical variables. An approach based on CART analyses may facilitate screening for HIV-associated neurocognitive disorders and complement clinical information about risk and protective factors for NCI in HIV-infected patients.
Resumo:
RATIONALE: Patients with acute symptomatic pulmonary embolism (PE) deemed to be at low risk for early complications might be candidates for partial or complete outpatient treatment. OBJECTIVES: To develop and validate a clinical prediction rule that accurately identifies patients with PE and low risk of short-term complications and to compare its prognostic ability with two previously validated models (i.e., the Pulmonary Embolism Severity Index [PESI] and the Simplified PESI [sPESI]) METHODS: Multivariable logistic regression of a large international cohort of patients with PE prospectively enrolled in the RIETE (Registro Informatizado de la Enfermedad TromboEmbólica) registry. MEASUREMENTS AND MAIN RESULTS: All-cause mortality, recurrent PE, and major bleeding up to 10 days after PE diagnosis were determined. Of 18,707 eligible patients with acute symptomatic PE, 46 (0.25%) developed recurrent PE, 203 (1.09%) bled, and 471 (2.51%) died. Predictors included in the final model were chronic heart failure, recent immobilization, recent major bleeding, cancer, hypotension, tachycardia, hypoxemia, renal insufficiency, and abnormal platelet count. The area under receiver-operating characteristic curve was 0.77 (95% confidence interval [CI], 0.75-0.78) for the RIETE score, 0.72 (95% CI, 0.70-0.73) for PESI (P < 0.05), and 0.71 (95% CI, 0.69-0.73) for sPESI (P < 0.05). Our RIETE score outperformed the prognostic value of PESI in terms of net reclassification improvement (P < 0.001), integrated discrimination improvement (P < 0.001), and sPESI (net reclassification improvement, P < 0.001; integrated discrimination improvement, P < 0.001). CONCLUSIONS: We built a new score, based on widely available variables, that can be used to identify patients with PE at low risk of short-term complications, assisting in triage and potentially shortening duration of hospital stay.
Resumo:
Rust, caused by Puccinia psidii, is one of the most important diseases affecting eucalyptus in Brazil. This pathogen causes disease in mini-clonal garden and in young plants in the field, especially in leaves and juvenile shoots. Favorable climate conditions for infection by this pathogen in eucalyptus include temperature between 18 and 25 ºC, together with at least 6-hour leaf wetness periods, for 5 to 7 consecutive days. Considering the interaction between the environment and the pathogen, this study aimed to evaluate the potential impact of global climate changes on the spatial distribution of areas of risk for the occurrence of eucalyptus rust in Brazil. Thus, monthly maps of the areas of risk for the occurrence of this disease were elaborated, considering the current climate conditions, based on a historic series between 1961 and 1990, and the future scenarios A2 and B2, predicted by IPCC. The climate conditions were classified into three categories, according to the potential risk for the disease occurrence, considering temperature (T) and air relative humidity (RH): i) high risk (18 < T < 25 ºC and RH > 90%); ii) medium risk (18 < T < 25 ºC and RH < 90%; T< 18 or T > 25 ºC and RH > 90%); and iii) low risk (T < 18 or T > 25 ºC and RH < 90%). Data about the future climate scenarios were supplied by GCM Change Fields. In this study, the simulation model Hadley Centers for Climate Prediction and Research (HadCm3) was adopted, using the software Idrisi 32. The obtained results led to the conclusion that there will be a reduction in the area favorable to eucalyptus rust occurrence, and such a reduction will be gradual for the decades of 2020, 2050 and 2080 but more marked in scenario A2 than in B2. However, it is important to point out that extensive areas will still be favorable to the disease development, especially in the coldest months of the year, i.e., June and July. Therefore, the zoning of areas and periods of higher occurrence risk, considering the global climate changes, becomes important knowledge for the elaboration of predicting models and an alert for the integrated management of this disease.
Resumo:
Mathematical models often contain parameters that need to be calibrated from measured data. The emergence of efficient Markov Chain Monte Carlo (MCMC) methods has made the Bayesian approach a standard tool in quantifying the uncertainty in the parameters. With MCMC, the parameter estimation problem can be solved in a fully statistical manner, and the whole distribution of the parameters can be explored, instead of obtaining point estimates and using, e.g., Gaussian approximations. In this thesis, MCMC methods are applied to parameter estimation problems in chemical reaction engineering, population ecology, and climate modeling. Motivated by the climate model experiments, the methods are developed further to make them more suitable for problems where the model is computationally intensive. After the parameters are estimated, one can start to use the model for various tasks. Two such tasks are studied in this thesis: optimal design of experiments, where the task is to design the next measurements so that the parameter uncertainty is minimized, and model-based optimization, where a model-based quantity, such as the product yield in a chemical reaction model, is optimized. In this thesis, novel ways to perform these tasks are developed, based on the output of MCMC parameter estimation. A separate topic is dynamical state estimation, where the task is to estimate the dynamically changing model state, instead of static parameters. For example, in numerical weather prediction, an estimate of the state of the atmosphere must constantly be updated based on the recently obtained measurements. In this thesis, a novel hybrid state estimation method is developed, which combines elements from deterministic and random sampling methods.
Resumo:
The purpose of this research is to draw up a clear construction of an anticipatory communicative decision-making process and a successful implementation of a Bayesian application that can be used as an anticipatory communicative decision-making support system. This study is a decision-oriented and constructive research project, and it includes examples of simulated situations. As a basis for further methodological discussion about different approaches to management research, in this research, a decision-oriented approach is used, which is based on mathematics and logic, and it is intended to develop problem solving methods. The approach is theoretical and characteristic of normative management science research. Also, the approach of this study is constructive. An essential part of the constructive approach is to tie the problem to its solution with theoretical knowledge. Firstly, the basic definitions and behaviours of an anticipatory management and managerial communication are provided. These descriptions include discussions of the research environment and formed management processes. These issues define and explain the background to further research. Secondly, it is processed to managerial communication and anticipatory decision-making based on preparation, problem solution, and solution search, which are also related to risk management analysis. After that, a solution to the decision-making support application is formed, using four different Bayesian methods, as follows: the Bayesian network, the influence diagram, the qualitative probabilistic network, and the time critical dynamic network. The purpose of the discussion is not to discuss different theories but to explain the theories which are being implemented. Finally, an application of Bayesian networks to the research problem is presented. The usefulness of the prepared model in examining a problem and the represented results of research is shown. The theoretical contribution includes definitions and a model of anticipatory decision-making. The main theoretical contribution of this study has been to develop a process for anticipatory decision-making that includes management with communication, problem-solving, and the improvement of knowledge. The practical contribution includes a Bayesian Decision Support Model, which is based on Bayesian influenced diagrams. The main contributions of this research are two developed processes, one for anticipatory decision-making, and the other to produce a model of a Bayesian network for anticipatory decision-making. In summary, this research contributes to decision-making support by being one of the few publicly available academic descriptions of the anticipatory decision support system, by representing a Bayesian model that is grounded on firm theoretical discussion, by publishing algorithms suitable for decision-making support, and by defining the idea of anticipatory decision-making for a parallel version. Finally, according to the results of research, an analysis of anticipatory management for planned decision-making is presented, which is based on observation of environment, analysis of weak signals, and alternatives to creative problem solving and communication.
Resumo:
PURPOSE: It was to assess the risk of cardiovascular disease (CVD) in breast cancer survivors (BCS).METHODS: This cross-sectional study analyzed 67 BCS, aged 45 -65 years, who underwent complete oncological treatment, but had not received hormone therapy, tamoxifen or aromatase inhibitors during the previous 6 months. Lipid profile and CVD risk were evaluated, the latter using the Framingham and Systematic COronary Risk Evaluation (SCORE) models. The agreement between cardiovascular risk models was analyzed by calculating a kappa coefficient and its 95% confidence interval (CI).RESULTS: Mean subject age was 53.2±6.0 years, with rates of obesity, hypertension, and dyslipidemia of 25, 34 and 90%, respectively. The most frequent lipid abnormalities were high total cholesterol (70%), high LDL-C (51%) and high non-HDL-C (48%) concentrations. Based on the Framingham score, 22% of the participants had a high risk for coronary artery disease. According to the SCORE model, 100 and 93% of the participants were at low risk for fatal CVD in populations at low and high risk, respectively, for CVD. The agreement between the Framingham and SCORE risk models was poor (kappa: 0.1; 95%CI 0.01 -0.2) for populations at high risk for CVD.CONCLUSIONS: These findings indicate the need to include lipid profile and CVD risk assessment in the follow-up of BCS, focusing on adequate control of serum lipid concentrations.
Resumo:
This thesis is concerned with the state and parameter estimation in state space models. The estimation of states and parameters is an important task when mathematical modeling is applied to many different application areas such as the global positioning systems, target tracking, navigation, brain imaging, spread of infectious diseases, biological processes, telecommunications, audio signal processing, stochastic optimal control, machine learning, and physical systems. In Bayesian settings, the estimation of states or parameters amounts to computation of the posterior probability density function. Except for a very restricted number of models, it is impossible to compute this density function in a closed form. Hence, we need approximation methods. A state estimation problem involves estimating the states (latent variables) that are not directly observed in the output of the system. In this thesis, we use the Kalman filter, extended Kalman filter, Gauss–Hermite filters, and particle filters to estimate the states based on available measurements. Among these filters, particle filters are numerical methods for approximating the filtering distributions of non-linear non-Gaussian state space models via Monte Carlo. The performance of a particle filter heavily depends on the chosen importance distribution. For instance, inappropriate choice of the importance distribution can lead to the failure of convergence of the particle filter algorithm. In this thesis, we analyze the theoretical Lᵖ particle filter convergence with general importance distributions, where p ≥2 is an integer. A parameter estimation problem is considered with inferring the model parameters from measurements. For high-dimensional complex models, estimation of parameters can be done by Markov chain Monte Carlo (MCMC) methods. In its operation, the MCMC method requires the unnormalized posterior distribution of the parameters and a proposal distribution. In this thesis, we show how the posterior density function of the parameters of a state space model can be computed by filtering based methods, where the states are integrated out. This type of computation is then applied to estimate parameters of stochastic differential equations. Furthermore, we compute the partial derivatives of the log-posterior density function and use the hybrid Monte Carlo and scaled conjugate gradient methods to infer the parameters of stochastic differential equations. The computational efficiency of MCMC methods is highly depend on the chosen proposal distribution. A commonly used proposal distribution is Gaussian. In this kind of proposal, the covariance matrix must be well tuned. To tune it, adaptive MCMC methods can be used. In this thesis, we propose a new way of updating the covariance matrix using the variational Bayesian adaptive Kalman filter algorithm.
Resumo:
The SEARCH-RIO study prospectively investigated electrocardiogram (ECG)-derived variables in chronic Chagas disease (CCD) as predictors of cardiac death and new onset ventricular tachycardia (VT). Cardiac arrhythmia is a major cause of death in CCD, and electrical markers may play a significant role in risk stratification. One hundred clinically stable outpatients with CCD were enrolled in this study. They initially underwent a 12-lead resting ECG, signal-averaged ECG, and 24-h ambulatory ECG. Abnormal Q-waves, filtered QRS duration, intraventricular electrical transients (IVET), 24-h standard deviation of normal RR intervals (SDNN), and VT were assessed. Echocardiograms assessed left ventricular ejection fraction. Predictors of cardiac death and new onset VT were identified in a Cox proportional hazard model. During a mean follow-up of 95.3 months, 36 patients had adverse events: 22 new onset VT (mean±SD, 18.4±4‰/year) and 20 deaths (26.4±1.8‰/year). In multivariate analysis, only Q-wave (hazard ratio, HR=6.7; P<0.001), VT (HR=5.3; P<0.001), SDNN<100 ms (HR=4.0; P=0.006), and IVET+ (HR=3.0; P=0.04) were independent predictors of the composite endpoint of cardiac death and new onset VT. A prognostic score was developed by weighting points proportional to beta coefficients and summing-up: Q-wave=2; VT=2; SDNN<100 ms=1; IVET+=1. Receiver operating characteristic curve analysis optimized the cutoff value at >1. In 10,000 bootstraps, the C-statistic of this novel score was non-inferior to a previously validated (Rassi) score (0.89±0.03 and 0.80±0.05, respectively; test for non-inferiority: P<0.001). In CCD, surface ECG-derived variables are predictors of cardiac death and new onset VT.
Resumo:
Solid mixtures for refreshment are already totally integrated to the Brazilian consumers' daily routine, because of their quick preparation method, yield and reasonable price - quite lower if compared to 'ready-to-drink' products or products for prompt consumption, what makes them economically more accessible to low-income populations. Within such a context, the aim of this work was to evaluate the physicochemical and mineral composition, as well as the hygroscopic behavior of four different brands of solid mixture for mango refreshment. The BET, GAB, Oswim and Henderson mathematical models were built through the adjustment of experimental data to the isotherms of adsorption. Results from the physiochemical evaluation showed that the solid mixtures for refreshments are considerable sources of ascorbic acid and reductor sugar; and regarding mineral compounds, they are significant sources of calcium, sodium and potassium. It was also verified that the solid mixtures for refreshments of the four studied brands are considered highly hygroscopic.