886 resultados para least absolute deviation (LAD) fitting
Resumo:
The impact of the Parkinson's disease and its treatment on the patients' health-related quality of life can be estimated either by means of generic measures such as the european quality of Life-5 Dimensions (EQ-5D) or specific measures such as the 8-item Parkinson's disease questionnaire (PDQ-8). In clinical studies, PDQ-8 could be used in detriment of EQ-5D due to the lack of resources, time or clinical interest in generic measures. Nevertheless, PDQ-8 cannot be applied in cost-effectiveness analyses which require generic measures and quantitative utility scores, such as EQ-5D. To deal with this problem, a commonly used solution is the prediction of EQ-5D from PDQ-8. In this paper, we propose a new probabilistic method to predict EQ-5D from PDQ-8 using multi-dimensional Bayesian network classifiers. Our approach is evaluated using five-fold cross-validation experiments carried out on a Parkinson's data set containing 488 patients, and is compared with two additional Bayesian network-based approaches, two commonly used mapping methods namely, ordinary least squares and censored least absolute deviations, and a deterministic model. Experimental results are promising in terms of predictive performance as well as the identification of dependence relationships among EQ-5D and PDQ-8 items that the mapping approaches are unable to detect
Resumo:
Different types of spin–spin coupling constants (SSCCs) for several representative small molecules are evaluated and analyzed using a combination of 10 exchange functionals with 12 correlation functionals. For comparison, calculations performed using MCSCF, SOPPA, other common DFT methods, and also experimental data are considered. A detailed study of the percentage of Hartree–Fock exchange energy in SSCCs and in its four contributions is carried out. From the above analysis, a combined functional formed with local Slater (34%), Hartree–Fock exchange (66%), and P86 correlation functional (S66P86) is proposed in this paper. The accuracy of the values obtained with this hybrid functional (mean absolute deviation of 4.5 Hz) is similar to that of the SOPPA method (mean absolute deviation of 4.6 Hz).
Resumo:
Tese de mestrado, Bioinformática e Biologia Computacional (Bioinformática), Universidade de Lisboa, Faculdade de Ciências, 2016
Resumo:
Correlation and regression are two of the statistical procedures most widely used by optometrists. However, these tests are often misused or interpreted incorrectly, leading to erroneous conclusions from clinical experiments. This review examines the major statistical tests concerned with correlation and regression that are most likely to arise in clinical investigations in optometry. First, the use, interpretation and limitations of Pearson's product moment correlation coefficient are described. Second, the least squares method of fitting a linear regression to data and for testing how well a regression line fits the data are described. Third, the problems of using linear regression methods in observational studies, if there are errors associated in measuring the independent variable and for predicting a new value of Y for a given X, are discussed. Finally, methods for testing whether a non-linear relationship provides a better fit to the data and for comparing two or more regression lines are considered.
Resumo:
Feature selection is important in medical field for many reasons. However, selecting important variables is a difficult task with the presence of censoring that is a unique feature in survival data analysis. This paper proposed an approach to deal with the censoring problem in endovascular aortic repair survival data through Bayesian networks. It was merged and embedded with a hybrid feature selection process that combines cox's univariate analysis with machine learning approaches such as ensemble artificial neural networks to select the most relevant predictive variables. The proposed algorithm was compared with common survival variable selection approaches such as; least absolute shrinkage and selection operator LASSO, and Akaike information criterion AIC methods. The results showed that it was capable of dealing with high censoring in the datasets. Moreover, ensemble classifiers increased the area under the roc curves of the two datasets collected from two centers located in United Kingdom separately. Furthermore, ensembles constructed with center 1 enhanced the concordance index of center 2 prediction compared to the model built with a single network. Although the size of the final reduced model using the neural networks and its ensembles is greater than other methods, the model outperformed the others in both concordance index and sensitivity for center 2 prediction. This indicates the reduced model is more powerful for cross center prediction.
Resumo:
This thesis studies survival analysis techniques dealing with censoring to produce predictive tools that predict the risk of endovascular aortic aneurysm repair (EVAR) re-intervention. Censoring indicates that some patients do not continue follow up, so their outcome class is unknown. Methods dealing with censoring have drawbacks and cannot handle the high censoring of the two EVAR datasets collected. Therefore, this thesis presents a new solution to high censoring by modifying an approach that was incapable of differentiating between risks groups of aortic complications. Feature selection (FS) becomes complicated with censoring. Most survival FS methods depends on Cox's model, however machine learning classifiers (MLC) are preferred. Few methods adopted MLC to perform survival FS, but they cannot be used with high censoring. This thesis proposes two FS methods which use MLC to evaluate features. The two FS methods use the new solution to deal with censoring. They combine factor analysis with greedy stepwise FS search which allows eliminated features to enter the FS process. The first FS method searches for the best neural networks' configuration and subset of features. The second approach combines support vector machines, neural networks, and K nearest neighbor classifiers using simple and weighted majority voting to construct a multiple classifier system (MCS) for improving the performance of individual classifiers. It presents a new hybrid FS process by using MCS as a wrapper method and merging it with the iterated feature ranking filter method to further reduce the features. The proposed techniques outperformed FS methods based on Cox's model such as; Akaike and Bayesian information criteria, and least absolute shrinkage and selector operator in the log-rank test's p-values, sensitivity, and concordance. This proves that the proposed techniques are more powerful in correctly predicting the risk of re-intervention. Consequently, they enable doctors to set patients’ appropriate future observation plan.
Resumo:
Correct specification of the simple location quotients in regionalizing the national direct requirements table is essential to the accuracy of regional input-output multipliers. The purpose of this research is to examine the relative accuracy of these multipliers when earnings, employment, number of establishments, and payroll data specify the simple location quotients.^ For each specification type, I derive a column of total output multipliers and a column of total income multipliers. These multipliers are based on the 1987 benchmark input-output accounts of the U.S. economy and 1988-1992 state of Florida data.^ Error sign tests, and Standardized Mean Absolute Deviation (SMAD) statistics indicate that the output multiplier estimates overestimate the output multipliers published by the Department of Commerce-Bureau of Economic Analysis (BEA) for the state of Florida. In contrast, the income multiplier estimates underestimate the BEA's income multipliers. For a given multiplier type, the Spearman-rank correlation analysis shows that the multiplier estimates and the BEA multipliers have statistically different rank ordering of row elements. The above tests also find no significant different differences, both in size and ranking distributions, among the vectors of multiplier estimates. ^
Resumo:
Correct specification of the simple location quotients in regionalizing the national direct requirements table is essential to the accuracy of regional input-output multipliers. The purpose of this research is to examine the relative accuracy of these multipliers when earnings, employment, number of establishments, and payroll data specify the simple location quotients. For each specification type, I derive a column of total output multipliers and a column of total income multipliers. These multipliers are based on the 1987 benchmark input-output accounts of the U.S. economy and 1988-1992 state of Florida data. Error sign tests, and Standardized Mean Absolute Deviation (SMAD) statistics indicate that the output multiplier estimates overestimate the output multipliers published by the Department of Commerce-Bureau of Economic Analysis (BEA) for the state of Florida. In contrast, the income multiplier estimates underestimate the BEA's income multipliers. For a given multiplier type, the Spearman-rank correlation analysis shows that the multiplier estimates and the BEA multipliers have statistically different rank ordering of row elements. The above tests also find no significant different differences, both in size and ranking distributions, among the vectors of multiplier estimates.
Resumo:
Seagrass is expected to benefit from increased carbon availability under future ocean acidification. This hypothesis has been little tested by in situ manipulation. To test for ocean acidification effects on seagrass meadows under controlled CO2/pH conditions, we used a Free Ocean Carbon Dioxide Enrichment (FOCE) system which allows for the manipulation of pH as continuous offset from ambient. It was deployed in a Posidonia oceanica meadow at 11 m depth in the Northwestern Mediterranean Sea. It consisted of two benthic enclosures, an experimental and a control unit both 1.7 m**3, and an additional reference plot in the ambient environment (2 m**2) to account for structural artifacts. The meadow was monitored from April to November 2014. The pH of the experimental enclosure was lowered by 0.26 pH units for the second half of the 8-month study. The greatest magnitude of change in P. oceanica leaf biometrics, photosynthesis, and leaf growth accompanied seasonal changes recorded in the environment and values were similar between the two enclosures. Leaf thickness may change in response to lower pH but this requires further testing. Results are congruent with other short-term and natural studies that have investigated the response of P. oceanica over a wide range of pH. They suggest any benefit from ocean acidification, over the next century (at a pH of 7.7 on the total scale), on Posidonia physiology and growth may be minimal and difficult to detect without increased replication or longer experimental duration. The limited stimulation, which did not surpass any enclosure or seasonal effect, casts doubts on speculations that elevated CO2 would confer resistance to thermal stress and increase the buffering capacity of meadows.
Resumo:
Reliable and fine resolution estimates of surface net-radiation are required for estimating latent and sensible heat fluxes between the land surface and the atmosphere. However, currently, fine resolution estimates of net-radiation are not available and consequently it is challenging to develop multi-year estimates of evapotranspiration at scales that can capture land surface heterogeneity and are relevant for policy and decision-making. We developed and evaluated a global net-radiation product at 5 km and 8-day resolution by combining mutually consistent atmosphere and land data from the Moderate Resolution Imaging Spectroradiometer (MODIS) on board Terra. Comparison with net-radiation measurements from 154 globally distributed sites (414 site-years) from the FLUXNET and Surface Radiation budget network (SURFRAD) showed that the net-radiation product agreed well with measurements across seasons and climate types in the extratropics (Wilmott’s index ranged from 0.74 for boreal to 0.63 for Mediterranean sites). Mean absolute deviation between the MODIS and measured net-radiation ranged from 38.0 ± 1.8 W∙m−2 in boreal to 72.0 ± 4.1 W∙m−2 in the tropical climates. The mean bias was small and constituted only 11%, 0.7%, 8.4%, 4.2%, 13.3%, and 5.4% of the mean absolute error in daytime net-radiation in boreal, Mediterranean, temperate-continental, temperate, semi-arid, and tropical climate, respectively. To assess the accuracy of the broader spatiotemporal patterns, we upscaled error-quantified MODIS net-radiation and compared it with the net-radiation estimates from the coarse spatial (1° × 1°) but high temporal resolution gridded net-radiation product from the Clouds and Earth’s Radiant Energy System (CERES). Our estimates agreed closely with the net-radiation estimates from the CERES. Difference between the two was less than 10 W•m−2 in 94% of the total land area. MODIS net-radiation product will be a valuable resource for the science community studying turbulent fluxes and energy budget at the Earth’s surface.
Resumo:
The viscosity of ionic liquids (ILs) has been modeled as a function of temperature and at atmospheric pressure using a new method based on the UNIFAC–VISCO method. This model extends the calculations previously reported by our group (see Zhao et al. J. Chem. Eng. Data 2016, 61, 2160–2169) which used 154 experimental viscosity data points of 25 ionic liquids for regression of a set of binary interaction parameters and ion Vogel–Fulcher–Tammann (VFT) parameters. Discrepancies in the experimental data of the same IL affect the quality of the correlation and thus the development of the predictive method. In this work, mathematical gnostics was used to analyze the experimental data from different sources and recommend one set of reliable data for each IL. These recommended data (totally 819 data points) for 70 ILs were correlated using this model to obtain an extended set of binary interaction parameters and ion VFT parameters, with a regression accuracy of 1.4%. In addition, 966 experimental viscosity data points for 11 binary mixtures of ILs were collected from literature to establish this model. All the binary data consist of 128 training data points used for the optimization of binary interaction parameters and 838 test data points used for the comparison of the pure evaluated values. The relative average absolute deviation (RAAD) for training and test is 2.9% and 3.9%, respectively.
Resumo:
BACKGROUND: The purpose of the present study was to investigate the diagnostic value of T2-mapping in acute myocarditis (ACM) and to define cut-off values for edema detection. METHODS: Cardiovascular magnetic resonance (CMR) data of 31 patients with ACM were retrospectively analyzed. 30 healthy volunteers (HV) served as a control. Additionally to the routine CMR protocol, T2-mapping data were acquired at 1.5 T using a breathhold Gradient-Spin-Echo T2-mapping sequence in six short axis slices. T2-maps were segmented according to the 16-segments AHA-model and segmental T2 values as well as the segmental pixel-standard deviation (SD) were analyzed. RESULTS: Mean differences of global myocardial T2 or pixel-SD between HV and ACM patients were only small, lying in the normal range of HV. In contrast, variation of segmental T2 values and pixel-SD was much larger in ACM patients compared to HV. In random forests and multiple logistic regression analyses, the combination of the highest segmental T2 value within each patient (maxT2) and the mean absolute deviation (MAD) of log-transformed pixel-SD (madSD) over all 16 segments within each patient proved to be the best discriminators between HV and ACM patients with an AUC of 0.85 in ROC-analysis. In classification trees, a combined cut-off of 0.22 for madSD and of 68 ms for maxT2 resulted in 83% specificity and 81% sensitivity for detection of ACM. CONCLUSIONS: The proposed cut-off values for maxT2 and madSD in the setting of ACM allow edema detection with high sensitivity and specificity and therefore have the potential to overcome the hurdles of T2-mapping for its integration into clinical routine.
Resumo:
Dissertação (mestrado)—Universidade de Brasília, Faculdade de Tecnoloigia, 2016.
Resumo:
We have sought to determine the nature of the free-radical precursors to ring-opened hydrocarbon 5 and ring-closed hydrocarbon 6. Reasonable alternative formulations involve the postulation of hydrogen abstraction (a) by a pair of rapidly equilibrating classical radicals (the ring-opened allylcarbinyl-type radical 3 and the ring-closed cyclopropylcarbinyl-type 4), or (b) by a nonclassical radical such as homoallylic radical 7.
[Figure not reproduced.]
Entry to the radical system is gained via degassed thermal decomposition of peresters having the ring-opened and the ring-closed structures. The ratio of 6:5 is essentially independent of the hydrogen donor concentration for decomposition of the former at 125° in the presence of triethyltin hydrdride. A deuterium labeling study showed that the α and β methylene groups in 3 (or the equivalent) are rapidly interchanged under these conditions.
Existence of two (or more) product-forming intermediates is indicated (a) by dependence of the ratio 6:5 on the tin hydride concentration for decomposition of the ring-closed perester at 10 and 35°, and (b) by formation of cage products having largely or wholly the structure (ring-opened or ring-closed) of the starting perester.
Relative rates of hydrogen abstraction by 3 could be inferred by comparison of ratios of rate constants for hydrogen abstraction and ortho-ring cyclization:
[Figure not reproduced.]
At 100° values of ka/kr are 0.14 for hydrogen abstraction from 1,4-cyclohexadiene and 7 for abstraction from triethyltin hydride. The ratio 6:5 at the same temperature is ~0.0035 for hydrogen abstraction from 1,4-cyclohexadiene, ~0.078 for abstraction from the tin hydride, and ≥ 5 for abstraction from cyclohexadienyl radicals. These data indicate that abstraction of hydrogen from triethyltin hydride is more rapid than from 1,4-cyclohexadiene by a factor of ~1000 for 4, but only ~50 for 3.
Measurements of product ratios at several temperatures allowed the construction of an approximate energy-level scheme. A major inference is that isomerization of 3 to 4 is exothermic by 8 ± 3 kcal/mole, in good agreement with expectations based on bond dissociation energies. Absolute rate-constant estimates are also given.
The results are nicely compatible with a classical-radical mechanism, but attempted interpretation in terms of a nonclassical radical precursor of product ratios formed even from equilibrated radical intermediates leads, it is argued, to serious difficulties.
The roles played by hydrogen abstraction from 1,4,-cyclohexadiene and from the derived cyclohexadienyl radicals were probed by fitting observed ratios of 6:5 and 5:10 in the sense of least-squares to expressions derived for a complex mechanistic scheme. Some 30 to 40 measurements on each product ratio, obtained under a variety of experimental conditions, could be fit with an average deviation of ~6%. Significant systematic deviations were found, but these could largely be redressed by assuming (a) that the rate constant for reaction of 4 with cyclohexadienyl radical is inversely proportional to the viscosity of the medium (i.e., is diffusion-controlled), and (b) that ka/kr for hydrogen abstraction from 1,4-cyclohexadiene depends slightly on the composition of the medium. An average deviation of 4.4% was thereby attained.
Degassed thermal decomposition of the ring-opened perester in the presence of the triethyltin hydride occurs primarily by attack on perester of triethyltin radicals, presumably at the –O-O- bond, even at 0.01 M tin hydride at 100 and 125°. Tin ester and tin ether are apparently formed in closely similar amounts under these conditions, but the tin ester predominates at room temperature in the companion air-induced decomposition, indicating that attack on perester to give the tin ether requires an activation energy approximately 5 kcal/mole in excess of that for the formation of tin ester.