909 resultados para Nonparametric Bayes
Resumo:
An increase in daily mortality from myocardial infarction has been observed in association with meteorological factors and air pollution in several cities in the world, mainly in the northern hemisphere. The objective of the present study was to analyze the independent effects of environmental variables on daily counts of death from myocardial infarction in a subtropical region in South America. We used the robust Poisson regression to investigate associations between weather (temperature, humidity and barometric pressure), air pollution (sulfur dioxide, carbon monoxide, and inhalable particulate), and the daily death counts attributed to myocardial infarction in the city of São Paulo in Brazil, where 12,007 fatal events were observed from 1996 to 1998. The model was adjusted in a linear fashion for relative humidity and day-of-week, while nonparametric smoothing factors were used for seasonal trend and temperature. We found a significant association of daily temperature with deaths due to myocardial infarction (P < 0.001), with the lowest mortality being observed at temperatures between 21.6 and 22.6ºC. Relative humidity appeared to exert a protective effect. Sulfur dioxide concentrations correlated linearly with myocardial infarction deaths, increasing the number of fatal events by 3.4% (relative risk of 1.03; 95% confidence interval = 1.02-1.05) for each 10 µg/m³ increase. In conclusion, this study provides evidence of important associations between daily temperature and air pollution and mortality from myocardial infarction in a subtropical region, even after a comprehensive control for confounding factors.
Resumo:
Several methods are used to estimate anaerobic threshold (AT) during exercise. The aim of the present study was to compare AT obtained by a graphic visual method for the estimate of ventilatory and metabolic variables (gold standard), to a bi-segmental linear regression mathematical model of Hinkley's algorithm applied to heart rate (HR) and carbon dioxide output (VCO2) data. Thirteen young (24 ± 2.63 years old) and 16 postmenopausal (57 ± 4.79 years old) healthy and sedentary women were submitted to a continuous ergospirometric incremental test on an electromagnetic braking cycloergometer with 10 to 20 W/min increases until physical exhaustion. The ventilatory variables were recorded breath-to-breath and HR was obtained beat-to-beat over real time. Data were analyzed by the nonparametric Friedman test and Spearman correlation test with the level of significance set at 5%. Power output (W), HR (bpm), oxygen uptake (VO2; mL kg-1 min-1), VO2 (mL/min), VCO2 (mL/min), and minute ventilation (VE; L/min) data observed at the AT level were similar for both methods and groups studied (P > 0.05). The VO2 (mL kg-1 min-1) data showed significant correlation (P < 0.05) between the gold standard method and the mathematical model when applied to HR (r s = 0.75) and VCO2 (r s = 0.78) data for the subjects as a whole (N = 29). The proposed mathematical method for the detection of changes in response patterns of VCO2 and HR was adequate and promising for AT detection in young and middle-aged women, representing a semi-automatic, non-invasive and objective AT measurement.
Resumo:
Our objective was to determine the presence of vascular endothelial growth factor (VEGF), matrix metalloproteinase-2 (MMP-2) and MMP-9 and specific tissue inhibitors of matrix metalloproteinase (TIMP-1 and TIMP-2) in tumor samples obtained from patients with primary breast cancer. We attempted to correlate these findings with the status of the sentinel lymph node (SLN) and clinical-pathological characteristics such as age, tumor size, histological type, histological grade, and vascular invasion. Tumor samples from 88 patients with primary breast cancer were analyzed. The immunoreactivity of VEGF, MMP-2, MMP-9, TIMP-1, and TIMP-2 in tumors was correlated with clinical and pathological features, as well as SLN status. Nonparametric, Mann-Whittney, Kruskal-Wallis, and Spearmann tests were used. Categorical variables were analyzed by the Pearson test. No statistically significant correlation was found between the amount of VEGF, MMP-2, MMP-9, TIMP-1, and TIMP-2 and the presence of tumor cells in the SLN. However, larger tumor diameter (P < 0.01) and the presence of vascular invasion (P < 0.01) were correlated positively with a positive SLN. A significant correlation of higher VEGF levels (P = 0.04) and lower TIMP-1 levels (P = 0.04) with ductal histology was also observed. Furthermore, lower TIMP-2 levels showed a statistically significant correlation with younger age (<50 years) and larger tumor diameter (2.0-5.0 cm). A positive SLN correlated significantly with a larger tumor diameter and the presence of vascular invasion. Higher VEGF and lower TIMP-1 levels were observed in patients with ductal tumors, while higher TIMP-1 levels were observed in lobular tumors.
Resumo:
Research on molecular mechanisms of carcinogenesis plays an important role in diagnosing and treating gastric cancer. Metabolic profiling may offer the opportunity to understand the molecular mechanism of carcinogenesis and help to non-invasively identify the potential biomarkers for the early diagnosis of human gastric cancer. The aims of this study were to explore the underlying metabolic mechanisms of gastric cancer and to identify biomarkers associated with morbidity. Gas chromatography/mass spectrometry (GC/MS) was used to analyze the serum metabolites of 30 Chinese gastric cancer patients and 30 healthy controls. Diagnostic models for gastric cancer were constructed using orthogonal partial least squares discriminant analysis (OPLS-DA). Acquired metabolomic data were analyzed by the nonparametric Wilcoxon test to find serum metabolic biomarkers for gastric cancer. The OPLS-DA model showed adequate discrimination between cancer and non-cancer cohorts while the model failed to discriminate different pathological stages (I-IV) of gastric cancer patients. A total of 44 endogenous metabolites such as amino acids, organic acids, carbohydrates, fatty acids, and steroids were detected, of which 18 differential metabolites were identified with significant differences. A total of 13 variables were obtained for their greatest contribution in the discriminating OPLS-DA model [variable importance in the projection (VIP) value >1.0], among which 11 metabolites were identified using both VIP values (VIP >1) and the Wilcoxon test. These metabolites potentially revealed perturbations of glycolysis and of amino acid, fatty acid, cholesterol, and nucleotide metabolism of gastric cancer patients. These results suggest that gastric cancer serum metabolic profiling has great potential in detecting this disease and helping to understand its metabolic mechanisms.
Resumo:
Hepatitis E virus (HEV) is classified within the family Hepeviridae, genus Hepevirus. HEV genotype 3 (Gt3) infections are endemic in pigs in Western Europe and in North and South America and cause zoonotic infections in humans. Several serological assays to detect HEV antibodies in pigs have been developed, at first mainly based on HEV genotype 1 (Gt1) antigens. To develop a sensitive HEV Gt3 ELISA, a recombinant baculovirus expression product of HEV Gt3 open reading frame-2 was produced and coated onto polystyrene ELISA plates. After incubation of porcine sera, bound HEV antibodies were detected with anti-porcine anti-IgG and anti-IgM conjugates. For primary estimation of sensitivity and specificity of the assay, sets of sera were used from pigs experimentally infected with HEV Gt3. For further validation of the assay and to set the cutoff value, a batch of 1100 pig sera was used. All pig sera were tested using the developed HEV Gt3 assay and two other serologic assays based on HEV Gt1 antigens. Since there is no gold standard available for HEV antibody testing, further validation and a definite setting of the cutoff of the developed HEV Gt3 assay were performed using a statistical approach based on Bayes' theorem. The developed and validated HEV antibody assay showed effective detection of HEV-specific antibodies. This assay can contribute to an improved detection of HEV antibodies and enable more reliable estimates of the prevalence of HEV Gt3 in swine in different regions.
Resumo:
This study aimed to analyze the agreement between measurements of unloaded oxygen uptake and peak oxygen uptake based on equations proposed by Wasserman and on real measurements directly obtained with the ergospirometry system. We performed an incremental cardiopulmonary exercise test (CPET), which was applied to two groups of sedentary male subjects: one apparently healthy group (HG, n=12) and the other had stable coronary artery disease (n=16). The mean age in the HG was 47±4 years and that in the coronary artery disease group (CG) was 57±8 years. Both groups performed CPET on a cycle ergometer with a ramp-type protocol at an intensity that was calculated according to the Wasserman equation. In the HG, there was no significant difference between measurements predicted by the formula and real measurements obtained in CPET in the unloaded condition. However, at peak effort, a significant difference was observed between oxygen uptake (V˙O2)peak(predicted)and V˙O2peak(real)(nonparametric Wilcoxon test). In the CG, there was a significant difference of 116.26 mL/min between the predicted values by the formula and the real values obtained in the unloaded condition. A significant difference in peak effort was found, where V˙O2peak(real)was 40% lower than V˙O2peak(predicted)(nonparametric Wilcoxon test). There was no agreement between the real and predicted measurements as analyzed by Lin’s coefficient or the Bland and Altman model. The Wasserman formula does not appear to be appropriate for prediction of functional capacity of volunteers. Therefore, this formula cannot precisely predict the increase in power in incremental CPET on a cycle ergometer.
Resumo:
The growing population in cities increases the energy demand and affects the environment by increasing carbon emissions. Information and communications technology solutions which enable energy optimization are needed to address this growing energy demand in cities and to reduce carbon emissions. District heating systems optimize the energy production by reusing waste energy with combined heat and power plants. Forecasting the heat load demand in residential buildings assists in optimizing energy production and consumption in a district heating system. However, the presence of a large number of factors such as weather forecast, district heating operational parameters and user behavioural parameters, make heat load forecasting a challenging task. This thesis proposes a probabilistic machine learning model using a Naive Bayes classifier, to forecast the hourly heat load demand for three residential buildings in the city of Skellefteå, Sweden over a period of winter and spring seasons. The district heating data collected from the sensors equipped at the residential buildings in Skellefteå, is utilized to build the Bayesian network to forecast the heat load demand for horizons of 1, 2, 3, 6 and 24 hours. The proposed model is validated by using four cases to study the influence of various parameters on the heat load forecast by carrying out trace driven analysis in Weka and GeNIe. Results show that current heat load consumption and outdoor temperature forecast are the two parameters with most influence on the heat load forecast. The proposed model achieves average accuracies of 81.23 % and 76.74 % for a forecast horizon of 1 hour in the three buildings for winter and spring seasons respectively. The model also achieves an average accuracy of 77.97 % for three buildings across both seasons for the forecast horizon of 1 hour by utilizing only 10 % of the training data. The results indicate that even a simple model like Naive Bayes classifier can forecast the heat load demand by utilizing less training data.
Resumo:
Kandidaatintyö tehtiin osana PulpVision-tutkimusprojektia, jonka tarkoituksena on kehittää kuvapohjaisia laskenta- ja luokittelumetodeja sellun laaduntarkkailuun paperin valmistuksessa. Tämän tutkimusprojektin osana on aiemmin kehitetty metodi, jolla etsittiin kaarevia rakenteita kuvista, ja tätä metodia hyödynnettiin kuitujen etsintään kuvista. Tätä metodia käytettiin lähtökohtana kandidaatintyölle. Työn tarkoituksena oli tutkia, voidaanko erilaisista kuitukuvista laskettujen piirteiden avulla tunnistaa kuvassa olevien kuitujen laji. Näissä kuitukuvissa oli kuituja neljästä eri puulajista ja yhdestä kasvista. Nämä lajit olivat akasia, koivu, mänty, eukalyptus ja vehnä. Jokaisesta lajista valittiin 100 kuitukuvaa ja nämä kuvat jaettiin kahteen ryhmään, joista ensimmäistä käytettiin opetusryhmänä ja toista testausryhmänä. Opetusryhmän avulla jokaiselle kuitulajille laskettiin näitä kuvaavia piirteitä, joiden avulla pyrittiin tunnistamaan testausryhmän kuvissa olevat kuitulajit. Nämä kuvat oli tuottanut CEMIS-Oulu (Center for Measurement and Information Systems), joka on mittaustekniikkaan keskittynyt yksikkö Oulun yliopistossa. Yksittäiselle opetusryhmän kuitukuvalle laskettiin keskiarvot ja keskihajonnat kolmesta eri piirteestä, jotka olivat pituus, leveys ja kaarevuus. Lisäksi laskettiin, kuinka monta kuitua kuvasta löydettiin. Näiden piirteiden eri yhdistelmien avulla testattiin tunnistamisen tarkkuutta käyttämällä k:n lähimmän naapurin menetelmää ja Naiivi Bayes -luokitinta testausryhmän kuville. Testeistä saatiin lupaavia tuloksia muun muassa pituuden ja leveyden keskiarvoja käytettäessä saavutettiin jopa noin 98 %:n tarkkuus molemmilla algoritmeilla. Tunnistuksessa kuitujen keskimäärinen pituus vaikutti olevan kuitukuvia parhaiten kuvaava piirre. Käytettyjen algoritmien välillä ei ollut suurta vaihtelua tarkkuudessa. Testeissä saatujen tulosten perusteella voidaan todeta, että kuitukuvien tunnistaminen on mahdollista. Testien perusteella kuitukuvista tarvitsee laskea vain kaksi piirrettä, joilla kuidut voidaan tunnistaa tarkasti. Käytetyt lajittelualgoritmit olivat hyvin yksinkertaisia, mutta ne toimivat testeissä hyvin.
Resumo:
The effects oftwo types of small-group communication, synchronous computer-mediated and face-to-face, on the quantity and quality of verbal output were con^ared. Quantity was deiSned as the number of turns taken per minute, the number of Analysis-of-Speech units (AS-units) produced per minute, and the number ofwords produced per minute. Quality was defined as the number of words produced per AS-unit. In addition, the interaction of gender and type of communication was explored for any differences that existed in the output produced. Questionnaires were also given to participants to determine attitudes toward computer-mediated and face-to-face communication. Thirty intermediate-level students fi-om the Intensive English Language Program (lELP) at Brock University participated in the study, including 15 females and 15 males. Nonparametric tests, including the Wilcoxon matched-pairs test, Mann-Whitney U test, and Friedman test were used to test for significance at the p < .05 level. No significant differences were found in the effects of computer-mediated and face-to-face communication on the output produced during follow-up speaking sessions. However, the quantity and quality of interaction was significantly higher during face-to-face sessions than computer-mediated sessions. No significant differences were found in the output produced by males and females in these 2 conditions. While participants felt that the use of computer-mediated communication may aid in the development of certain language skills, they generally preferred face-to-face communication. These results differed fi-om previous studies that found a greater quantity and quality of output in addition to a greater equality of interaction produced during computer-mediated sessions in comparison to face-to-face sessions (Kern, 1995; Warschauer, 1996).
Resumo:
This project examines students in a private school in southwestern Ontario on a 17 -day Costa Rica Outward Bound Rainforest multielement course. The study attempted to discover whether voluntary teenage participants could increase their self-perceptions of life effectiveness by participating in a 17-day expedition. A total of9 students participated in the study. The experimental design that was implemented was a mixed methods design. Participants filled in a Life Effectiveness Questionnaire (LEQ) at four predesignated times during the study. These time intervals occurred (a) before the trip commenced, (b) the first day of the trip, ( c) the last day of the trip, and (d) 1 month after the trip ended. Fieldnotes and recordings from informal group debriefing sessions were also used to gather information. Data collected in this study were analyzed in a variety of ways by the researcher. Analyses that were run on the data included the Friedman test for covariance, means, medians, and the Wilcoxon Pairs Test. The questionnaires were analyzed quantitatively, and the fieldnotes were analyzed qualitatively. Nonparametric statistical analysis was implemented as a result of the small group size of participants. Both sets of data were grouped and discussed according to similarities and differences. The data indicate that voluntary teenage participants experience significant changes over time in the areas of time management, social competency, emotional control, active initiative, and self-confidence. The types of outcomes from this study illustrate that Outward Bound-type opportunities should be offered to teenagers in Ontario schools as a means to bring about self-development.
Resumo:
The purpose of this study is to examine the impact of the choice of cut-off points, sampling procedures, and the business cycle on the accuracy of bankruptcy prediction models. Misclassification can result in erroneous predictions leading to prohibitive costs to firms, investors and the economy. To test the impact of the choice of cut-off points and sampling procedures, three bankruptcy prediction models are assessed- Bayesian, Hazard and Mixed Logit. A salient feature of the study is that the analysis includes both parametric and nonparametric bankruptcy prediction models. A sample of firms from Lynn M. LoPucki Bankruptcy Research Database in the U. S. was used to evaluate the relative performance of the three models. The choice of a cut-off point and sampling procedures were found to affect the rankings of the various models. In general, the results indicate that the empirical cut-off point estimated from the training sample resulted in the lowest misclassification costs for all three models. Although the Hazard and Mixed Logit models resulted in lower costs of misclassification in the randomly selected samples, the Mixed Logit model did not perform as well across varying business-cycles. In general, the Hazard model has the highest predictive power. However, the higher predictive power of the Bayesian model, when the ratio of the cost of Type I errors to the cost of Type II errors is high, is relatively consistent across all sampling methods. Such an advantage of the Bayesian model may make it more attractive in the current economic environment. This study extends recent research comparing the performance of bankruptcy prediction models by identifying under what conditions a model performs better. It also allays a range of user groups, including auditors, shareholders, employees, suppliers, rating agencies, and creditors' concerns with respect to assessing failure risk.
Resumo:
This research evaluated (a) the correlation between math anxiety, math attitudes, and achievement in math and (b) comparison among these variables in terms of gender among grade 9 students in a high school located in southern Ontario. Data were compiled from participant responses to the Attitudes Toward Math Inventory (ATMI) and the Math Anxiety Rating Scale for Adolescents (MARS-A), and achievement data were gathered from participants’ grade 9 academic math course marks and the EQAO Grade 9 Assessment of Mathematics. Nonparametric tests were conducted to determine whether there were relationships between the variables and to explore whether gender differences in anxiety, attitudes, and achievement existed for this sample. Results indicated that math anxiety was not related to math achievement but was a strong correlate of attitudes toward math. A strong positive relationship was found between math attitudes and achievement in math. Specifically, self-confidence in math, enjoyment of math, value of math, and motivation were all positive correlates of achievement in math. Also, results for gender comparisons were nonsignificant, indicating that gender differences in math anxiety, math attitudes, and math achievement scores were not prevalent in this group of grade 9 students. Therefore, attitudes toward math were considered to be a stronger predictor of performance than math anxiety or gender for this group.
Resumo:
This paper studies seemingly unrelated linear models with integrated regressors and stationary errors. By adding leads and lags of the first differences of the regressors and estimating this augmented dynamic regression model by feasible generalized least squares using the long-run covariance matrix, we obtain an efficient estimator of the cointegrating vector that has a limiting mixed normal distribution. Simulation results suggest that this new estimator compares favorably with others already proposed in the literature. We apply these new estimators to the testing of purchasing power parity (PPP) among the G-7 countries. The test based on the efficient estimates rejects the PPP hypothesis for most countries.
Resumo:
Dans ce texte, nous revoyons certains développements récents de l’économétrie qui peuvent être intéressants pour des chercheurs dans des domaines autres que l’économie et nous soulignons l’éclairage particulier que l’économétrie peut jeter sur certains thèmes généraux de méthodologie et de philosophie des sciences, tels la falsifiabilité comme critère du caractère scientifique d’une théorie (Popper), la sous-détermination des théories par les données (Quine) et l’instrumentalisme. En particulier, nous soulignons le contraste entre deux styles de modélisation - l’approche parcimonieuse et l’approche statistico-descriptive - et nous discutons les liens entre la théorie des tests statistiques et la philosophie des sciences.
Resumo:
A wide range of tests for heteroskedasticity have been proposed in the econometric and statistics literature. Although a few exact homoskedasticity tests are available, the commonly employed procedures are quite generally based on asymptotic approximations which may not provide good size control in finite samples. There has been a number of recent studies that seek to improve the reliability of common heteroskedasticity tests using Edgeworth, Bartlett, jackknife and bootstrap methods. Yet the latter remain approximate. In this paper, we describe a solution to the problem of controlling the size of homoskedasticity tests in linear regression contexts. We study procedures based on the standard test statistics [e.g., the Goldfeld-Quandt, Glejser, Bartlett, Cochran, Hartley, Breusch-Pagan-Godfrey, White and Szroeter criteria] as well as tests for autoregressive conditional heteroskedasticity (ARCH-type models). We also suggest several extensions of the existing procedures (sup-type of combined test statistics) to allow for unknown breakpoints in the error variance. We exploit the technique of Monte Carlo tests to obtain provably exact p-values, for both the standard and the new tests suggested. We show that the MC test procedure conveniently solves the intractable null distribution problem, in particular those raised by the sup-type and combined test statistics as well as (when relevant) unidentified nuisance parameter problems under the null hypothesis. The method proposed works in exactly the same way with both Gaussian and non-Gaussian disturbance distributions [such as heavy-tailed or stable distributions]. The performance of the procedures is examined by simulation. The Monte Carlo experiments conducted focus on : (1) ARCH, GARCH, and ARCH-in-mean alternatives; (2) the case where the variance increases monotonically with : (i) one exogenous variable, and (ii) the mean of the dependent variable; (3) grouped heteroskedasticity; (4) breaks in variance at unknown points. We find that the proposed tests achieve perfect size control and have good power.