980 resultados para empirical correlation
Whence a healthy mind: Correlation of physical fitness and academic performance among schoolchildren
Resumo:
Background. Public schools are a key forum in the fight for child health because of the opportunities they present for physical activity and fitness surveillance. However, because schools are evaluated and funded on the basis of standardized academic performance rather than physical activity, empirical research evaluating the connections between fitness and academic performance is needed to justify curriculum allocations to physical activity. ^ Methods. Analyses were based on a convenience sample of 315,092 individually-matched standardized academic (TAKS™) and fitness (FITNESSGRAM®) test records collected by 13 Texas school districts under state mandates. We categorized each fitness result in quintiles by age and gender and used a mixed effects regression model to compare the academic performance of the top and bottom fitness groups for each fitness test and grade level combination. ^ Results. All fitness variables except BMI showed significant, positive associations with academic performance after sociodemographic covariate adjustments, with effect sizes ranging from 0.07 (95% CI: 0.05,0.08) in girls trunklift-TAKS reading to 0.34 (0.32,0.35) in boys cardiovascular-TAKS math. Cardiovascular fitness showed the largest inter-quintile difference in TAKS score (32-75 points), followed by curl-ups. After an additional adjustment for BMI and curl-ups, cardiovascular associations peaked in 8th-9 th grades (maximum inter-quintile difference 142 TAKS points; effect size 0.75 (0.69,0.82) for 8th grade girls math) and showed dose-response characteristics across quintiles (p<0.001 for both genders and outcomes). BMI analysis demonstrated limited, non-linear association with academic performance after adjustment for sociodemographic, cardiovascular fitness and curl-up variables. Low-BMI Hispanic high school boys showed significantly lower TAKS scores than the moderate (but not high) BMI group. High-BMI non-Hispanic white high school girls showed significantly lower scores than the moderate (but not low) BMI group. ^ Conclusions. In this study, fitness was strongly and significantly related to academic performance. Cardiovascular fitness showed a distinct dose-response association with academic performance independent of other sociodemographic and fitness variables. The association peaked in late middle to early high school. The independent association of BMI to academic performance was only found in two sub-groups and was non-linear, with both low and high BMI posing risk relative to moderate BMI but not to each other. In light of our findings, we recommend that policymakers consider PE mandates in middle-high school and require linkage of academic and fitness records to facilitate longitudinal surveillance. School administrators should consider increasing PE time in pursuit of higher academic test scores, and PE practitioners should emphasize cardiovascular fitness over BMI reduction.^
Resumo:
The research project is an extension of a series of administrative science and health care research projects evaluating the influence of external context, organizational strategy, and organizational structure upon organizational success or performance. The research will rely on the assumption that there is not one single best approach to the management of organizations (the contingency theory). As organizational effectiveness is dependent on an appropriate mix of factors, organizations may be equally effective based on differing combinations of factors. The external context of the organization is expected to influence internal organizational strategy and structure and in turn the internal measures affect performance (discriminant theory). The research considers the relationship of external context and organization performance.^ The unit of study for the research will be the health maintenance organization (HMO); an organization the accepts in exchange for a fixed, advance capitation payment, contractual responsibility to assure the delivery of a stated range of health sevices to a voluntary enrolled population. With the current Federal resurgence of interest in the Health Maintenance Organization (HMO) as a major component in the health care system, attention must be directed at maximizing development of HMOs from the limited resources available. Increased skills are needed in both Federal and private evaluation of HMO feasibility in order to prevent resource investment and in projects that will fail while concurrently identifying potentially successful projects that will not be considered using current standards.^ The research considers 192 factors measuring contextual milieu (social, educational, economic, legal, demographic, health and technological factors). Through intercorrelation and principle components data reduction techniques this was reduced to 12 variables. Two measures of HMO performance were identified, they are (1) HMO status (operational or defunct), and (2) a principle components factor score considering eight measures of performance. The relationship between HMO context and performance was analysed using correlation and stepwise multiple regression methods. In each case it has been concluded that the external contextual variables are not predictive of success or failure of study Health Maintenance Organizations. This suggests that performance of an HMO may rely on internal organizational factors. These findings have policy implications as contextual measures are used as a major determinant in HMO feasibility analysis, and as a factor in the allocation of limited Federal funds. ^
Resumo:
An increasing number of neuroimaging studies are concerned with the identification of interactions or statistical dependencies between brain areas. Dependencies between the activities of different brain regions can be quantified with functional connectivity measures such as the cross-correlation coefficient. An important factor limiting the accuracy of such measures is the amount of empirical data available. For event-related protocols, the amount of data also affects the temporal resolution of the analysis. We use analytical expressions to calculate the amount of empirical data needed to establish whether a certain level of dependency is significant when the time series are autocorrelated, as is the case for biological signals. These analytical results are then contrasted with estimates from simulations based on real data recorded with magnetoencephalography during a resting-state paradigm and during the presentation of visual stimuli. Results indicate that, for broadband signals, 50–100 s of data is required to detect a true underlying cross-correlations coefficient of 0.05. This corresponds to a resolution of a few hundred milliseconds for typical event-related recordings. The required time window increases for narrow band signals as frequency decreases. For instance, approximately 3 times as much data is necessary for signals in the alpha band. Important implications can be derived for the design and interpretation of experiments to characterize weak interactions, which are potentially important for brain processing.
Resumo:
This thesis uses models of firm-heterogeneity to complete empirical analyses in economic history and agricultural economics. In Chapter 2, a theoretical model of firm heterogeneity is used to derive a statistic that summarizes the welfare gains from the introduction of a new technology. The empirical application considers the use of mechanical steam power in the Canadian manufacturing sector during the late nineteenth century. I exploit exogenous variation in geography to estimate several parameters of the model. My results indicate that the use of steam power resulted in a 15.1 percent increase in firm-level productivity and a 3.0-5.2 percent increase in aggregate welfare. Chapter 3 considers various policy alternatives to price ceiling legislation in the market for production quotas in the dairy farming sector in Quebec. I develop a dynamic model of the demand for quotas with farmers that are heterogeneous in their marginal cost of milk production. The econometric analysis uses farm-level data and estimates a parameter of the theoretical model that is required for the counterfactual experiments. The results indicate that the price of quotas could be reduced to the ceiling price through a 4.16 percent expansion of the aggregate supply of quotas, or through moderate trade liberalization of Canadian dairy products. In Chapter 4, I study the relationship between farm-level productivity and participation in the Commercial Export Milk (CEM) program. I use a difference-in-difference research design with inverse propensity weights to test for causality between participation in the CEM program and total factor productivity (TFP). I find a positive correlation between participation in the CEM program and TFP, however I find no statistically significant evidence that the CEM program affected TFP.
Resumo:
Sediment cores collected from the Eastern Equatorial Pacific Ocean display a clear positive second-order relationship between wet bulk density (WBD) and carbonate content. This has long interested the paleoceanography community because detailed Gamma Ray Attenuation Porosity Evaluator (GRAPE) measurements, which approximate WBD, might be used to determine records of carbonate content at very high temporal resolution. Although general causes for the relationship are known, they have not been presented and discussed systematically on the basis of first principles. In this study, we measure the mass and carbonate content of 50 sediment samples with known WBD from Site U1338, before and after rinsing with de-ionized water; we also determine the mass related proportion of coarse (> 63 µm) material. Samples exhibit clear relationships between WBD, carbonate content, mass loss upon rinsing, and grain size. We develop a series of mathematical expressions to describe these relationships, and solve them numerically. As noted by previous workers, the second-order relationship between WBD and carbonate content results from the mixing of biogenic carbonate and biogenic silica, which have different grain densities and different porosities. However, at high carbonate content, a wide range in WBD occurs because samples with greater amounts of coarse carbonate have higher porosity. Moreover compaction impacts carbonate particles more than biogenic silica particles. As such, a single two-component equation cannot be used to determine carbonate content accurately across depth intervals where both the porosity and type of carbonate vary. Instead, the WBD-carbonate relationship is described by an infinite series of curves, each which represents mixing of multiple sediment components with different densities and porosities. Dissolved ions also precipitate from pore space during sample drying, which adds mass to the sediment. Without rinsing samples, simple empirical relationships between WBD and carbonate content are further skewed by salt dilution.
Resumo:
Published also as thesis (PH. D.) Columbia University, 1921.
Resumo:
Automaticity (in this essay defined as short response time) and fluency in language use are closely connected to each other and some research has been conducted regarding some of the aspects involved. In fact, the notion of automaticity is still debated and many definitions and opinions on what automaticity is have been suggested (Andersson,1987, 1992, 1993, Logan, 1988, Segalowitz, 2010). One aspect that still needs more research is the correlation between vocabulary proficiency (a person’s knowledge about words and ability to use them correctly) and response time in word recognition. Therefore, the aim of this study has been to investigate this correlation using two different tests; one vocabulary size test (Paul Nation) and one lexical decision task (SuperLab) that measures both response time and accuracy. 23 Swedish students partaking in the English 7 course in upper secondary Swedish school were tested. The data were analyzed using a quantitative method where the average values and correlations from the test were used to compare the results. The correlations were calculated using Pearson’s Coefficient Correlations Calculator. The empirical study indicates that vocabulary proficiency is not strongly correlated with shorter response times in word recognition. Rather, the data indicate that L2 learners instead are sensitive to the frequency levels of the vocabulary. The accuracy (number of correct recognized words) and response times correlate with the frequency level of the tested words. This indicates that factors other than vocabulary proficiency are important for the ability to recognize words quickly.
Resumo:
Water-sampler equilibrium partitioning coefficients and aqueous boundary layer mass transfer coefficients for atrazine, diuron, hexazionone and fluometuron onto C18 and SDB-RPS Empore disk-based aquatic passive samplers have been determined experimentally under a laminar flow regime (Re = 5400). The method involved accelerating the time to equilibrium of the samplers by exposing them to three water concentrations, decreasing stepwise to 50% and then 25% of the original concentration. Assuming first-order Fickian kinetics across a rate-limiting aqueous boundary layer, both parameters are determined computationally by unconstrained nonlinear optimization. In addition, a method of estimation of mass transfer coefficients-therefore sampling rates-using the dimensionless Sherwood correlation developed for laminar flow over a flat plate is applied. For each of the herbicides, this correlation is validated to within 40% of the experimental data. The study demonstrates that for trace concentrations (sub 0.1 mu g/L) and these flow conditions, a naked Empore disk performs well as an integrative sampler over short deployments (up to 7 days) for the range of polar herbicides investigated. The SDB-RPS disk allows a longer integrative period than the C18 disk due to its higher sorbent mass and/or its more polar sorbent chemistry. This work also suggests that for certain passive sampler designs, empirical estimation of sampling rates may be possible using correlations that have been available in the chemical engineering literature for some time.
Resumo:
In this paper the performance of opening and closing returns, for the components of the FT-30 will be studied. It will be shown that for these stocks opening returns have higher volatility and a greater tendency towards negative serial correlation than closing returns. Unlike previous studies this contrasting performance cannot solely be attributed to differences in the trading mechanism across the trading day. All the stocks used in our sample trade thought the day using a uniform trading mechanism. In this paper, we suggest that it is differences in the speed that closing and opening returns adjust to new information that causes differences in return performance. By estimating the Amihud and Mendelson (1987) [Amihud, Yakov, & Mendelson, Haim (1987). Trading mechanisms and stock returns: An empirical investigation, Journal of Finance, 62 533-553.] partial adjustment model with noise, we show that opening returns have a tendency towards over-reaction, while closing returns have a tendency towards under-reaction. We suggest that it is these differences that cause a substantial proportion (although not all) of the asymmetric return patterns associated with opening and closing returns. © 2005 Elsevier Inc. All rights reserved.
Resumo:
An increasing number of neuroimaging studies are concerned with the identification of interactions or statistical dependencies between brain areas. Dependencies between the activities of different brain regions can be quantified with functional connectivity measures such as the cross-correlation coefficient. An important factor limiting the accuracy of such measures is the amount of empirical data available. For event-related protocols, the amount of data also affects the temporal resolution of the analysis. We use analytical expressions to calculate the amount of empirical data needed to establish whether a certain level of dependency is significant when the time series are autocorrelated, as is the case for biological signals. These analytical results are then contrasted with estimates from simulations based on real data recorded with magnetoencephalography during a resting-state paradigm and during the presentation of visual stimuli. Results indicate that, for broadband signals, 50-100 s of data is required to detect a true underlying cross-correlations coefficient of 0.05. This corresponds to a resolution of a few hundred milliseconds for typical event-related recordings. The required time window increases for narrow band signals as frequency decreases. For instance, approximately 3 times as much data is necessary for signals in the alpha band. Important implications can be derived for the design and interpretation of experiments to characterize weak interactions, which are potentially important for brain processing.
Resumo:
The techniques and insights from two distinct areas of financial economic modelling are combined to provide evidence of the influence of firm size on the volatility of stock portfolio returns. Portfolio returns are characterized by positive serial correlation induced by the varying levels of non-synchronous trading among the component stocks. This serial correlation is greatest for portfolios of small firms. The conditional volatility of stock returns has been shown to be well represented by the GARCH family of statistical processes. Using a GARCH model of the variance of capitalization-based portfolio returns, conditioned on the autocorrelation structure in the conditional mean, striking differences related to firm size are uncovered.
Resumo:
Correct specification of the simple location quotients in regionalizing the national direct requirements table is essential to the accuracy of regional input-output multipliers. The purpose of this research is to examine the relative accuracy of these multipliers when earnings, employment, number of establishments, and payroll data specify the simple location quotients.^ For each specification type, I derive a column of total output multipliers and a column of total income multipliers. These multipliers are based on the 1987 benchmark input-output accounts of the U.S. economy and 1988-1992 state of Florida data.^ Error sign tests, and Standardized Mean Absolute Deviation (SMAD) statistics indicate that the output multiplier estimates overestimate the output multipliers published by the Department of Commerce-Bureau of Economic Analysis (BEA) for the state of Florida. In contrast, the income multiplier estimates underestimate the BEA's income multipliers. For a given multiplier type, the Spearman-rank correlation analysis shows that the multiplier estimates and the BEA multipliers have statistically different rank ordering of row elements. The above tests also find no significant different differences, both in size and ranking distributions, among the vectors of multiplier estimates. ^
Resumo:
FDI is believed to be a conduit of new technologies between countries. The first chapter of this dissertation studies the advantages of outward FDI for the home country of multinationals conducting research and development abroad. We use patent citations as a proxy for technology spillovers and we bring empirical evidence that supports the hypothesis that a U.S. subsidiary conducting research and development overseas facilitates the flow of knowledge between its host and home countries.^ The second chapter examines the impact of intellectual property rights (IPR) reforms on the technology flows between the U.S. and host countries of U.S. affiliates. We again use patent citations to examine whether the diffusion of new technology between the host countries and the U.S. is accelerated by the reforms. Our results suggest that the reforms favor innovative efforts of domestic firms in the reforming countries rather than U.S. affiliates efforts. In other words, reforms mediate the technology flows from the U.S. to the reforming countries.^ The third chapter deals with another form of IPR, open source (OS) licenses. These differ in the conditions under which licensors and OS contributors are allowed to modify and redistribute the source code. We measure OS project quality by the speed with which programming bugs are fixed and test whether the license chosen by project leaders influences bug resolution rates. In initial regressions, we find a strong correlation between the hazard of bug resolution and the use of highly restrictive licenses. However, license choices are likely to be endogenous. We instrument license choice using (i) the human language in which contributors operate and (ii) the license choice of the project leaders for a previous project. We then find weak evidence that restrictive licenses adversely affect project success.^
Resumo:
This study examined Kirkpatrick’s training evaluation model (Kirkpatrick & Kirkpatrick, 2006) by assessing a sales training program conducted at an organization in the hospitality industry. The study assessed the employees’ training outcomes of knowledge and skills, job performance, and the impact of the training upon the organization. By assessing these training outcomes and their relationships, the study demonstrated whether Kirkpatrick’s theories are supported and the lower evaluation levels can be used to predict organizational impact. The population for this study was a group of reservations sales agents from a leading luxury hotel chain’s reservations center. During the study period from January 2005 to May 2007, there were 335 reservations sales agents employed in this Global Reservations Center (GRC). The number of reservations sales agents who had completed a sales training program/intervention during this period and had data available for at least two months pre and post training composed the sample for this study. The number of agents was 69 ( N = 69). Four hypotheses were tested through paired-samples t tests, correlation, and hierarchical regression analytic procedures. Results from the analyses supported the hypotheses in this study. The significant improvement in the call score supported hypothesis one that the reservations sales agents who completed the training improved their knowledge of content and required skills in handling calls (Level 2). Hypothesis two was accepted in part as there was significant improvement in call conversion, but there was no significant improvement of time usage. The significant improvement in the sales per call supported hypothesis three that the reservations agents who completed the training contributed to increased organizational impact (Level 4), i.e., made significantly more sales. Last, findings supported hypothesis four that Level 2 and Level 3 variables can be used for predicting Level 4 organizational impact. The findings supported the theory of Kirkpatrick’s evaluation model that in order to expect organizational results, a positive change in behavior (job performance) and learning must occur. The examinations of Levels 2 and 3 helped to partially explain and predict Level 4 results.
Resumo:
This ex post facto study (N = 209) examined the relationships between employer job strategies and job retention among organizations participating in Florida welfare-to-work network programs and associated the strategies with job retention data to determine best practices. ^ An internet-based self-report survey battery was administered to a heterogeneous sampling of organizations participating in the Florida welfare-to-work network program. Hypotheses were tested through correlational and hierarchical regression analytic procedures. The partial correlation results linked each of the job retention strategies to job retention. Wages, benefits, training and supervision, communication, job growth, work/life balance, fairness and respect were all significantly related to job retention. Hierarchical regression results indicated that the training and supervision variable was the best predictor of job retention in the regression equation. ^ The size of the organization was also a significant predictor of job retention. Large organizations reported higher job retention rates than small organizations. There was no statistical difference between the types of organizations (profit-making and non-profit) and job retention. The standardized betas ranged from to .26 to .41 in the regression equation. Twenty percent of the variance in job retention was explained by the combination of demographic and job retention strategy predictors, supporting the theoretical, empirical, and practical relevance of understanding the association between employer job strategies and job retention outcomes. Implications for adult education and human resource development theory, research, and practice are highlighted as possible strategic leverage points for creating conditions that facilitate the development of job strategies as a means for improving former welfare workers’ job retention.^