965 resultados para A posteriori error estimation


Relevância:

20.00% 20.00%

Publicador:

Resumo:

Minimax lower bounds for concept learning state, for example, thatfor each sample size $n$ and learning rule $g_n$, there exists a distributionof the observation $X$ and a concept $C$ to be learnt such that the expectederror of $g_n$ is at least a constant times $V/n$, where $V$ is the VC dimensionof the concept class. However, these bounds do not tell anything about therate of decrease of the error for a {\sl fixed} distribution--concept pair.\\In this paper we investigate minimax lower bounds in such a--stronger--sense.We show that for several natural $k$--parameter concept classes, includingthe class of linear halfspaces, the class of balls, the class of polyhedrawith a certain number of faces, and a class of neural networks, for any{\sl sequence} of learning rules $\{g_n\}$, there exists a fixed distributionof $X$ and a fixed concept $C$ such that the expected error is larger thana constant times $k/n$ for {\sl infinitely many n}. We also obtain suchstrong minimax lower bounds for the tail distribution of the probabilityof error, which extend the corresponding minimax lower bounds.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Returns to scale to capital and the strength of capital externalities play a key role for the empirical predictions and policy implications of different growth theories. We show that both can be identified with individual wage data and implement our approach at the city-level using US Census data on individuals in 173 cities for 1970, 1980, and 1990. Estimation takes into account fixed effects, endogeneity of capital accumulation, and measurement error. We find no evidence for human or physical capital externalities and decreasing aggregate returns to capital. Returns to scale to physical and human capital are around 80 percent. We also find strong complementarities between human capital and labor and substantial total employment externalities.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

We introduce simple nonparametric density estimators that generalize theclassical histogram and frequency polygon. The new estimators are expressed as linear combination of density functions that are piecewisepolynomials, where the coefficients are optimally chosen in order to minimize the integrated square error of the estimator. We establish the asymptotic behaviour of the proposed estimators, and study theirperformance in a simulation study.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

BACKGROUND: Creatinine clearance is the most common method used to assess glomerular filtration rate (GFR). In children, GFR can also be estimated without urine collection, using the formula GFR (mL/min x 1.73 m2) = K x height [cm]/Pcr [mumol/L]), where Pcr represents the plasma creatinine concentration. K is usually calculated using creatinine clearance (Ccr) as an index of GFR. The aim of the present study was to evaluate the reliability of the formula, using the standard UV/P inulin clearance to calculate K. METHODS: Clearance data obtained in 200 patients (1 month to 23 years) during the years 1988-1994 were used to calculate the factor K as a function of age. Forty-four additional patients were studied prospectively in conditions of either hydropenia or water diuresis in order to evaluate the possible variation of K as a function of urine flow rate. RESULTS: When GFR was estimated by the standard inulin clearance, the calculated values of K was 39 (infants less than 6 months), 44 (1-2 years) and 47 (2-12 years). The correlation between the values of GFR, as estimated by the formula, and the values measured by the standard clearance of inulin was highly significant; the scatter of individual values was however substantial. When K was calculated using Ccr, the formula overestimated Cin at all urine flow rates. When calculated from Ccr, K varied as a function of urine flow rate (K = 50 at urine flow rates of 3.5 and K = 64 at urine flow rates of 8.5 mL/min x 1.73 m2). When calculated from Cin, in the same conditions, K remained constant with a value of 50. CONCLUSIONS: The formula GFR = K x H/Pcr can be used to estimate GFR. The scatter of values precludes however the use of the formula to estimate GFR in pathophysiological studies. The formula should only be used when K is calculated from Cin, and the plasma creatinine concentration is measured in well defined conditions of hydration.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Medical errors compromise patient safety in ambulatory practice. These errors must be faced in a framework that reduces to a minimum their consequences for the patients. This approach relies on the implementation of a new culture without stigmatization and where errors are disclosed to the patients; this culture implies the build up of a system for reporting errors associated to an in-depth analysis of the system, looking for root causes and insufficient barriers with the aim to fix them. A useful education tool is the "critical situations" meeting during which physicians are encouraged to openly present adverse events and "near misses". Their analysis, with supportive attitude towards involved staff members, allows to reveal systems failures within the institution or the private practice.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

For the standard kernel density estimate, it is known that one can tune the bandwidth such that the expected L1 error is within a constant factor of the optimal L1 error (obtained when one is allowed to choose the bandwidth with knowledge of the density). In this paper, we pose the same problem for variable bandwidth kernel estimates where the bandwidths are allowed to depend upon the location. We show in particular that for positive kernels on the real line, for any data-based bandwidth, there exists a densityfor which the ratio of expected L1 error over optimal L1 error tends to infinity. Thus, the problem of tuning the variable bandwidth in an optimal manner is ``too hard''. Moreover, from the class of counterexamples exhibited in the paper, it appears thatplacing conditions on the densities (monotonicity, convexity, smoothness) does not help.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The central message of this paper is that nobody should be using the samplecovariance matrix for the purpose of portfolio optimization. It containsestimation error of the kind most likely to perturb a mean-varianceoptimizer. In its place, we suggest using the matrix obtained from thesample covariance matrix through a transformation called shrinkage. Thistends to pull the most extreme coefficients towards more central values,thereby systematically reducing estimation error where it matters most.Statistically, the challenge is to know the optimal shrinkage intensity,and we give the formula for that. Without changing any other step in theportfolio optimization process, we show on actual stock market data thatshrinkage reduces tracking error relative to a benchmark index, andsubstantially increases the realized information ratio of the activeportfolio manager.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

We compare a set of empirical Bayes and composite estimators of the population means of the districts (small areas) of a country, and show that the natural modelling strategy of searching for a well fitting empirical Bayes model and using it for estimation of the area-level means can be inefficient.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This paper proposes to estimate the covariance matrix of stock returnsby an optimally weighted average of two existing estimators: the samplecovariance matrix and single-index covariance matrix. This method isgenerally known as shrinkage, and it is standard in decision theory andin empirical Bayesian statistics. Our shrinkage estimator can be seenas a way to account for extra-market covariance without having to specifyan arbitrary multi-factor structure. For NYSE and AMEX stock returns from1972 to 1995, it can be used to select portfolios with significantly lowerout-of-sample variance than a set of existing estimators, includingmulti-factor models.