922 resultados para Measurement Error Estimation


Relevância:

80.00% 80.00%

Publicador:

Resumo:

Models of snow processes in areas of possible large-scale change need to be site independent and physically based. Here, the accumulation and ablation of the seasonal snow cover beneath a fir canopy has been simulated with a new physically based snow-soil vegetation-atmosphere transfer scheme (Snow-SVAT) called SNOWCAN. The model was formulated by coupling a canopy optical and thermal radiation model to a physically based multilayer snow model. Simple representations of other forest effects were included. These include the reduction of wind speed and hence turbulent transfer beneath the canopy, sublimation of intercepted snow, and deposition of debris on the surface. This paper tests this new modeling approach fully at a fir site within Reynolds Creek Experimental Watershed, Idaho. Model parameters were determined at an open site and subsequently applied to the fir site. SNOWCAN was evaluated using measurements of snow depth, subcanopy solar and thermal radiation, and snowpack profiles of temperature, density, and grain size. Simulations showed good agreement with observations (e.g., fir site snow depth was estimated over the season with r(2) = 0.96), generally to within measurement error. However, the simulated temperature profiles were less accurate after a melt-freeze event, when the temperature discrepancy resulted from underestimation of the rate of liquid water flow and/or the rate of refreeze. This indicates both that the general modeling approach is applicable and that a still more complete representation of liquid water in the snowpack will be important.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Accurately measured peptide masses can be used for large-scale protein identification from bacterial whole-cell digests as an alternative to tandem mass spectrometry (MS/MS) provided mass measurement errors of a few parts-per-million (ppm) are obtained. Fourier transform ion cyclotron resonance (FTICR) mass spectrometry (MS) routinely achieves such mass accuracy either with internal calibration or by regulating the charge in the analyzer cell. We have developed a novel and automated method for internal calibration of liquid chromatography (LC)/FTICR data from whole-cell digests using peptides in the sample identified by concurrent MS/MS together with ambient polydimethyl-cyclosiloxanes as internal calibrants in the mass spectra. The method reduced mass measurement error from 4.3 +/- 3.7 ppm to 0.3 +/- 2.3 ppm in an E. coli LC/FTICR dataset of 1000 MS and MS/MS spectra and is applicable to all analyses of complex protein digests by FTICRMS. Copyright (c) 2006 John Wiley & Sons, Ltd.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

The purpose of this study was to apply and compare two time-domain analysis procedures in the determination of oxygen uptake (VO2) kinetics in response to a pseudorandom binary sequence (PRBS) exercise test. PRBS exercise tests have typically been analysed in the frequency domain. However, the complex interpretation of frequency responses may have limited the application of this procedure in both sporting and clinical contexts, where a single time measurement would facilitate subject comparison. The relative potential of both a mean response time (MRT) and a peak cross-correlation time (PCCT) was investigated. This study was divided into two parts: a test-retest reliability study (part A), in which 10 healthy male subjects completed two identical PRBS exercise tests, and a comparison of the VO2 kinetics of 12 elite endurance runners (ER) and 12 elite sprinters (SR; part B). In part A, 95% limits of agreement were calculated for comparison between MRT and PCCT. The results of part A showed no significant difference between test and retest as assessed by MRT [mean (SD) 42.2 (4.2) s and 43.8 (6.9) s] or by PCCT [21.8 (3.7) s and 22.7 (4.5) s]. Measurement error (%) was lower for MRT in comparison with PCCT (16% and 25%, respectively). In part B of the study, the VO2 kinetics of ER were significantly faster than those of SR, as assessed by MRT [33.4 (3.4) s and 39.9 (7.1) s, respectively; P<0.01] and PCCT [20.9 (3.8) s and 24.8 (4.5) s; P < 0.05]. It is possible that either analysis procedure could provide a single test measurement Of VO2 kinetics; however, the greater reliability of the MRT data suggests that this method has more potential for development in the assessment Of VO2 kinetics by PRBS exercise testing.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

For the first time, vertical column measurements of (HNO3) above the Arctic Stratospheric Ozone Observatory (AStrO) at Eureka (80N, 86W), Canada, have been made during polar night using lunar spectra recorded with a Fourier Transform Infrared (FTIR) spectrometer, from October 2001 to March 2002. AStrO is part of the primary Arctic station of the Network for the Detection of Stratospheric Change (NDSC). These measurements were compared with FTIR measurements at two other NDSC Arctic sites: Thule, Greenland (76.5N, 68.8W) and Kiruna, Sweden (67.8N, 20.4E). The measurements were also compared with two atmospheric models: the Canadian Middle Atmosphere Model (CMAM) and SLIMCAT. This is the first time that CMAM HNO3 columns have been compared with observations in the Arctic. Eureka lunar measurements are in good agreement with solar ones made with the same instrument. Eureka and Thule HNO3 columns are consistent within measurement error. Differences among HNO3 columns measured at Kiruna and those measured at Eureka and Thule can be explained on the basis of the available sunlight hours and the polar vortex location. The comparison of CMAM HNO3 columns with Eureka and Kiruna data shows good agreement, considering CMAM small inter-annual variability. The warm 2001/02 winter with almost no Polar Stratospheric Clouds (PSCs) makes the comparison of the warm climate version of CMAM with these observations a good test for CMAM under no PSC conditions. SLIMCAT captures the magnitude of HNO3 columns at Eureka, and the day-to-day variability, but generally reports higher HNO3 columns than the CMAM climatological mean columns.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

We present a novel method for retrieving high-resolution, three-dimensional (3-D) nonprecipitating cloud fields in both overcast and broken-cloud situations. The method uses scanning cloud radar and multiwavelength zenith radiances to obtain gridded 3-D liquid water content (LWC) and effective radius (re) and 2-D column mean droplet number concentration (Nd). By using an adaption of the ensemble Kalman filter, radiances are used to constrain the optical properties of the clouds using a forward model that employs full 3-D radiative transfer while also providing full error statistics given the uncertainty in the observations. To evaluate the new method, we first perform retrievals using synthetic measurements from a challenging cumulus cloud field produced by a large-eddy simulation snapshot. Uncertainty due to measurement error in overhead clouds is estimated at 20% in LWC and 6% in re, but the true error can be greater due to uncertainties in the assumed droplet size distribution and radiative transfer. Over the entire domain, LWC and re are retrieved with average error 0.05–0.08 g m-3 and ~2 μm, respectively, depending on the number of radiance channels used. The method is then evaluated using real data from the Atmospheric Radiation Measurement program Mobile Facility at the Azores. Two case studies are considered, one stratocumulus and one cumulus. Where available, the liquid water path retrieved directly above the observation site was found to be in good agreement with independent values obtained from microwave radiometer measurements, with an error of 20 g m-2.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Satellite-based (e.g., Synthetic Aperture Radar [SAR]) water level observations (WLOs) of the floodplain can be sequentially assimilated into a hydrodynamic model to decrease forecast uncertainty. This has the potential to keep the forecast on track, so providing an Earth Observation (EO) based flood forecast system. However, the operational applicability of such a system for floods developed over river networks requires further testing. One of the promising techniques for assimilation in this field is the family of ensemble Kalman (EnKF) filters. These filters use a limited-size ensemble representation of the forecast error covariance matrix. This representation tends to develop spurious correlations as the forecast-assimilation cycle proceeds, which is a further complication for dealing with floods in either urban areas or river junctions in rural environments. Here we evaluate the assimilation of WLOs obtained from a sequence of real SAR overpasses (the X-band COSMO-Skymed constellation) in a case study. We show that a direct application of a global Ensemble Transform Kalman Filter (ETKF) suffers from filter divergence caused by spurious correlations. However, a spatially-based filter localization provides a substantial moderation in the development of the forecast error covariance matrix, directly improving the forecast and also making it possible to further benefit from a simultaneous online inflow error estimation and correction. Additionally, we propose and evaluate a novel along-network metric for filter localization, which is physically-meaningful for the flood over a network problem. Using this metric, we further evaluate the simultaneous estimation of channel friction and spatially-variable channel bathymetry, for which the filter seems able to converge simultaneously to sensible values. Results also indicate that friction is a second order effect in flood inundation models applied to gradually varied flow in large rivers. The study is not conclusive regarding whether in an operational situation the simultaneous estimation of friction and bathymetry helps the current forecast. Overall, the results indicate the feasibility of stand-alone EO-based operational flood forecasting.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

The Natural History of Human Papillomavirus (HPV) Infection in Men: The HIM Study is a prospective multi-center cohort study that, among other factors, analyzes participants` diet. A parallel cross-sectional study was designed to evaluate the validity and reproducibility of the quantitative food frequency questionnaire (QFFQ) used in the Brazilian center from the HIM Study. For this, a convenience subsample of 98 men aged 18 to 70 years from the HIM Study in Brazil answered three 54-item QFFQ and three 24-hour recall interviews, with 6-month intervals between them (data collection January to September 2007). A Bland-Altman analysis indicated that the difference between instruments was dependent on the magnitude of the intake for energy and most nutrients included in the validity analysis, with the exception of carbohydrates, fiber, polyunsaturated fat, vitamin C, and vitamin E. The correlation between the QFFQ and the 24-hour recall for the deattenuated and energy-adjusted data ranged from 0.05 (total fat) to 0.57 (calcium). For the energy and nutrients consumption included in the validity analysis, 33.5% of participants on average were correctly classified into quartiles, and the average value of 0.26 for weighted kappa shows a reasonable agreement. The intraclass correlation coefficients for all nutrients were greater than 0.40 in the reproducibility analysis. The QFFQ demonstrated good reproducibility and acceptable validity. The results support the use of this instrument in the HIM Study. J Am Diet Assoc. 2011;111:1045-1051.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

This paper presents a new framework for generating triangular meshes from textured color images. The proposed framework combines a texture classification technique, called W-operator, with Imesh, a method originally conceived to generate simplicial meshes from gray scale images. An extension of W-operators to handle textured color images is proposed, which employs a combination of RGB and HSV channels and Sequential Floating Forward Search guided by mean conditional entropy criterion to extract features from the training data. The W-operator is built into the local error estimation used by Imesh to choose the mesh vertices. Furthermore, the W-operator also enables to assign a label to the triangles during the mesh construction, thus allowing to obtain a segmented mesh at the end of the process. The presented results show that the combination of W-operators with Imesh gives rise to a texture classification-based triangle mesh generation framework that outperforms pixel based methods. Crown Copyright (C) 2009 Published by Elsevier Inc. All rights reserved.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

In chemical analyses performed by laboratories, one faces the problem of determining the concentration of a chemical element in a sample. In practice, one deals with the problem using the so-called linear calibration model, which considers that the errors associated with the independent variables are negligible compared with the former variable. In this work, a new linear calibration model is proposed assuming that the independent variables are subject to heteroscedastic measurement errors. A simulation study is carried out in order to verify some properties of the estimators derived for the new model and it is also considered the usual calibration model to compare it with the new approach. Three applications are considered to verify the performance of the new approach. Copyright (C) 2010 John Wiley & Sons, Ltd.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

In this paper we deal with the issue of performing accurate testing inference on a scalar parameter of interest in structural errors-in-variables models. The error terms are allowed to follow a multivariate distribution in the class of the elliptical distributions, which has the multivariate normal distribution as special case. We derive a modified signed likelihood ratio statistic that follows a standard normal distribution with a high degree of accuracy. Our Monte Carlo results show that the modified test is much less size distorted than its unmodified counterpart. An application is presented.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Esta dissertação de mestrado em economia foi motivada por uma questão complexa bastante estudada na literatura de economia política nos dias de hoje: as formas como campanhas políticas afetam votação em uma eleição. estudo procura modelar mercado eleitoral brasileiro para deputados federais senadores. Através de um modelo linear, conclui-se que os gastos em campanha eleitoral são fatores decisivos para eleição de um candidato deputado federal. Após reconhecer que variável que mede os gastos em campanha possui erro de medida (devido ao famoso "caixa dois", por exemplo), além de ser endógena uma vez que candidatos com maiores possibilidades de conseguir votos conseguem mais fontes de financiamento -, modelo foi estimado por variáveis instrumentais. Para senadores, utilizando modelos lineares modelos com variável resposta binaria, verifica-se também importância, ainda que em menor escala, da campanha eleitoral, sendo que um fator mais importante para corrida ao senado parece ser uma percepção priori da qualidade do candidato.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

O objetivo deste trabalho é verificar se o ajustamento das condições de paridade de juros por expectativa do mercado (paridade descoberta) e por prêmios de risco (paridades coberta e descoberta) leva à validação da relação de não-arbitragem subjacente, ou pelo menos a resultados econométricos mais próximos de sua validação. Para isso, combinamos taxas de retornos de instrumentos de renda fixa domésticos e norte-americanos e aplicamos o arcabouço econométrico de séries de tempo. Como primeiro passo de investigação, aplicamos a paridade de juros (descoberta e coberta) na sua forma tradicional. No passo seguinte aplicamos os testes econométricos às condições de paridade ajustadas por um prêmio de risco. No caso da PDJ, não obtivemos resultados satisfatórios, mesmo ajustando pelos prêmios de risco. Esse ajuste propiciou uma mudança nos sinais dos coeficientes na direção correta, mas a magnitude do coeficiente da desvalorização cambial efetiva passou a destoar bastante da magnitude das outras séries. Apesar de termos obtido a validade da PCJ na forma tradicional, não esperaríamos este resultado, pois isso implicaria que o prêmio de risco país seria nulo para este período. Ajustando a PCJ pelo prêmio de risco de não-pagamento passa-se a não obter co integração entre as séries, ou seja, o prêmio de risco de não-pagamento teria um comportamento independente do prêmio futuro e do diferencial de juros. As possíveis causas para a não obtenção dos resultados esperados são: intervalo amostraI menor que 3 anos, erro de medida dos dados de survey ou tentativa do Banco Central de controlar a taxa de câmbio nominal e as taxas de juros domésticas simultaneamente.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

In this work, we study the survival cure rate model proposed by Yakovlev et al. (1993), based on a competing risks structure concurring to cause the event of interest, and the approach proposed by Chen et al. (1999), where covariates are introduced to model the risk amount. We focus the measurement error covariates topics, considering the use of corrected score method in order to obtain consistent estimators. A simulation study is done to evaluate the behavior of the estimators obtained by this method for finite samples. The simulation aims to identify not only the impact on the regression coefficients of the covariates measured with error (Mizoi et al. 2007) but also on the coefficients of covariates measured without error. We also verify the adequacy of the piecewise exponential distribution to the cure rate model with measurement error. At the end, model applications involving real data are made

Relevância:

80.00% 80.00%

Publicador:

Resumo:

A common approach used to estimate landscape resistance involves comparing correlations of ecological and genetic distances calculated among individuals of a species. However, the location of sampled individuals may contain some degree of spatial uncertainty due to the natural variation of animals moving through their home range or measurement error in plant or animal locations. In this study, we evaluate the ways that spatial uncertainty, landscape characteristics, and genetic stochasticity interact to influence the strength and variability of conclusions about landscape-genetics relationships. We used a neutral landscape model to generate 45 landscapes composed of habitat and non-habitat, varying in percent habitat, aggregation, and structural connectivity (patch cohesion). We created true and alternate locations for 500 individuals, calculated ecological distances (least-cost paths), and simulated genetic distances among individuals. We compared correlations between ecological distances for true and alternate locations. We then simulated genotypes at 15 neutral loci and investigated whether the same influences could be detected in simple Mantel tests and while controlling for the effects of isolation-by distance using the partial Mantel test. Spatial uncertainty interacted with the percentage of habitat in the landscape, but led to only small reductions in correlations. Furthermore, the strongest correlations occurred with low percent habitat, high aggregation, and low to intermediate levels of cohesion. Overall genetic stochasticity was relatively low and was influenced by landscape characteristics.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Com o objetivo de verificar a variabilidade temporal e espacial do tamanho de amostra da radiação solar global média decendial, de 22 locais do Estado do Rio Grande do Sul, utilizaram-se séries de dados de radiação solar global do período de 1956 a 2003. Determinou-se o tamanho de amostra da radiação solar global média decendial em cada decêndio e local e agruparam-se os decêndios e os locais pelo método hierárquico 'vizinho mais distante'. Há variabilidade do tamanho de amostra (número de anos) para a estimativa da radiação solar global média decendial no Estado do Rio Grande do Sul no tempo e no espaço. Maior tamanho é necessário nos decêndios dos meses de junho, julho, agosto e setembro em relação aos outros meses. Para os locais e decêndios estudados, 30 anos de observações são suficientes para estimar a média (µ) de radiação solar global média decendial, para um erro de estimação igual a 12.3%, com coeficiente de confiança de 95%.