49 resultados para Kernel estimator and ROC-GLM methodology
em CentAUR: Central Archive University of Reading - UK
Resumo:
This correspondence introduces a new orthogonal forward regression (OFR) model identification algorithm using D-optimality for model structure selection and is based on an M-estimators of parameter estimates. M-estimator is a classical robust parameter estimation technique to tackle bad data conditions such as outliers. Computationally, The M-estimator can be derived using an iterative reweighted least squares (IRLS) algorithm. D-optimality is a model structure robustness criterion in experimental design to tackle ill-conditioning in model Structure. The orthogonal forward regression (OFR), often based on the modified Gram-Schmidt procedure, is an efficient method incorporating structure selection and parameter estimation simultaneously. The basic idea of the proposed approach is to incorporate an IRLS inner loop into the modified Gram-Schmidt procedure. In this manner, the OFR algorithm for parsimonious model structure determination is extended to bad data conditions with improved performance via the derivation of parameter M-estimators with inherent robustness to outliers. Numerical examples are included to demonstrate the effectiveness of the proposed algorithm.
Resumo:
This paper makes a theoretical case for using these two systems approaches together. The theoretical and methodological assumptions of system dynamics (SD) and soft system methodology (SSM) are briefly described and a partial critique is presented. SSM generates and represents diverse perspectives on a problem situation and addresses the socio-political elements of an intervention. However, it is weak in ensuring `dynamic coherence'. consistency between the intuitive behaviour resulting from proposed changes and behaviour deduced from ideas on causal structure. Conversely, SD examines causal structures and dynamic behaviours. However, whilst emphasising the need for a clear issue focus, it has little theory for generating and representing diverse issues. Also, there is no theory for facilitating sensitivity to socio-political elements. A synthesis of the two called ‘Holon Dynamics' is proposed. After an SSM intervention, a second stage continues the socio-political analysis and also operates within a new perspective which values dynamic coherence of the mental construct - the holon - which is capable of expressing the proposed changes. A model of this holon is constructed using SD and the changes are thus rendered `systemically desirable' in the additional sense that dynamic consistency has been confirmed. The paper closes with reflections on the proposal and the need for theoretical consistency when mixing tools is emphasised.
Resumo:
Collocations between two satellite sensors are occasions where both sensors observe the same place at roughly the same time. We study collocations between the Microwave Humidity Sounder (MHS) on-board NOAA-18 and the Cloud Profiling Radar (CPR) on-board CloudSat. First, a simple method is presented to obtain those collocations and this method is compared with a more complicated approach found in literature. We present the statistical properties of the collocations, with particular attention to the effects of the differences in footprint size. For 2007, we find approximately two and a half million MHS measurements with CPR pixels close to their centrepoints. Most of those collocations contain at least ten CloudSat pixels and image relatively homogeneous scenes. In the second part, we present three possible applications for the collocations. Firstly, we use the collocations to validate an operational Ice Water Path (IWP) product from MHS measurements, produced by the National Environment Satellite, Data and Information System (NESDIS) in the Microwave Surface and Precipitation Products System (MSPPS). IWP values from the CloudSat CPR are found to be significantly larger than those from the MSPPS. Secondly, we compare the relation between IWP and MHS channel 5 (190.311 GHz) brightness temperature for two datasets: the collocated dataset, and an artificial dataset. We find a larger variability in the collocated dataset. Finally, we use the collocations to train an Artificial Neural Network and describe how we can use it to develop a new MHS-based IWP product. We also study the effect of adding measurements from the High Resolution Infrared Radiation Sounder (HIRS), channels 8 (11.11 μm) and 11 (8.33 μm). This shows a small improvement in the retrieval quality. The collocations described in the article are available for public use.
Resumo:
Many kernel classifier construction algorithms adopt classification accuracy as performance metrics in model evaluation. Moreover, equal weighting is often applied to each data sample in parameter estimation. These modeling practices often become problematic if the data sets are imbalanced. We present a kernel classifier construction algorithm using orthogonal forward selection (OFS) in order to optimize the model generalization for imbalanced two-class data sets. This kernel classifier identification algorithm is based on a new regularized orthogonal weighted least squares (ROWLS) estimator and the model selection criterion of maximal leave-one-out area under curve (LOO-AUC) of the receiver operating characteristics (ROCs). It is shown that, owing to the orthogonalization procedure, the LOO-AUC can be calculated via an analytic formula based on the new regularized orthogonal weighted least squares parameter estimator, without actually splitting the estimation data set. The proposed algorithm can achieve minimal computational expense via a set of forward recursive updating formula in searching model terms with maximal incremental LOO-AUC value. Numerical examples are used to demonstrate the efficacy of the algorithm.
Resumo:
Dense deployments of wireless local area networks (WLANs) are becoming a norm in many cities around the world. However, increased interference and traffic demands can severely limit the aggregate throughput achievable unless an effective channel assignment scheme is used. In this work, a simple and effective distributed channel assignment (DCA) scheme is proposed. It is shown that in order to maximise throughput, each access point (AP) simply chooses the channel with the minimum number of active neighbour nodes (i.e. nodes associated with neighbouring APs that have packets to send). However, application of such a scheme to practice depends critically on its ability to estimate the number of neighbour nodes in each channel, for which no practical estimator has been proposed before. In view of this, an extended Kalman filter (EKF) estimator and an estimate of the number of nodes by AP are proposed. These not only provide fast and accurate estimates but can also exploit channel switching information of neighbouring APs. Extensive packet level simulation results show that the proposed minimum neighbour and EKF estimator (MINEK) scheme is highly scalable and can provide significant throughput improvement over other channel assignment schemes.
Resumo:
This paper presents the major characteristics of the Institut Pierre Simon Laplace (IPSL) coupled ocean–atmosphere general circulation model. The model components and the coupling methodology are described, as well as the main characteristics of the climatology and interannual variability. The model results of the standard version used for IPCC climate projections, and for intercomparison projects like the Paleoclimate Modeling Intercomparison Project (PMIP 2) are compared to those with a higher resolution in the atmosphere. A focus on the North Atlantic and on the tropics is used to address the impact of the atmosphere resolution on processes and feedbacks. In the North Atlantic, the resolution change leads to an improved representation of the storm-tracks and the North Atlantic oscillation. The better representation of the wind structure increases the northward salt transports, the deep-water formation and the Atlantic meridional overturning circulation. In the tropics, the ocean–atmosphere dynamical coupling, or Bjerknes feedback, improves with the resolution. The amplitude of ENSO (El Niño-Southern oscillation) consequently increases, as the damping processes are left unchanged.
Resumo:
The problem of estimating the individual probabilities of a discrete distribution is considered. The true distribution of the independent observations is a mixture of a family of power series distributions. First, we ensure identifiability of the mixing distribution assuming mild conditions. Next, the mixing distribution is estimated by non-parametric maximum likelihood and an estimator for individual probabilities is obtained from the corresponding marginal mixture density. We establish asymptotic normality for the estimator of individual probabilities by showing that, under certain conditions, the difference between this estimator and the empirical proportions is asymptotically negligible. Our framework includes Poisson, negative binomial and logarithmic series as well as binomial mixture models. Simulations highlight the benefit in achieving normality when using the proposed marginal mixture density approach instead of the empirical one, especially for small sample sizes and/or when interest is in the tail areas. A real data example is given to illustrate the use of the methodology.
Resumo:
Background: The present paper investigates the question of a suitable basic model for the number of scrapie cases in a holding and applications of this knowledge to the estimation of scrapie-ffected holding population sizes and adequacy of control measures within holding. Is the number of scrapie cases proportional to the size of the holding in which case it should be incorporated into the parameter of the error distribution for the scrapie counts? Or, is there a different - potentially more complex - relationship between case count and holding size in which case the information about the size of the holding should be better incorporated as a covariate in the modeling? Methods: We show that this question can be appropriately addressed via a simple zero-truncated Poisson model in which the hypothesis of proportionality enters as a special offset-model. Model comparisons can be achieved by means of likelihood ratio testing. The procedure is illustrated by means of surveillance data on classical scrapie in Great Britain. Furthermore, the model with the best fit is used to estimate the size of the scrapie-affected holding population in Great Britain by means of two capture-recapture estimators: the Poisson estimator and the generalized Zelterman estimator. Results: No evidence could be found for the hypothesis of proportionality. In fact, there is some evidence that this relationship follows a curved line which increases for small holdings up to a maximum after which it declines again. Furthermore, it is pointed out how crucial the correct model choice is when applied to capture-recapture estimation on the basis of zero-truncated Poisson models as well as on the basis of the generalized Zelterman estimator. Estimators based on the proportionality model return very different and unreasonable estimates for the population sizes. Conclusion: Our results stress the importance of an adequate modelling approach to the association between holding size and the number of cases of classical scrapie within holding. Reporting artefacts and speculative biological effects are hypothesized as the underlying causes of the observed curved relationship. The lack of adjustment for these artefacts might well render ineffective the current strategies for the control of the disease.
Resumo:
Purpose – The main aim of this paper is to present the results of a study examining managers' attitudes towards the deployment and use of information and communications technology (ICT) in their organisations. The study comes at a time when ICT is being recognised as a major enabler of innovation and new business models, which have the potential to have major impact on western economies and jobs. Design/methodology/approach – A questionnaire was specially designed to collect data relating to three research questions. The questionnaire also included a number of open-ended questions. A total of 181 managers from a wide range of industries across a number of countries participated in the electronic survey. The quantitative responses to the survey were analysed using SPSS. Exploratory factor analysis using Varimax rotation was used and ANOVA to compare responses by different groups. Findings – The survey showed that many of the respondents appeared equipped to work “any place, any time”. However, it also highlighted the challenges managers face in working in a connected operation. Also, the data suggested that many managers were less than confident about their companies' policies and practices in relation to information management. Originality/value – A next step from this exploratory research could be the development of a model exploring the impact of ICT on management and organisational performance in terms of personal characteristics of the manager, the role performed, the context and the ICT provision. Also, further research could focus on examining in more detail differences between management levels.
Resumo:
Using the classical Parzen window (PW) estimate as the desired response, the kernel density estimation is formulated as a regression problem and the orthogonal forward regression technique is adopted to construct sparse kernel density (SKD) estimates. The proposed algorithm incrementally minimises a leave-one-out test score to select a sparse kernel model, and a local regularisation method is incorporated into the density construction process to further enforce sparsity. The kernel weights of the selected sparse model are finally updated using the multiplicative nonnegative quadratic programming algorithm, which ensures the nonnegative and unity constraints for the kernel weights and has the desired ability to reduce the model size further. Except for the kernel width, the proposed method has no other parameters that need tuning, and the user is not required to specify any additional criterion to terminate the density construction procedure. Several examples demonstrate the ability of this simple regression-based approach to effectively construct a SKID estimate with comparable accuracy to that of the full-sample optimised PW density estimate. (c) 2007 Elsevier B.V. All rights reserved.
Resumo:
Epidemiological evidence suggests that polyphenols may, in part, explain the cardioprotective properties of fruits. This review aims to summarise the evidence for the effects of fruit polyphenols on four risk factors of CVD: platelet function, blood pressure, vascular function and blood lipids. This review includes human dietary intervention studies investigating fruits and their polyphenols. There was some evidence to suggest that fruits containing relatively high concentrations of flavonols, anthocyanins and procyanindins, such as pomegranate, purple grapes and berries, were effective at reducing CVD risk factors, particularly with respect to anti-hypertensive effects, inhibition of platelet aggregation and increasing endothelial-dependent vasodilation than other fruits investigated. Flavanone-rich fruits, such as oranges and grapefruits, were reported to have hypocholesterolaemic effects, with little impact on other risk factors being examined. However, the evidence was limited, inconsistent and often inconclusive. This is in part due to the heterogeneity in the design of studies, the lack of controls, the relatively short intervention periods and low power in several studies. Details of the polyphenol content of the fruits investigated were also omitted in some studies, negating comparison of data. It is recommended that large, well-powered, long-term human dietary intervention studies investigating a wider range of fruits are required to confirm these observations. Investigations into the potential synergistic effects of polyphenols on a combination of CVD risk markers, dose–response relationships and standardisation in methodology would facilitate the comparison of studies and also provide valuable information on the types of fruits which could confer protection against CVD.
Resumo:
Information on the distribution and behavior of C fractions in soil particle sizes is crucial for understanding C dynamics in soil. At present little is known about the behavior of the C associated with silt-size particles. We quantified the concentrations, distribution, and enrichment of total C (TC), readily oxidizable C (ROC), hotwater- extractable C (HWC), and cold-water-extractable C (CWC) fractions in coarse (63–20-mm), medium (20–6.3-mm), and fine (6.3–2-mm) silt-size subfractions and in coarse (2000–250 mm) and fine (250–63 mm) sand and clay (<2-mm) soil fractions isolated from bulk soil (<2 mm), and 2- to 4-mm aggregate-size fraction of surface (0–25 cm) and subsurface (25–55 cm) soils under different land uses. All measured C fractions varied significantly across all soil particle-size fractions. The highest C concentrations were associated with the <20-mm soil fractions and peaked in the medium (20–6.3-mm) and fine (6.3–2-mm) silt subfractions in most treatments. Carbon enrichment ratios (ERC) revealed the dual behavior of the C fractions associated with the medium silt-size fraction, demonstrating the simultaneous enrichment of TC and ROC, and the depletion of HWC and CWC fractions. The medium silt (20–6.3-mm) subfraction was identified in this study as a zone where the associated C fractions exhibit transitory qualities. Our results show that investigating subfractions within the silt-size particle fraction provides better understanding of the behavior of C fractions in this soil fraction.
Resumo:
Background FFAR1 receptor is a long chain fatty acid G-protein coupled receptor which is expressed widely, but found in high density in the pancreas and central nervous system. It has been suggested that FFAR1 may play a role in insulin sensitivity, lipotoxicity and is associated with type 2 diabetes. Here we investigate the effect of three common SNPs of FFAR1 (rs2301151; rs16970264; rs1573611) on pancreatic function, BMI, body composition and plasma lipids. Methodology/Principal Findings For this enquiry we used the baseline RISCK data, which provides a cohort of overweight subjects at increased cardiometabolic risk with detailed phenotyping. The key findings were SNPs of the FFAR1 gene region were associated with differences in body composition and lipids, and the effects of the 3 SNPs combined were cumulative on BMI, body composition and total cholesterol. The effects on BMI and body fat were predominantly mediated by rs1573611 (1.06 kg/m2 higher (P = 0.009) BMI and 1.53% higher (P = 0.002) body fat per C allele). Differences in plasma lipids were also associated with the BMI-increasing allele of rs2301151 including higher total cholesterol (0.2 mmol/L per G allele, P = 0.01) and with the variant A allele of rs16970264 associated with lower total (0.3 mmol/L, P = 0.02) and LDL (0.2 mmol/L, P<0.05) cholesterol, but also with lower HDL-cholesterol (0.09 mmol/L, P<0.05) although the difference was not apparent when controlling for multiple testing. There were no statistically significant effects of the three SNPs on insulin sensitivity or beta cell function. However accumulated risk allele showed a lower beta cell function on increasing plasma fatty acids with a carbon chain greater than six. Conclusions/Significance Differences in body composition and lipids associated with common SNPs in the FFAR1 gene were apparently not mediated by changes in insulin sensitivity or beta-cell function.
Resumo:
Purpose – The purpose of this paper is to consider prospects for UK REITs, which were introduced on 1 January 2007. It specifically focuses on the potential influence of depreciation and expenditure on income and distributions. Design/methodology/approach – First, the ways in which depreciation can affect vehicle earnings and value are discussed. This is then set in the context of the specific rules and features of REITs. An analysis using property income and expenditure data from the Investment Property Databank (IPD) then assesses what gross and net income for a UK REIT might have been like for the period 1984-2003. Findings – A UK REIT must distribute at least 90 per cent of net income from its property rental business. Expenditure therefore plays a significant part in determining what funds remain for distribution. Over 1984-2003, expenditure has absorbed 20 per cent of gross income and been a source of earnings volatility, which would have been exacerbated by gearing. Practical implications – Expenditure must take place to help UK REITs maintain and renew their real estate portfolios. In view of this, investors should moderate expectations of a high and stable income return, although it may well still be so relative to alternative investments. Originality/value – Previous literature on depreciation has not quantified amounts spent on portfolios to keep depreciation at those rates. Nor, to our knowledge, has its ideas been placed in the indirect investor context.