862 resultados para horizons of expectation
Resumo:
This paper considers the problem of estimation when one of a number of populations, assumed normal with known common variance, is selected on the basis of it having the largest observed mean. Conditional on selection of the population, the observed mean is a biased estimate of the true mean. This problem arises in the analysis of clinical trials in which selection is made between a number of experimental treatments that are compared with each other either with or without an additional control treatment. Attempts to obtain approximately unbiased estimates in this setting have been proposed by Shen [2001. An improved method of evaluating drug effect in a multiple dose clinical trial. Statist. Medicine 20, 1913–1929] and Stallard and Todd [2005. Point estimates and confidence regions for sequential trials involving selection. J. Statist. Plann. Inference 135, 402–419]. This paper explores the problem in the simple setting in which two experimental treatments are compared in a single analysis. It is shown that in this case the estimate of Stallard and Todd is the maximum-likelihood estimate (m.l.e.), and this is compared with the estimate proposed by Shen. In particular, it is shown that the m.l.e. has infinite expectation whatever the true value of the mean being estimated. We show that there is no conditionally unbiased estimator, and propose a new family of approximately conditionally unbiased estimators, comparing these with the estimators suggested by Shen.
Resumo:
Investigations were conducted during the 2003, 2004 and 2005 growing seasons in northern Greece to evaluate effects of tillage regime (mouldboard plough, chisel plough and rotary tiller), cropping sequence (continuous cotton, cotton-sugar beet rotation and continuous tobacco) and herbicide treatment on weed seedbank dynamics. Amaranthus spp. and Portulaca oleracea were the most abundant species, ranging from 76% to 89% of total weed seeds found in 0-15 and 15-30 cm soil depths during the 3 years. With the mouldboard plough, 48% and 52% of the weed seedbank was found in the 0-15 and 15-30 cm soil horizons, while approximately 60% was concentrated in the upper 15 cm soil horizon for chisel plough and rotary tillage. Mouldboard ploughing significantly buried more Echinochloa crus-galli seeds in the 15-30 cm soil horizon compared with the other tillage regimes. Total seedbank (0-30 cm) of P. oleracea was significantly reduced in cotton-sugar beet rotation compared with cotton and tobacco monocultures, while the opposite occurred for E. crus-galli. Total seed densities of most annual broad-leaved weed species (Amaranthus spp., P. oleracea, Solanum nigrum) and E. crus-galli were lower in herbicide treated than in untreated plots. The results suggest that in light textured soils, conventional tillage with herbicide use gradually reduces seed density of small seeded weed species in the top 15 cm over several years. In contrast, crop rotation with the early established sugar beet favours spring-germinating grass weed species, but also prevents establishment of summer-germinating weed species by the early developing crop canopy.
Resumo:
Microsatellite lengths change over evolutionary time through a process of replication slippage. A recently proposed model of this process holds that the expansionary tendencies of slippage mutation are balanced by point mutations breaking longer microsatellites into smaller units and that this process gives rise to the observed frequency distributions of uninterrupted microsatellite lengths. We refer to this as the slippage/point-mutation theory. Here we derive the theory's predictions for interrupted microsatellites comprising regions of perfect repeats, labeled segments, separated by dinucleotide interruptions containing point mutations. These predictions are tested by reference to the frequency distributions of segments of AC microsatellite in the human genome, and several predictions are shown not to be supported by the data, as follows. The estimated slippage rates are relatively low for the first four repeats, and then rise initially linearly with length, in accordance with previous work. However, contrary to expectation and the experimental evidence, the inferred slippage rates decline in segments above 10 repeats. Point mutation rates are also found to be higher within microsatellites than elsewhere. The theory provides an excellent fit to the frequency distribution of peripheral segment lengths but fails to explain why internal segments are shorter. Furthermore, there are fewer microsatellites with many segments than predicted. The frequencies of interrupted microsatellites decline geometrically with microsatellite size measured in number of segments, so that for each additional segment, the number of microsatellites is 33.6% less. Overall we conclude that the detailed structure of interrupted microsatellites cannot be reconciled with the existing slippage/point-mutation theory of microsatellite evolution, and we suggest that microsatellites are stabilized by processes acting on interior rather than on peripheral segments.
Resumo:
The Engineering and Physical Sciences Research Council (EPSRC) extending quality of life (EQUAL) initiative, specifically supports interdisciplinary user-focused design, engineering and technology research concerned with enhancing the independence and quality of life of older and disabled people. This paper briefly describes a recent programme to encourage the adoption of a broader perspective on the lives and needs of older people that have been pursued by EPSRC through its extending quality life (EQUAL) initiative. EPSRC is the principle supporter design, engineering and technology research in UK universities. The paper illustrates the scope of EQUAL.
Resumo:
Over-involved parenting is commonly hypothesized to be it risk factor for the development of anxiety disorders in childhood. This parenting style may result from parental attempts to prevent child distress based on expectations that the child will be unable to cope in a challenging situation. Naturalistic studies are limited in their ability to disentangle the overlapping contribution of child and parent factors in driving parental behaviours. To overcome this difficulty, an experimental study was conducted in which parental expectations of child distress were manipulated and the effects on parent behaviour and child mood were assessed. Fifty-two children (aged 7 - 11 years) and their primary caregiver participated. Parents were allocated to either a "positive" or a "negative" expectation group. Observations were made of the children and their parents interacting whilst completing a difficult anagram task. Parents given negative expectations of their child's response displayed higher levels of involvement. No differences were found on indices of child mood and behaviour and possible explanations for this are considered. The findings are consistent with suggestions that increased parental involvement may be a "natural" reaction to enhanced perceptions of child vulnerability and an attempt to avoid child distress.
Resumo:
The main activity carried out by the geophysicist when interpreting seismic data, in terms of both importance and time spent is tracking (or picking) seismic events. in practice, this activity turns out to be rather challenging, particularly when the targeted event is interrupted by discontinuities such as geological faults or exhibits lateral changes in seismic character. In recent years, several automated schemes, known as auto-trackers, have been developed to assist the interpreter in this tedious and time-consuming task. The automatic tracking tool available in modem interpretation software packages often employs artificial neural networks (ANN's) to identify seismic picks belonging to target events through a pattern recognition process. The ability of ANNs to track horizons across discontinuities largely depends on how reliably data patterns characterise these horizons. While seismic attributes are commonly used to characterise amplitude peaks forming a seismic horizon, some researchers in the field claim that inherent seismic information is lost in the attribute extraction process and advocate instead the use of raw data (amplitude samples). This paper investigates the performance of ANNs using either characterisation methods, and demonstrates how the complementarity of both seismic attributes and raw data can be exploited in conjunction with other geological information in a fuzzy inference system (FIS) to achieve an enhanced auto-tracking performance.
Resumo:
Information on the distribution and behavior of C fractions in soil particle sizes is crucial for understanding C dynamics in soil. At present little is known about the behavior of the C associated with silt-size particles. We quantified the concentrations, distribution, and enrichment of total C (TC), readily oxidizable C (ROC), hotwater- extractable C (HWC), and cold-water-extractable C (CWC) fractions in coarse (63–20-mm), medium (20–6.3-mm), and fine (6.3–2-mm) silt-size subfractions and in coarse (2000–250 mm) and fine (250–63 mm) sand and clay (<2-mm) soil fractions isolated from bulk soil (<2 mm), and 2- to 4-mm aggregate-size fraction of surface (0–25 cm) and subsurface (25–55 cm) soils under different land uses. All measured C fractions varied significantly across all soil particle-size fractions. The highest C concentrations were associated with the <20-mm soil fractions and peaked in the medium (20–6.3-mm) and fine (6.3–2-mm) silt subfractions in most treatments. Carbon enrichment ratios (ERC) revealed the dual behavior of the C fractions associated with the medium silt-size fraction, demonstrating the simultaneous enrichment of TC and ROC, and the depletion of HWC and CWC fractions. The medium silt (20–6.3-mm) subfraction was identified in this study as a zone where the associated C fractions exhibit transitory qualities. Our results show that investigating subfractions within the silt-size particle fraction provides better understanding of the behavior of C fractions in this soil fraction.
Resumo:
Although the use of climate scenarios for impact assessment has grown steadily since the 1990s, uptake of such information for adaptation is lagging by nearly a decade in terms of scientific output. Nonetheless, integration of climate risk information in development planning is now a priority for donor agencies because of the need to prepare for climate change impacts across different sectors and countries. This urgency stems from concerns that progress made against Millennium Development Goals (MDGs) could be threatened by anthropogenic climate change beyond 2015. Up to this time the human signal, though detectable and growing, will be a relatively small component of climate variability and change. This implies the need for a twin-track approach: on the one hand, vulnerability assessments of social and economic strategies for coping with present climate extremes and variability, and, on the other hand, development of climate forecast tools and scenarios to evaluate sector-specific, incremental changes in risk over the next few decades. This review starts by describing the climate outlook for the next couple of decades and the implications for adaptation assessments. We then review ways in which climate risk information is already being used in adaptation assessments and evaluate the strengths and weaknesses of three groups of techniques. Next we identify knowledge gaps and opportunities for improving the production and uptake of climate risk information for the 2020s. We assert that climate change scenarios can meet some, but not all, of the needs of adaptation planning. Even then, the choice of scenario technique must be matched to the intended application, taking into account local constraints of time, resources, human capacity and supporting infrastructure. We also show that much greater attention should be given to improving and critiquing models used for climate impact assessment, as standard practice. Finally, we highlight the over-arching need for the scientific community to provide more information and guidance on adapting to the risks of climate variability and change over nearer time horizons (i.e. the 2020s). Although the focus of the review is on information provision and uptake in developing regions, it is clear that many developed countries are facing the same challenges. Copyright © 2009 Royal Meteorological Society
Resumo:
Decision theory is the study of models of judgement involved in, and leading to, deliberate and (usually) rational choice. In real estate investment there are normative models for the allocation of assets. These asset allocation models suggest an optimum allocation between the respective asset classes based on the investors’ judgements of performance and risk. Real estate is selected, as other assets, on the basis of some criteria, e.g. commonly its marginal contribution to the production of a mean variance efficient multi asset portfolio, subject to the investor’s objectives and capital rationing constraints. However, decisions are made relative to current expectations and current business constraints. Whilst a decision maker may believe in the required optimum exposure levels as dictated by an asset allocation model, the final decision may/will be influenced by factors outside the parameters of the mathematical model. This paper discusses investors' perceptions and attitudes toward real estate and highlights the important difference between theoretical exposure levels and pragmatic business considerations. It develops a model to identify “soft” parameters in decision making which will influence the optimal allocation for that asset class. This “soft” information may relate to behavioural issues such as the tendency to mirror competitors; a desire to meet weight of money objectives; a desire to retain the status quo and many other non-financial considerations. The paper aims to establish the place of property in multi asset portfolios in the UK and examine the asset allocation process in practice, with a view to understanding the decision making process and to look at investors’ perceptions based on an historic analysis of market expectation; a comparison with historic data and an analysis of actual performance.
Resumo:
This study considers the consistency of the role of both the private and public real estate markets within a mixed-asset context. While a vast literature has developed that has examined the potential role of both the private and public real estate markets, most studies have largely relied on both single time horizons and single sample periods. This paper builds upon the analysis of Lee and Stevenson (2005) who examined the consistency of REITs in a US capital market portfolio. The current paper extends that by also analyzing the role of the private market. To address the question, the allocation of both the private and traded markets is evaluated over different holding periods varying from 5- to 20-years. In general the results show that optimum mixed-asset portfolios already containing private real estate have little place for public real estate securities, especially in low risk portfolios and for longer investment horizons. Additionally, mixed-asset portfolios with public real estate either see the allocations to REITs diminished or eliminated if private real estate is also considered. The results demonstrate that there is a still a strong case for private real estate in the mixed-asset portfolio on the basis of an increase in risk-adjusted performance, even if the investor is already holding REITs, but that the reverse is not always the case.
Resumo:
The performance of various statistical models and commonly used financial indicators for forecasting securitised real estate returns are examined for five European countries: the UK, Belgium, the Netherlands, France and Italy. Within a VAR framework, it is demonstrated that the gilt-equity yield ratio is in most cases a better predictor of securitized returns than the term structure or the dividend yield. In particular, investors should consider in their real estate return models the predictability of the gilt-equity yield ratio in Belgium, the Netherlands and France, and the term structure of interest rates in France. Predictions obtained from the VAR and univariate time-series models are compared with the predictions of an artificial neural network model. It is found that, whilst no single model is universally superior across all series, accuracy measures and horizons considered, the neural network model is generally able to offer the most accurate predictions for 1-month horizons. For quarterly and half-yearly forecasts, the random walk with a drift is the most successful for the UK, Belgian and Dutch returns and the neural network for French and Italian returns. Although this study underscores market context and forecast horizon as parameters relevant to the choice of the forecast model, it strongly indicates that analysts should exploit the potential of neural networks and assess more fully their forecast performance against more traditional models.