895 resultados para data types and operators
Resumo:
Svalgaard and Cliver (2010) recently reported a consensus between the various reconstructions of the heliospheric field over recent centuries. This is a significant development because, individually, each has uncertainties introduced by instrument calibration drifts, limited numbers of observatories, and the strength of the correlations employed. However, taken collectively, a consistent picture is emerging. We here show that this consensus extends to more data sets and methods than reported by Svalgaard and Cliver, including that used by Lockwood et al. (1999), when their algorithm is used to predict the heliospheric field rather than the open solar flux. One area where there is still some debate relates to the existence and meaning of a floor value to the heliospheric field. From cosmogenic isotope abundances, Steinhilber et al. (2010) have recently deduced that the near-Earth IMF at the end of the Maunder minimum was 1.80 ± 0.59 nT which is considerably lower than the revised floor of 4nT proposed by Svalgaard and Cliver. We here combine cosmogenic and geomagnetic reconstructions and modern observations (with allowance for the effect of solar wind speed and structure on the near-Earth data) to derive an estimate for the open solar flux of (0.48 ± 0.29) × 1014 Wb at the end of the Maunder minimum. By way of comparison, the largest and smallest annual means recorded by instruments in space between 1965 and 2010 are 5.75 × 1014 Wb and 1.37 × 1014 Wb, respectively, set in 1982 and 2009, and the maximum of the 11 year running means was 4.38 × 1014 Wb in 1986. Hence the average open solar flux during the Maunder minimum is found to have been 11% of its peak value during the recent grand solar maximum.
Resumo:
Virulence for bean and soybean is determined by effector genes in a plasmid-borne pathogenicity island (PAI) in race 7 strain 1449B of Pseudomonas syringae pv. phaseolicola. One of the effector genes, avrPphF, confers either pathogenicity, virulence, or avirulence depending on the plant host and is absent from races 2, 3, 4, 6, and 8 of this pathogen. Analysis of cosmid clones and comparison of DNA sequences showed that the absence of avrPphF from strain 1448A is due to deletion of a continuous 9.5-kb fragment. The remainder of the PAI is well conserved in strains 1448A and 1449B. The left junction of the deleted region consists of a chimeric transposable element generated from the fusion of homologs of IS1492 from Pseudomonas putida and IS1090 from Ralstonia eutropha. The borders of the deletion were conserved in 66 P. syringae pv. phaseolicola strains isolated in different countries and representing the five races lacking avrPphF. However, six strains isolated in Spain had a 10.5-kb deletion that extended 1 kb further from the right junction. The perfect conservation of the 28-nucleotide right repeat of the IS1090 homolog in the two deletion types and in the other 47 insertions of the IS1090 homolog in the 1448A genome strongly suggests that the avrPphF deletions were mediated by the activity of the chimeric mobile element. Our data strongly support a clonal origin for the races of P. syringae pv. phaseolicola lacking avrPphF.
Resumo:
This paper re-examines the relative importance of sector and regional effects in determining property returns. Using the largest property database currently available in the world, we decompose the returns on individual properties into a national effect, common to all properties, and a number of sector and regional factors. However, unlike previous studies, we categorise the individual property data into an ever-increasing number of property-types and regions, from a simple 3-by-3 classification, up to a 10 by 63 sector/region classification. In this way we can test the impact that a finer classification has on the sector and regional effects. We confirm the earlier findings of previous studies that sector-specific effects have a greater influence on property returns than regional effects. We also find that the impact of the sector effect is robust across different classifications of sectors and regions. Nonetheless, the more refined sector and regional partitions uncover some interesting sector and regional differences, which were obscured in previous studies. All of which has important implications for property portfolio construction and analysis.
Resumo:
As part of a large European coastal operational oceanography project (ECOOP), we have developed a web portal for the display and comparison of model and in situ marine data. The distributed model and in situ datasets are accessed via an Open Geospatial Consortium Web Map Service (WMS) and Web Feature Service (WFS) respectively. These services were developed independently and readily integrated for the purposes of the ECOOP project, illustrating the ease of interoperability resulting from adherence to international standards. The key feature of the portal is the ability to display co-plotted timeseries of the in situ and model data and the quantification of misfits between the two. By using standards-based web technology we allow the user to quickly and easily explore over twenty model data feeds and compare these with dozens of in situ data feeds without being concerned with the low level details of differing file formats or the physical location of the data. Scientific and operational benefits to this work include model validation, quality control of observations, data assimilation and decision support in near real time. In these areas it is essential to be able to bring different data streams together from often disparate locations.
Resumo:
A dynamic, mechanistic model of enteric fermentation was used to investigate the effect of type and quality of grass forage, dry matter intake (DMI) and proportion of concentrates in dietary dry matter (DM) on variation in methane (CH(4)) emission from enteric fermentation in dairy cows. The model represents substrate degradation and microbial fermentation processes in rumen and hindgut and, in particular, the effects of type of substrate fermented and of pH oil the production of individual volatile fatty acids and CH, as end-products of fermentation. Effects of type and quality of fresh and ensiled grass were evaluated by distinguishing two N fertilization rates of grassland and two stages of grass maturity. Simulation results indicated a strong impact of the amount and type of grass consumed oil CH(4) emission, with a maximum difference (across all forage types and all levels of DM 1) of 49 and 77% in g CH(4)/kg fat and protein corrected milk (FCM) for diets with a proportion of concentrates in dietary DM of 0.1 and 0.4, respectively (values ranging from 10.2 to 19.5 g CH(4)/kg FCM). The lowest emission was established for early Cut, high fertilized grass silage (GS) and high fertilized grass herbage (GH). The highest emission was found for late cut, low-fertilized GS. The N fertilization rate had the largest impact, followed by stage of grass maturity at harvesting and by the distinction between GH and GS. Emission expressed in g CH(4)/kg FCM declined oil average 14% with an increase of DMI from 14 to 18 kg/day for grass forage diets with a proportion of concentrates of 0.1, and on average 29% with an increase of DMI from 14 to 23 kg/day for diets with a proportion of concentrates of 0.4. Simulation results indicated that a high proportion of concentrates in dietary DM may lead to a further reduction of CH, emission per kg FCM mainly as a result of a higher DM I and milk yield, in comparison to low concentrate diets. Simulation results were evaluated against independent data obtained at three different laboratories in indirect calorimetry trials with COWS consuming GH mainly. The model predicted the average of observed values reasonably, but systematic deviations remained between individual laboratories and root mean squared prediction error was a proportion of 0.12 of the observed mean. Both observed and predicted emission expressed in g CH(4)/kg DM intake decreased upon an increase in dietary N:organic matter (OM) ratio. The model reproduced reasonably well the variation in measured CH, emission in cattle sheds oil Dutch dairy farms and indicated that oil average a fraction of 0.28 of the total emissions must have originated from manure under these circumstances.
Resumo:
We describe a model-data fusion (MDF) inter-comparison project (REFLEX), which compared various algorithms for estimating carbon (C) model parameters consistent with both measured carbon fluxes and states and a simple C model. Participants were provided with the model and with both synthetic net ecosystem exchange (NEE) of CO2 and leaf area index (LAI) data, generated from the model with added noise, and observed NEE and LAI data from two eddy covariance sites. Participants endeavoured to estimate model parameters and states consistent with the model for all cases over the two years for which data were provided, and generate predictions for one additional year without observations. Nine participants contributed results using Metropolis algorithms, Kalman filters and a genetic algorithm. For the synthetic data case, parameter estimates compared well with the true values. The results of the analyses indicated that parameters linked directly to gross primary production (GPP) and ecosystem respiration, such as those related to foliage allocation and turnover, or temperature sensitivity of heterotrophic respiration, were best constrained and characterised. Poorly estimated parameters were those related to the allocation to and turnover of fine root/wood pools. Estimates of confidence intervals varied among algorithms, but several algorithms successfully located the true values of annual fluxes from synthetic experiments within relatively narrow 90% confidence intervals, achieving >80% success rate and mean NEE confidence intervals <110 gC m−2 year−1 for the synthetic case. Annual C flux estimates generated by participants generally agreed with gap-filling approaches using half-hourly data. The estimation of ecosystem respiration and GPP through MDF agreed well with outputs from partitioning studies using half-hourly data. Confidence limits on annual NEE increased by an average of 88% in the prediction year compared to the previous year, when data were available. Confidence intervals on annual NEE increased by 30% when observed data were used instead of synthetic data, reflecting and quantifying the addition of model error. Finally, our analyses indicated that incorporating additional constraints, using data on C pools (wood, soil and fine roots) would help to reduce uncertainties for model parameters poorly served by eddy covariance data.
Resumo:
Data assimilation refers to the problem of finding trajectories of a prescribed dynamical model in such a way that the output of the model (usually some function of the model states) follows a given time series of observations. Typically though, these two requirements cannot both be met at the same time–tracking the observations is not possible without the trajectory deviating from the proposed model equations, while adherence to the model requires deviations from the observations. Thus, data assimilation faces a trade-off. In this contribution, the sensitivity of the data assimilation with respect to perturbations in the observations is identified as the parameter which controls the trade-off. A relation between the sensitivity and the out-of-sample error is established, which allows the latter to be calculated under operational conditions. A minimum out-of-sample error is proposed as a criterion to set an appropriate sensitivity and to settle the discussed trade-off. Two approaches to data assimilation are considered, namely variational data assimilation and Newtonian nudging, also known as synchronization. Numerical examples demonstrate the feasibility of the approach.
Resumo:
The P-found protein folding and unfolding simulation repository is designed to allow scientists to perform data mining and other analyses across large, distributed simulation data sets. There are two storage components in P-found: a primary repository of simulation data that is used to populate the second component, and a data warehouse that contains important molecular properties. These properties may be used for data mining studies. Here we demonstrate how grid technologies can support multiple, distributed P-found installations. In particular, we look at two aspects: firstly, how grid data management technologies can be used to access the distributed data warehouses; and secondly, how the grid can be used to transfer analysis programs to the primary repositories — this is an important and challenging aspect of P-found, due to the large data volumes involved and the desire of scientists to maintain control of their own data. The grid technologies we are developing with the P-found system will allow new large data sets of protein folding simulations to be accessed and analysed in novel ways, with significant potential for enabling scientific discovery.
Resumo:
The majority of vegetation reconstructions from the Neotropics are derived from fossil pollen records extracted from lake sediments. However, the interpretation of these records is restricted by limited knowledge of the contemporary relationships between the vegetation and pollen rain of Neotropical ecosystems, especially for more open vegetation such as savannas. This research aims to improve the interpretation of these records by investigating the vegetation and modern pollen rain of different savanna ecosystems in Bolivia using vegetation inventories, artificial pollen traps and surface lake sediments. Two types of savanna were studied, upland savannas (cerrado), occurring on well drained soils, and seasonally-inundated savannas occurring on seasonally water-logged soils. Quantitative vegetation data are used to identify taxa that are floristically important in the different savanna types and to allow modern pollen/vegetation ratios to be calculated. Artificial pollen traps from the upland savanna site are dominated by Moraceae (35%), Poaceae (30%), Alchornea (6%) and Cecropia (4%). The two seasonally-inundated savanna sites are dominated by Moraceae (37%), Poaceae (20%), Alchornea (8%) and Cecropia (7%), and Moraceae (25%), Cyperaceae (22%), Poaceae (19%) and Cecropia (9%), respectively. The modern pollen rain of seasonally-inundated savannas from surface lake sediments is dominated by Cyperaceae (35%), Poaceae (33%), Moraceae (9%) and Asteraceae (5%). Upland and seasonally-flooded savannas were found to be only subtly distinct from each other palynologically. All sites have a high proportion of Moraceae pollen due to effective wind dispersal of this pollen type from areas of evergreen forest close to the study sites. Modern pollen/vegetation ratios show that many key woody plant taxa are absent/under-represented in the modern pollen rain (e.g., Caryocar and Tabebuia). The lower-than-expected percentages of Poaceae pollen, and the scarcity of savanna indicators, in the modern pollen rain of these ecosystems mean that savannas could potentially be overlooked in fossil pollen records without consideration of the full pollen spectrum available.
Resumo:
In this article, we review the state-of-the-art techniques in mining data streams for mobile and ubiquitous environments. We start the review with a concise background of data stream processing, presenting the building blocks for mining data streams. In a wide range of applications, data streams are required to be processed on small ubiquitous devices like smartphones and sensor devices. Mobile and ubiquitous data mining target these applications with tailored techniques and approaches addressing scarcity of resources and mobility issues. Two categories can be identified for mobile and ubiquitous mining of streaming data: single-node and distributed. This survey will cover both categories. Mining mobile and ubiquitous data require algorithms with the ability to monitor and adapt the working conditions to the available computational resources. We identify the key characteristics of these algorithms and present illustrative applications. Distributed data stream mining in the mobile environment is then discussed, presenting the Pocket Data Mining framework. Mobility of users stimulates the adoption of context-awareness in this area of research. Context-awareness and collaboration are discussed in the Collaborative Data Stream Mining, where agents share knowledge to learn adaptive accurate models.
Resumo:
The purpose of this study was to develop an understanding of the current state of scientific data sharing that stakeholders could use to develop and implement effective data sharing strategies and policies. The study developed a conceptual model to describe the process of data sharing, and the drivers, barriers, and enablers that determine stakeholder engagement. The conceptual model was used as a framework to structure discussions and interviews with key members of all stakeholder groups. Analysis of data obtained from interviewees identified a number of themes that highlight key requirements for the development of a mature data sharing culture.
Resumo:
Data assimilation (DA) systems are evolving to meet the demands of convection-permitting models in the field of weather forecasting. On 19 April 2013 a special interest group meeting of the Royal Meteorological Society brought together UK researchers looking at different aspects of the data assimilation problem at high resolution, from theory to applications, and researchers creating our future high resolution observational networks. The meeting was chaired by Dr Sarah Dance of the University of Reading and Dr Cristina Charlton-Perez from the MetOffice@Reading. The purpose of the meeting was to help define the current state of high resolution data assimilation in the UK. The workshop assembled three main types of scientists: observational network specialists, operational numerical weather prediction researchers and those developing the fundamental mathematical theory behind data assimilation and the underlying models. These three working areas are intrinsically linked; therefore, a holistic view must be taken when discussing the potential to make advances in high resolution data assimilation.
Resumo:
Population modelling is increasingly recognised as a useful tool for pesticide risk assessment. For vertebrates that may ingest pesticides with their food, such as woodpigeon (Columba palumbus), population models that simulate foraging behaviour explicitly can help predicting both exposure and population-level impact. Optimal foraging theory is often assumed to explain the individual-level decisions driving distributions of individuals in the field, but it may not adequately predict spatial and temporal characteristics of woodpigeon foraging because of the woodpigeons’ excellent memory, ability to fly long distances, and distinctive flocking behaviour. Here we present an individual-based model (IBM) of the woodpigeon. We used the model to predict distributions of foraging woodpigeons that use one of six alternative foraging strategies: optimal foraging, memory-based foraging and random foraging, each with or without flocking mechanisms. We used pattern-oriented modelling to determine which of the foraging strategies is best able to reproduce observed data patterns. Data used for model evaluation were gathered during a long-term woodpigeon study conducted between 1961 and 2004 and a radiotracking study conducted in 2003 and 2004, both in the UK, and are summarised here as three complex patterns: the distributions of foraging birds between vegetation types during the year, the number of fields visited daily by individuals, and the proportion of fields revisited by them on subsequent days. The model with a memory-based foraging strategy and a flocking mechanism was the only one to reproduce these three data patterns, and the optimal foraging model produced poor matches to all of them. The random foraging strategy reproduced two of the three patterns but was not able to guarantee population persistence. We conclude that with the memory-based foraging strategy including a flocking mechanism our model is realistic enough to estimate the potential exposure of woodpigeons to pesticides. We discuss how exposure can be linked to our model, and how the model could be used for risk assessment of pesticides, for example predicting exposure and effects in heterogeneous landscapes planted seasonally with a variety of crops, while accounting for differences in land use between landscapes.
Resumo:
Using five climate model simulations of the response to an abrupt quadrupling of CO2, the authors perform the first simultaneous model intercomparison of cloud feedbacks and rapid radiative adjustments with cloud masking effects removed, partitioned among changes in cloud types and gross cloud properties. Upon CO2 quadrupling, clouds exhibit a rapid reduction in fractional coverage, cloud-top pressure, and optical depth, with each contributing equally to a 1.1 W m−2 net cloud radiative adjustment, primarily from shortwave radiation. Rapid reductions in midlevel clouds and optically thick clouds are important in reducing planetary albedo in every model. As the planet warms, clouds become fewer, higher, and thicker, and global mean net cloud feedback is positive in all but one model and results primarily from increased trapping of longwave radiation. As was true for earlier models, high cloud changes are the largest contributor to intermodel spread in longwave and shortwave cloud feedbacks, but low cloud changes are the largest contributor to the mean and spread in net cloud feedback. The importance of the negative optical depth feedback relative to the amount feedback at high latitudes is even more marked than in earlier models. The authors show that the negative longwave cloud adjustment inferred in previous studies is primarily caused by a 1.3 W m−2 cloud masking of CO2 forcing. Properly accounting for cloud masking increases net cloud feedback by 0.3 W m−2 K−1, whereas accounting for rapid adjustments reduces by 0.14 W m−2 K−1 the ensemble mean net cloud feedback through a combination of smaller positive cloud amount and altitude feedbacks and larger negative optical depth feedbacks.