40 resultados para Web Log Data


Relevância:

30.00% 30.00%

Publicador:

Resumo:

Various studies investigating the future impacts of integrating high levels of renewable energy make use of historical meteorological (met) station data to produce estimates of future generation. Hourly means of 10m horizontal wind are extrapolated to a standard turbine hub height using the wind profile power or log law and used to simulate the hypothetical power output of a turbine at that location; repeating this procedure using many viable locations can produce a picture of future electricity generation. However, the estimate of hub height wind speed is dependent on the choice of the wind shear exponent a or the roughness length z0, and requires a number of simplifying assumptions. This paper investigates the sensitivity of this estimation on generation output using a case study of a met station in West Freugh, Scotland. The results show that the choice of wind shear exponent is a particularly sensitive parameter which can lead to significant variation of estimated hub height wind speed and hence estimated future generation potential of a region.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

OBJECTIVES: The prediction of protein structure and the precise understanding of protein folding and unfolding processes remains one of the greatest challenges in structural biology and bioinformatics. Computer simulations based on molecular dynamics (MD) are at the forefront of the effort to gain a deeper understanding of these complex processes. Currently, these MD simulations are usually on the order of tens of nanoseconds, generate a large amount of conformational data and are computationally expensive. More and more groups run such simulations and generate a myriad of data, which raises new challenges in managing and analyzing these data. Because the vast range of proteins researchers want to study and simulate, the computational effort needed to generate data, the large data volumes involved, and the different types of analyses scientists need to perform, it is desirable to provide a public repository allowing researchers to pool and share protein unfolding data. METHODS: To adequately organize, manage, and analyze the data generated by unfolding simulation studies, we designed a data warehouse system that is embedded in a grid environment to facilitate the seamless sharing of available computer resources and thus enable many groups to share complex molecular dynamics simulations on a more regular basis. RESULTS: To gain insight into the conformational fluctuations and stability of the monomeric forms of the amyloidogenic protein transthyretin (TTR), molecular dynamics unfolding simulations of the monomer of human TTR have been conducted. Trajectory data and meta-data of the wild-type (WT) protein and the highly amyloidogenic variant L55P-TTR represent the test case for the data warehouse. CONCLUSIONS: Web and grid services, especially pre-defined data mining services that can run on or 'near' the data repository of the data warehouse, are likely to play a pivotal role in the analysis of molecular dynamics unfolding data.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Background: Expression microarrays are increasingly used to obtain large scale transcriptomic information on a wide range of biological samples. Nevertheless, there is still much debate on the best ways to process data, to design experiments and analyse the output. Furthermore, many of the more sophisticated mathematical approaches to data analysis in the literature remain inaccessible to much of the biological research community. In this study we examine ways of extracting and analysing a large data set obtained using the Agilent long oligonucleotide transcriptomics platform, applied to a set of human macrophage and dendritic cell samples. Results: We describe and validate a series of data extraction, transformation and normalisation steps which are implemented via a new R function. Analysis of replicate normalised reference data demonstrate that intrarray variability is small (only around 2 of the mean log signal), while interarray variability from replicate array measurements has a standard deviation (SD) of around 0.5 log(2) units (6 of mean). The common practise of working with ratios of Cy5/Cy3 signal offers little further improvement in terms of reducing error. Comparison to expression data obtained using Arabidopsis samples demonstrates that the large number of genes in each sample showing a low level of transcription reflect the real complexity of the cellular transcriptome. Multidimensional scaling is used to show that the processed data identifies an underlying structure which reflect some of the key biological variables which define the data set. This structure is robust, allowing reliable comparison of samples collected over a number of years and collected by a variety of operators. Conclusions: This study outlines a robust and easily implemented pipeline for extracting, transforming normalising and visualising transcriptomic array data from Agilent expression platform. The analysis is used to obtain quantitative estimates of the SD arising from experimental (non biological) intra- and interarray variability, and for a lower threshold for determining whether an individual gene is expressed. The study provides a reliable basis for further more extensive studies of the systems biology of eukaryotic cells.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The objective of this paper is to apply the mis-specification (M-S) encompassing perspective to the problem of choosing between linear and log-linear unit-root models. A simple M-S encompassing test, based on an auxiliary regression stemming from the conditional second moment, is proposed and its empirical size and power are investigated using Monte Carlo simulations. It is shown that by focusing on the conditional process the sampling distributions of the relevant statistics are well behaved under both the null and alternative hypotheses. The proposed M-S encompassing test is illustrated using US total disposable income quarterly data.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Improving lifestyle behaviours has considerable potential for reducing the global burden of non-communicable diseases, promoting better health across the life-course and increasing well-being. However, realising this potential will require the development, testing and implementation of much more effective behaviour change interventions than are used conventionally. Therefore, the aim of this study was to conduct a multi-centre, web-based, proof-of-principle study of personalised nutrition (PN) to determine whether providing more personalised dietary advice leads to greater improvements in eating patterns and health outcomes compared to conventional population-based advice. A total of 5,562 volunteers were screened across seven European countries; the first 1,607 participants who fulfilled the inclusion criteria were recruited into the trial. Participants were randomly assigned to one of the following intervention groups for a 6-month period: Level 0-control group-receiving conventional, non-PN advice; Level 1-receiving PN advice based on dietary intake data alone; Level 2-receiving PN advice based on dietary intake and phenotypic data; and Level 3-receiving PN advice based on dietary intake, phenotypic and genotypic data. A total of 1,607 participants had a mean age of 39.8 years (ranging from 18 to 79 years). Of these participants, 60.9 % were women and 96.7 % were from white-European background. The mean BMI for all randomised participants was 25.5 kg m(-2), and 44.8 % of the participants had a BMI ≥ 25.0 kg m(-2). Food4Me is the first large multi-centre RCT of web-based PN. The main outcomes from the Food4Me study will be submitted for publication during 2015.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Income growth in highly industrialised countries has resulted in consumer choice of foodstuffs no longer being primarily influenced by basic factors such as price and organoleptic features. From this perspective, the present study sets out to evaluate how and to what extent consumer choice is influenced by the possible negative effects on health and environment caused by the consumption of fruit containing deposits of pesticides and chemical products. The study describes the results of a survey which explores and estimates consumer willingness to pay in two forms: a yearly contribution for the abolition of the use of pesticides on fruit, and a premium price for organically grown apples guaranteed by a certified label. The same questionnaire was administered to two samples. The first was a conventional face-to-face survey of customers of large retail outlets located around Bologna (Italy); the second was an Internet sample. The discrete choice data were analysed by means of probit and tobit models to estimate the utility consumers attribute to organically grown fruit and to a pesticide ban. The research also addresses questions of validity and representativeness as a fundamental problem in web-based surveys.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Social network has gained remarkable attention in the last decade. Accessing social network sites such as Twitter, Facebook LinkedIn and Google+ through the internet and the web 2.0 technologies has become more affordable. People are becoming more interested in and relying on social network for information, news and opinion of other users on diverse subject matters. The heavy reliance on social network sites causes them to generate massive data characterised by three computational issues namely; size, noise and dynamism. These issues often make social network data very complex to analyse manually, resulting in the pertinent use of computational means of analysing them. Data mining provides a wide range of techniques for detecting useful knowledge from massive datasets like trends, patterns and rules [44]. Data mining techniques are used for information retrieval, statistical modelling and machine learning. These techniques employ data pre-processing, data analysis, and data interpretation processes in the course of data analysis. This survey discusses different data mining techniques used in mining diverse aspects of the social network over decades going from the historical techniques to the up-to-date models, including our novel technique named TRCM. All the techniques covered in this survey are listed in the Table.1 including the tools employed as well as names of their authors.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

For users of climate services, the ability to quickly determine the datasets that best fit one's needs would be invaluable. The volume, variety and complexity of climate data makes this judgment difficult. The ambition of CHARMe ("Characterization of metadata to enable high-quality climate services") is to give a wider interdisciplinary community access to a range of supporting information, such as journal articles, technical reports or feedback on previous applications of the data. The capture and discovery of this "commentary" information, often created by data users rather than data providers, and currently not linked to the data themselves, has not been significantly addressed previously. CHARMe applies the principles of Linked Data and open web standards to associate, record, search and publish user-derived annotations in a way that can be read both by users and automated systems. Tools have been developed within the CHARMe project that enable annotation capability for data delivery systems already in wide use for discovering climate data. In addition, the project has developed advanced tools for exploring data and commentary in innovative ways, including an interactive data explorer and comparator ("CHARMe Maps") and a tool for correlating climate time series with external "significant events" (e.g. instrument failures or large volcanic eruptions) that affect the data quality. Although the project focuses on climate science, the concepts are general and could be applied to other fields. All CHARMe system software is open-source, released under a liberal licence, permitting future projects to re-use the source code as they wish.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Background: The high prevalence of physical inactivity worldwide calls for innovative and more effective ways to promote physical activity (PA). There are limited objective data on the effectiveness of Web-based personalized feedback on increasing PA in adults. Objective: It is hypothesized that providing personalized advice based on PA measured objectively alongside diet, phenotype, or genotype information would lead to larger and more sustained changes in PA, compared with nonpersonalized advice. Methods: A total of 1607 adults in seven European countries were randomized to either a control group (nonpersonalized advice, Level 0, L0) or to one of three personalized groups receiving personalized advice via the Internet based on current PA plus diet (Level 1, L1), PA plus diet and phenotype (Level 2, L2), or PA plus diet, phenotype, and genotype (Level 3, L3). PA was measured for 6 months using triaxial accelerometers, and self-reported using the Baecke questionnaire. Outcomes were objective and self-reported PA after 3 and 6 months. Results: While 1270 participants (85.81% of 1480 actual starters) completed the 6-month trial, 1233 (83.31%) self-reported PA at both baseline and month 6, but only 730 (49.32%) had sufficient objective PA data at both time points. For the total cohort after 6 months, a greater improvement in self-reported total PA (P=.02) and PA during leisure (nonsport) (P=.03) was observed in personalized groups compared with the control group. For individuals advised to increase PA, we also observed greater improvements in those two self-reported indices (P=.006 and P=.008, respectively) with increased personalization of the advice (L2 and L3 vs L1). However, there were no significant differences in accelerometer results between personalized and control groups, and no significant effect of adding phenotypic or genotypic information to the tailored feedback at month 3 or 6. After 6 months, there were small but significant improvements in the objectively measured physical activity level (P<.05), moderate PA (P<.01), and sedentary time (P<.001) for individuals advised to increase PA, but these changes were similar across all groups. Conclusions: Different levels of personalization produced similar small changes in objective PA. We found no evidence that personalized advice is more effective than conventional “one size fits all” guidelines to promote changes in PA in our Web-based intervention when PA was measured objectively. Based on self-reports, PA increased to a greater extent with more personalized advice. Thus, it is crucial to measure PA objectively in any PA intervention study.