968 resultados para Short-text clustering


Relevância:

30.00% 30.00%

Publicador:

Resumo:

The Green Feed (GF) system (C-Lock Inc., Rapid City, USA) is used to estimate total daily methane emissions of individual cattle using short-term measurements obtained over several days. Our objective was to compare measurements of methane emission by growing cattle obtained using the GF system with measurements using respiration chambers (RC)or sulphur hexafluoride tracer (SF6). It was hypothesised that estimates of methane emission for individual animals and treatments would be similar for GF compared to RC or SF6 techniques. In experiment 1, maize or grass silage-based diets were fed to four growing Holstein heifers, whilst for experiment 2, four different heifers were fed four haylage treatments. Both experiments were a 4 × 4 Latin square design with 33 day periods. Green Feed measurements of methane emission were obtained over 7 days (days 22–28) and com-pared to subsequent RC measurements over 4 days (days 29–33). For experiment 3, 12growing heifers rotationally grazed three swards for 26 days, with simultaneous GF and SF6 measurements over two 4 day measurement periods (days 15–19 and days 22–26).Overall methane emissions (g/day and g/kg dry matter intake [DMI]) measured using GF in experiments 1 (198 and 26.6, respectively) and 2 (208 and 27.8, respectively) were similar to averages obtained using RC (218 and 28.3, respectively for experiment 1; and 209 and 27.7, respectively, for experiment 2); but there was poor concordance between the two methods (0.1043 for experiments 1 and 2 combined). Overall, methane emissions measured using SF6 were higher (P<0.001) than GF during grazing (186 vs. 164 g/day), but there was significant (P<0.01) concordance between the two methods (0.6017). There were fewer methane measurements by GF under grazing conditions in experiment 3 (1.60/day) com-pared to indoor measurements in experiments 1 (2.11/day) and 2 (2.34/day). Significant treatment effects on methane emission measured using RC and SF6 were not evident for GF measurements, and the ranking for treatments and individual animals differed using the GF system. We conclude that under our conditions of use the GF system was unable to detectsignificant treatment and individual animal differences in methane emissions that were identified using both RC and SF6techniques, in part due to limited numbers and timing ofmeasurements obtained. Our data suggest that successful use of the GF system is reliant on the number and timing of measurements obtained relative to diurnal patterns of methane emission.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Background: In many experimental pipelines, clustering of multidimensional biological datasets is used to detect hidden structures in unlabelled input data. Taverna is a popular workflow management system that is used to design and execute scientific workflows and aid in silico experimentation. The availability of fast unsupervised methods for clustering and visualization in the Taverna platform is important to support a data-driven scientific discovery in complex and explorative bioinformatics applications. Results: This work presents a Taverna plugin, the Biological Data Interactive Clustering Explorer (BioDICE), that performs clustering of high-dimensional biological data and provides a nonlinear, topology preserving projection for the visualization of the input data and their similarities. The core algorithm in the BioDICE plugin is Fast Learning Self Organizing Map (FLSOM), which is an improved variant of the Self Organizing Map (SOM) algorithm. The plugin generates an interactive 2D map that allows the visual exploration of multidimensional data and the identification of groups of similar objects. The effectiveness of the plugin is demonstrated on a case study related to chemical compounds. Conclusions: The number and variety of available tools and its extensibility have made Taverna a popular choice for the development of scientific data workflows. This work presents a novel plugin, BioDICE, which adds a data-driven knowledge discovery component to Taverna. BioDICE provides an effective and powerful clustering tool, which can be adopted for the explorative analysis of biological datasets.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The treatment of auditory-verbal short-term memory (STM) deficits in aphasia is a growing avenue of research (Martin & Reilly, 2012; Murray, 2012). STM treatment requires time precision, which is suited to computerised delivery. We have designed software, which provides STM treatment for aphasia. The treatment is based on matching listening span tasks (Howard & Franklin, 1990), aiming to improve the temporal maintenance of multi-word sequences (Salis, 2012). The person listens to pairs of word-lists that differ in word-order and decides if the pairs are the same or different. This approach does not require speech output and is suitable for persons with aphasia who have limited or no output. We describe the software and how its review from clinicians shaped its design.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

An analysis of diabatic heating and moistening processes from 12-36 hour lead time forecasts from 12 Global Circulation Models are presented as part of the "Vertical structure and physical processes of the Madden-Julian Oscillation (MJO)" project. A lead time of 12-36 hours is chosen to constrain the large scale dynamics and thermodynamics to be close to observations while avoiding being too close to the initial spin-up for the models as they adjust to being driven from the YOTC analysis. A comparison of the vertical velocity and rainfall with the observations and YOTC analysis suggests that the phases of convection associated with the MJO are constrained in most models at this lead time although the rainfall in the suppressed phase is typically overestimated. Although the large scale dynamics is reasonably constrained, moistening and heating profiles have large inter-model spread. In particular, there are large spreads in convective heating and moistening at mid-levels during the transition to active convection. Radiative heating and cloud parameters have the largest relative spread across models at upper levels during the active phase. A detailed analysis of time step behaviour shows that some models show strong intermittency in rainfall and differences in the precipitation and dynamics relationship between models. The wealth of model outputs archived during this project is a very valuable resource for model developers beyond the study of the MJO. In addition, the findings of this study can inform the design of process model experiments, and inform the priorities for field experiments and future observing systems.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Policies to control air quality focus on mitigating emissions of aerosols and their precursors, and other short-lived climate pollutants (SLCPs). On a local scale, these policies will have beneficial impacts on health and crop yields, by reducing particulate matter (PM) and surface ozone concentrations; however, the climate impacts of reducing emissions of SLCPs are less straightforward to predict. In this paper we consider a set of idealised, extreme mitigation strategies, in which the total anthropogenic emissions of individual SLCP emissions species are removed. This provides an upper bound on the potential climate impacts of such air quality strategies. We focus on evaluating the climate responses to changes in anthropogenic emissions of aerosol precursor species: black carbon (BC), organic carbon (OC) and sulphur dioxide (SO2). We perform climate integrations with four fully coupled atmosphere-ocean global climate models (AOGCMs), and examine the effects on global and regional climate of removing the total land-based anthropogenic emissions of each of the three aerosol precursor species. We find that the SO2 emissions reductions lead to the strongest response, with all three models showing an increase in surface temperature focussed in the northern hemisphere high latitudes, and a corresponding increase in global mean precipitation and run-off. Changes in precipitation and run-off patterns are driven mostly by a northward shift in the ITCZ, consistent with the hemispherically asymmetric warming pattern driven by the emissions changes. The BC and OC emissions reductions give a much weaker forcing signal, and there is some disagreement between models in the sign of the climate responses to these perturbations. These differences between models are due largely to natural variability in sea-ice extent, circulation patterns and cloud changes. This large natural variability component to the signal when the ocean circulation and sea-ice are free-running means that the BC and OC mitigation measures do not necessarily lead to a discernible climate response.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This study investigates the effects of a short-term pedagogic intervention on the development of L2 fluency among learners studying English for Academic purposes (EAP) at a university in the UK. It also examines the interaction between the development of fluency, and complexity and accuracy. Through a pre-test, post-test design, data were collected over a period of four weeks from learners performing monologic tasks. While the Control Group (CG) focused on developing general speaking and listening skills, the Experimental Group (EG) received awareness-raising activities and fluency strategy training in addition to general speaking and listening practice i.e following the syllabus. The data, coded in terms of a range of measures of fluency, accuracy and complexity, were subjected to repeated measures MANOVA, t-tests and correlations. The results indicate that after the intervention, while some fluency gains were achieved by the CG, the EG produced statistically more fluent language demonstrating a faster speech and articulation rate, longer runs and higher phonation time ratios. The significant correlations obtained between measures of accuracy and learners’ pauses in the CG suggest that pausing opportunities may have been linked to accuracy. The findings of the study have significant implications for L2 pedagogy, highlighting the effective impact of instruction on the development of fluency.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Seamless phase II/III clinical trials in which an experimental treatment is selected at an interim analysis have been the focus of much recent research interest. Many of the methods proposed are based on the group sequential approach. This paper considers designs of this type in which the treatment selection can be based on short-term endpoint information for more patients than have primary endpoint data available. We show that in such a case, the familywise type I error rate may be inflated if previously proposed group sequential methods are used and the treatment selection rule is not specified in advance. A method is proposed to avoid this inflation by considering the treatment selection that maximises the conditional error given the data available at the interim analysis. A simulation study is reported that illustrates the type I error rate inflation and compares the power of the new approach with two other methods: a combination testing approach and a group sequential method that does not use the short-term endpoint data, both of which also strongly control the type I error rate. The new method is also illustrated through application to a study in Alzheimer's disease. © 2015 The Authors. Statistics in Medicine Published by John Wiley & Sons Ltd.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This study has investigated serial (temporal) clustering of extra-tropical cyclones simulated by 17 climate models that participated in CMIP5. Clustering was estimated by calculating the dispersion (ratio of variance to mean) of 30 December-February counts of Atlantic storm tracks passing nearby each grid point. Results from single historical simulations of 1975-2005 were compared to those from historical ERA40 reanalyses from 1958-2001 ERA40 and single future model projections of 2069-2099 under the RCP4.5 climate change scenario. Models were generally able to capture the broad features in reanalyses reported previously: underdispersion/regularity (i.e. variance less than mean) in the western core of the Atlantic storm track surrounded by overdispersion/clustering (i.e. variance greater than mean) to the north and south and over western Europe. Regression of counts onto North Atlantic Oscillation (NAO) indices revealed that much of the overdispersion in the historical reanalyses and model simulations can be accounted for by NAO variability. Future changes in dispersion were generally found to be small and not consistent across models. The overdispersion statistic, for any 30 year sample, is prone to large amounts of sampling uncertainty that obscures the climate change signal. For example, the projected increase in dispersion for storm counts near London in the CNRMCM5 model is 0.1 compared to a standard deviation of 0.25. Projected changes in the mean and variance of NAO are insufficient to create changes in overdispersion that are discernible above natural sampling variations.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

ICT clusters have attracted much attention because of their rapid growth and their value for other economic activities. Using a nested multi-level model, we examine how conditions at the country level and at the city level affect ICT clustering activity in 227 cities across 22 European countries. We test for the influence of three country regulations (starting a business, registering property, enforcing contracts) and two city conditions (proximity to university, network density) on ICT clustering. We consider heterogeneity within the sector and study two types of ICT activities: ICT product firms and ICT content firms. Our results indicate that country conditions and city conditions each have idiosyncratic implications for ICT clustering, and further, that these can vary by activities in ICT products or ICT content manufacturing.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This paper presents a summary of the work done within the European Union's Seventh Framework Programme project ECLIPSE (Evaluating the Climate and Air Quality Impacts of Short-Lived Pollutants). ECLIPSE had a unique systematic concept for designing a realistic and effective mitigation scenario for short-lived climate pollutants (SLCPs; methane, aerosols and ozone, and their precursor species) and quantifying its climate and air quality impacts, and this paper presents the results in the context of this overarching strategy. The first step in ECLIPSE was to create a new emission inventory based on current legislation (CLE) for the recent past and until 2050. Substantial progress compared to previous work was made by including previously unaccounted types of sources such as flaring of gas associated with oil production, and wick lamps. These emission data were used for present-day reference simulations with four advanced Earth system models (ESMs) and six chemistry transport models (CTMs). The model simulations were compared with a variety of ground-based and satellite observational data sets from Asia, Europe and the Arctic. It was found that the models still underestimate the measured seasonality of aerosols in the Arctic but to a lesser extent than in previous studies. Problems likely related to the emissions were identified for northern Russia and India, in particular. To estimate the climate impacts of SLCPs, ECLIPSE followed two paths of research: the first path calculated radiative forcing (RF) values for a large matrix of SLCP species emissions, for different seasons and regions independently. Based on these RF calculations, the Global Temperature change Potential metric for a time horizon of 20 years (GTP20) was calculated for each SLCP emission type. This climate metric was then used in an integrated assessment model to identify all emission mitigation measures with a beneficial air quality and short-term (20-year) climate impact. These measures together defined a SLCP mitigation (MIT) scenario. Compared to CLE, the MIT scenario would reduce global methane (CH4) and black carbon (BC) emissions by about 50 and 80 %, respectively. For CH4, measures on shale gas production, waste management and coal mines were most important. For non-CH4 SLCPs, elimination of high-emitting vehicles and wick lamps, as well as reducing emissions from gas flaring, coal and biomass stoves, agricultural waste, solvents and diesel engines were most important. These measures lead to large reductions in calculated surface concentrations of ozone and particulate matter. We estimate that in the EU, the loss of statistical life expectancy due to air pollution was 7.5 months in 2010, which will be reduced to 5.2 months by 2030 in the CLE scenario. The MIT scenario would reduce this value by another 0.9 to 4.3 months. Substantially larger reductions due to the mitigation are found for China (1.8 months) and India (11–12 months). The climate metrics cannot fully quantify the climate response. Therefore, a second research path was taken. Transient climate ensemble simulations with the four ESMs were run for the CLE and MIT scenarios, to determine the climate impacts of the mitigation. In these simulations, the CLE scenario resulted in a surface temperature increase of 0.70 ± 0.14 K between the years 2006 and 2050. For the decade 2041–2050, the warming was reduced by 0.22 ± 0.07 K in the MIT scenario, and this result was in almost exact agreement with the response calculated based on the emission metrics (reduced warming of 0.22 ± 0.09 K). The metrics calculations suggest that non-CH4 SLCPs contribute ~ 22 % to this response and CH4 78 %. This could not be fully confirmed by the transient simulations, which attributed about 90 % of the temperature response to CH4 reductions. Attribution of the observed temperature response to non-CH4 SLCP emission reductions and BC specifically is hampered in the transient simulations by small forcing and co-emitted species of the emission basket chosen. Nevertheless, an important conclusion is that our mitigation basket as a whole would lead to clear benefits for both air quality and climate. The climate response from BC reductions in our study is smaller than reported previously, possibly because our study is one of the first to use fully coupled climate models, where unforced variability and sea ice responses cause relatively strong temperature fluctuations that may counteract (and, thus, mask) the impacts of small emission reductions. The temperature responses to the mitigation were generally stronger over the continents than over the oceans, and with a warming reduction of 0.44 K (0.39–0.49) K the largest over the Arctic. Our calculations suggest particularly beneficial climate responses in southern Europe, where surface warming was reduced by about 0.3 K and precipitation rates were increased by about 15 (6–21) mm yr−1 (more than 4 % of total precipitation) from spring to autumn. Thus, the mitigation could help to alleviate expected future drought and water shortages in the Mediterranean area. We also report other important results of the ECLIPSE project.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In this article, along with others, we take the position that the Null-Subject Parameter (NSP) (Chomsky 1981; Rizzi 1982) cluster of properties is narrower in scope than some originally contended. We test for the resetting of the NSP by English L2 learners of Spanish at the intermediate level, including poverty-of-the stimulus knowledge of the Overt Pronoun Constraint (Montalbetti 1984). Our participants are tested before and after five months' residency in Spain in an effort to see if increased amounts of native exposure are particularly beneficial for parameter resetting. Although we demonstrate NSP resetting for some of the L2 learners, our data essentially demonstrate that even with the advent of time/exposure to native input, there is no immediate gainful effect for NSP resetting.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Clustering methods are increasingly being applied to residential smart meter data, providing a number of important opportunities for distribution network operators (DNOs) to manage and plan the low voltage networks. Clustering has a number of potential advantages for DNOs including, identifying suitable candidates for demand response and improving energy profile modelling. However, due to the high stochasticity and irregularity of household level demand, detailed analytics are required to define appropriate attributes to cluster. In this paper we present in-depth analysis of customer smart meter data to better understand peak demand and major sources of variability in their behaviour. We find four key time periods in which the data should be analysed and use this to form relevant attributes for our clustering. We present a finite mixture model based clustering where we discover 10 distinct behaviour groups describing customers based on their demand and their variability. Finally, using an existing bootstrapping technique we show that the clustering is reliable. To the authors knowledge this is the first time in the power systems literature that the sample robustness of the clustering has been tested.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Parties to the United Nations Framework Convention on Climate Change (UNFCCC) have requested guidance on common greenhouse gas metrics in accounting for Nationally determined contributions (NDCs) to emission reductions1. Metric choice can affect the relative emphasis placed on reductions of ‘cumulative climate pollutants’ such as carbon dioxide versus ‘short-lived climate pollutants’ (SLCPs), including methane and black carbon2, 3, 4, 5, 6. Here we show that the widely used 100-year global warming potential (GWP100) effectively measures the relative impact of both cumulative pollutants and SLCPs on realized warming 20–40 years after the time of emission. If the overall goal of climate policy is to limit peak warming, GWP100 therefore overstates the importance of current SLCP emissions unless stringent and immediate reductions of all climate pollutants result in temperatures nearing their peak soon after mid-century7, 8, 9, 10, which may be necessary to limit warming to “well below 2 °C” (ref. 1). The GWP100 can be used to approximately equate a one-off pulse emission of a cumulative pollutant and an indefinitely sustained change in the rate of emission of an SLCP11, 12, 13. The climate implications of traditional CO2-equivalent targets are ambiguous unless contributions from cumulative pollutants and SLCPs are specified separately.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Subspace clustering groups a set of samples from a union of several linear subspaces into clusters, so that the samples in the same cluster are drawn from the same linear subspace. In the majority of the existing work on subspace clustering, clusters are built based on feature information, while sample correlations in their original spatial structure are simply ignored. Besides, original high-dimensional feature vector contains noisy/redundant information, and the time complexity grows exponentially with the number of dimensions. To address these issues, we propose a tensor low-rank representation (TLRR) and sparse coding-based (TLRRSC) subspace clustering method by simultaneously considering feature information and spatial structures. TLRR seeks the lowest rank representation over original spatial structures along all spatial directions. Sparse coding learns a dictionary along feature spaces, so that each sample can be represented by a few atoms of the learned dictionary. The affinity matrix used for spectral clustering is built from the joint similarities in both spatial and feature spaces. TLRRSC can well capture the global structure and inherent feature information of data, and provide a robust subspace segmentation from corrupted data. Experimental results on both synthetic and real-world data sets show that TLRRSC outperforms several established state-of-the-art methods.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Tensor clustering is an important tool that exploits intrinsically rich structures in real-world multiarray or Tensor datasets. Often in dealing with those datasets, standard practice is to use subspace clustering that is based on vectorizing multiarray data. However, vectorization of tensorial data does not exploit complete structure information. In this paper, we propose a subspace clustering algorithm without adopting any vectorization process. Our approach is based on a novel heterogeneous Tucker decomposition model taking into account cluster membership information. We propose a new clustering algorithm that alternates between different modes of the proposed heterogeneous tensor model. All but the last mode have closed-form updates. Updating the last mode reduces to optimizing over the multinomial manifold for which we investigate second order Riemannian geometry and propose a trust-region algorithm. Numerical experiments show that our proposed algorithm compete effectively with state-of-the-art clustering algorithms that are based on tensor factorization.