948 resultados para COMBINING CLASSIFIERS


Relevância:

20.00% 20.00%

Publicador:

Resumo:

The human immunodeficiency virus type-1 (HIV-1) genome contains multiple, highly conserved structural RNA domains that play key roles in essential viral processes. Interference with the function of these RNA domains either by disrupting their structures or by blocking their interaction with viral or cellular factors may seriously compromise HIV-1 viability. RNA aptamers are amongst the most promising synthetic molecules able to interact with structural domains of viral genomes. However, aptamer shortening up to their minimal active domain is usually necessary for scaling up production, what requires very time-consuming, trial-and-error approaches. Here we report on the in vitro selection of 64 nt-long specific aptamers against the complete 5' -untranslated region of HIV-1 genome, which inhibit more than 75% of HIV-1 production in a human cell line. The analysis of the selected sequences and structures allowed for the identification of a highly conserved 16 nt-long stem-loop motif containing a common 8 nt-long apical loop. Based on this result, an in silico designed 16 nt-long RNA aptamer, termed RNApt16, was synthesized, with sequence 5'-CCCCGGCAAGGAGGGG-3-'. The HIV-1 inhibition efficiency of such an aptamer was close to 85%, thus constituting the shortest RNA molecule so far described that efficiently interferes with HIV-1 replication.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

When combining remote sensing imagery with statistical classifiers to obtain categorical thematic maps it is not usual to provide data about the spatial distribution of the error and uncertainty of the resulting maps. This paper describes, in the context of GeoViQua FP7 project, feasible approaches for methods based on several steps such as hybrid classifiers. Both for “per pixel” and “per polygon” strategies, the proposal is based on the use of the available ground truth, which is used to properly model the spatial distribution of the errors. Results allow mapping the classification success with a very high level of reliability (R2>0,94), providing users a sound knowledge of the accuracy at every area of the map.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Social media has become an effective channel for communicating both trends and public opinion on current events. However the automatic topic classification of social media content pose various challenges. Topic classification is a common technique used for automatically capturing themes that emerge from social media streams. However, such techniques are sensitive to the evolution of topics when new event-dependent vocabularies start to emerge (e.g., Crimea becoming relevant to War Conflict during the Ukraine crisis in 2014). Therefore, traditional supervised classification methods which rely on labelled data could rapidly become outdated. In this paper we propose a novel transfer learning approach to address the classification task of new data when the only available labelled data belong to a previous epoch. This approach relies on the incorporation of knowledge from DBpedia graphs. Our findings show promising results in understanding how features age, and how semantic features can support the evolution of topic classifiers.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Several analysis protocols have been tested to identify early visual field losses in glaucoma patients using the mfVEP technique, some were successful in detection of field defects, which were comparable to the standard SAP visual field assessment, and others were not very informative and needed more adjustment and research work. In this study we implemented a novel analysis approach and evaluated its validity and whether it could be used effectively for early detection of visual field defects in glaucoma. The purpose of this study is to examine the benefit of adding mfVEP hemifield Intersector analysis protocol to the standard HFA test when there is suspicious glaucomatous visual field loss. 3 groups were tested in this study; normal controls (38 eyes), glaucoma patients (36 eyes) and glaucoma suspect patients (38 eyes). All subjects had a two standard Humphrey visual field HFA test 24-2, optical coherence tomography of the optic nerve head, and a single mfVEP test undertaken in one session. Analysis of the mfVEP results was done using the new analysis protocol; the Hemifield Sector Analysis HSA protocol. The retinal nerve fibre (RNFL) thickness was recorded to identify subjects with suspicious RNFL loss. The hemifield Intersector analysis of mfVEP results showed that signal to noise ratio (SNR) difference between superior and inferior hemifields was statistically significant between the 3 groups (ANOVA p<0.001 with a 95% CI). The difference between superior and inferior hemispheres in all subjects were all statistically significant in the glaucoma patient group 11/11 sectors (t-test p<0.001), partially significant 5/11 in glaucoma suspect group (t-test p<0.01) and no statistical difference between most sectors in normal group (only 1/11 was significant) (t-test p<0.9). Sensitivity and specificity of the HSA protocol in detecting glaucoma was 97% and 86% respectively, while for glaucoma suspect were 89% and 79%. The use of SAP and mfVEP results in subjects with suspicious glaucomatous visual field defects, identified by low RNFL thickness, is beneficial in confirming early visual field defects. The new HSA protocol used in the mfVEP testing can be used to detect glaucomatous visual field defects in both glaucoma and glaucoma suspect patient. Using this protocol in addition to SAP analysis can provide information about focal visual field differences across the horizontal midline, and confirm suspicious field defects. Sensitivity and specificity of the mfVEP test showed very promising results and correlated with other anatomical changes in glaucoma field loss. The Intersector analysis protocol can detect early field changes not detected by standard HFA test.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Background: DNA-binding proteins play a pivotal role in various intra- and extra-cellular activities ranging from DNA replication to gene expression control. Identification of DNA-binding proteins is one of the major challenges in the field of genome annotation. There have been several computational methods proposed in the literature to deal with the DNA-binding protein identification. However, most of them can't provide an invaluable knowledge base for our understanding of DNA-protein interactions. Results: We firstly presented a new protein sequence encoding method called PSSM Distance Transformation, and then constructed a DNA-binding protein identification method (SVM-PSSM-DT) by combining PSSM Distance Transformation with support vector machine (SVM). First, the PSSM profiles are generated by using the PSI-BLAST program to search the non-redundant (NR) database. Next, the PSSM profiles are transformed into uniform numeric representations appropriately by distance transformation scheme. Lastly, the resulting uniform numeric representations are inputted into a SVM classifier for prediction. Thus whether a sequence can bind to DNA or not can be determined. In benchmark test on 525 DNA-binding and 550 non DNA-binding proteins using jackknife validation, the present model achieved an ACC of 79.96%, MCC of 0.622 and AUC of 86.50%. This performance is considerably better than most of the existing state-of-the-art predictive methods. When tested on a recently constructed independent dataset PDB186, SVM-PSSM-DT also achieved the best performance with ACC of 80.00%, MCC of 0.647 and AUC of 87.40%, and outperformed some existing state-of-the-art methods. Conclusions: The experiment results demonstrate that PSSM Distance Transformation is an available protein sequence encoding method and SVM-PSSM-DT is a useful tool for identifying the DNA-binding proteins. A user-friendly web-server of SVM-PSSM-DT was constructed, which is freely accessible to the public at the web-site on http://bioinformatics.hitsz.edu.cn/PSSM-DT/.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

In wireless sensor networks where nodes are powered by batteries, it is critical to prolong the network lifetime by minimizing the energy consumption of each node. In this paper, the cooperative multiple-input-multiple-output (MIMO) and data-aggregation techniques are jointly adopted to reduce the energy consumption per bit in wireless sensor networks by reducing the amount of data for transmission and better using network resources through cooperative communication. For this purpose, we derive a new energy model that considers the correlation between data generated by nodes and the distance between them for a cluster-based sensor network by employing the combined techniques. Using this model, the effect of the cluster size on the average energy consumption per node can be analyzed. It is shown that the energy efficiency of the network can significantly be enhanced in cooperative MIMO systems with data aggregation, compared with either cooperative MIMO systems without data aggregation or data-aggregation systems without cooperative MIMO, if sensor nodes are properly clusterized. Both centralized and distributed data-aggregation schemes for the cooperating nodes to exchange and compress their data are also proposed and appraised, which lead to diverse impacts of data correlation on the energy performance of the integrated cooperative MIMO and data-aggregation systems.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Decomposing output into trend and cyclical components is an uncertain exercise and depends on the method applied. It is an especially dubious task for countries undergoing large structural changes, such as transition countries. Despite their deficiencies, however, univariate detrending methods are frequently adopted for both policy oriented and academic research. This paper proposes a new procedure for combining univariate detrending techniques which is based on revisions of the estimated output gaps adjusted by the variance of and the correlation among output gaps. The procedure is applied to the study of the similarity of business cycles between the euro area and new EU Member States.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

A cikk központi kérdése a vállalati versenyképesség. Két meghatározó menedzsmentmegközelítés – a vevői érték és az erőforrás-, képességalapú stratégiai menedzsment – fogalmi rendszerére, illetve koncepcióira alapozva a szerzők kialakítottak és a "Versenyben a világgal 2004–2006 – gazdasági versenyképességünk vállalati nézőpontból" című kutatás adatbázisát használva teszteltek egy a vállalati versenyképesség belső struktúrájára rávilágító elméleti modellt. A modell empirikus tesztelésének eredménye szerint a kialakított koncepció a vállalati versenyképesség elemzésének hasznos eszköze, mely alkalmas arra, hogy a vizsgált vállalatok esetében igazolja, vagy éppen megcáfolja a meghatározott, ún. összehangolt kompetenciák meglétét. ________The main issue of this article is the competitiveness of firms. Based on conceptual system and concepts of two determinant management approaches – the customer value and the resource and capability based strategic management – the authors have developed and tested a theoretical model of the inside structure of the competitiveness of firms using the database of Competing with the World 2004-2006 – the competitiveness of the Hungarian economy from enterprise point of view. According to the result of the empiric test of the model the developed concept has been a profitable tool of the analysis of the enterprise competitiveness and it has been suited for confirming or confuting determinated so called aligned competencies being in the case of examined firms.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Hyperthermia is usually used at a sub-lethal level in cancer treatment to potentiate the effects of chemotherapy. The purpose of this study is to investigate the role of heating rate in achieving synergistic cell killing by chemotherapy and hyperthermia. For this purpose, in vitro cell culture experiments with a uterine cancer cell line (MES-SA) and its multidrug resistant (MDR) variant MES-SA/Dx5 were conducted. The cytotoxicity, mode of cell death, induction of thermal tolerance and P-gp mediated MDR following the two different modes of heating were studied. Doxorubicin (DOX) was used as the chemotherapy drug. Indocyanine green (ICG), which absorbs near infrared light at 808nm (ideal for tissue penetration), was chosen for achieving rapid rate hyperthermia. A slow rate hyperthermia was provided by a cell culture incubator. The results show that the potentiating effect of hyperthermia to chemotherapy can be maximized by increasing the rate of heating as evident by the results from the cytotoxicity assay. When delivered at the same thermal dose, a rapid increase in temperature from 37°C to 43°C caused more cell membrane damage than gradually heating the cells from 37°C to 43°C and thus allowed for more intracellular accumulation of the chemotherapeutic agents. Different modes of cell death are observed by the two hyperthermia delivery methods. The rapid rate laser-ICG hyperthermia @ 43°C caused cell necrosis whereas the slow rate incubator hyperthermia @ 43°C induced very mild apoptosis. At 43°C a positive correlation between thermal tolerance and the length of hyperthermia exposure is identified. This study shows that by increasing the rate of heating, less thermal dose is needed in order to overcome P-gp mediated MDR.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Long-term management plans for restoration of natural flow conditions through the Everglades increase the importance of understanding potential nutrient impacts of increased freshwater delivery on Florida Bay biogeochemistry. Planktonic communities respond quickly to changes in water quality, thus spatial variability in community composition and relationships to nutrient parameters must be understood in order to evaluate future downstream impacts of modifications to Everglades hydrology. Here we present initial results combining flow cytometry analyses of phytoplankton and bacterial populations (0.1–50 μm size fraction) with measurements of δ13C and δ15N composition and dissolved inorganic nutrient concentrations to explore proxies for planktonic species assemblage compositions and nutrient cycling. Particulate organic material in the 0.1–50 μm size fraction was collected from five stations in Northeastern and Western Florida Bay to characterize spatial variability in species assemblage and stable isotopic composition. A dense bloom of the picocyanobacterium, Synechococcus elongatus, was observed at Western Florida Bay sites. Smaller Synechococcus sp. were present at Northeast sites in much lower abundance. Bacteria and detrital particles were also more abundant at Western Florida Bay stations than in the northeast region. The highest abundance of detritus occurred at Trout Creek, which receives freshwater discharge from the Everglades through Taylor Slough. In terms of nutrient availability and stable isotopic values, the S. elongatus population in the Western bay corresponded to low DIN (0.5 μM NH 4 + ; 0.2 μM NO 3 − ) concentrations and depleted δ15N signatures ranging from +0.3 to +0.8‰, suggesting that the bloom supported high productivity levels through N2-fixation. δ15N values from the Northeast bay were more enriched (+2.0 to +3.0‰), characteristic of N-recycling. δ13C values were similar for all marine Florida Bay stations, ranging from −17.6 to −14.4‰, however were more depleted at the mangrove ecotone station (−25.5 to −22.3‰). The difference in the isotopic values reflects differences in carbon sources. These findings imply that variations in resource availability and nutrient sources exert significant control over planktonic community composition, which is reflected by stable isotopic signatures.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Funding — Forest Enterprise Scotland and the University of Aberdeen provided funding for the project. The Carnegie Trust supported the lead author, E. McHenry, in this research through the award of a tuition fees bursary.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The chapter discusses both the complementary factors and contradictions of adopting ERP based systems with enterprise 2.0. ERP is characterized as achieving efficient business performance by enabling a standardized business process design, but at a cost of flexibility in operations. It is claimed that enterprise 2.0 can support flexible business process management and so incorporate informal and less structured interactions. A traditional view however is that efficiency and flexibility objectives are incompatible as they are different business objectives which are pursued separately in different organizational environments. Thus an ERP system with a primary objective of improving efficiency and an enterprise 2.0 system with a primary aim of improving flexibility may represent a contradiction and lead to a high risk of failure if adopted simultaneously. This chapter will use case study analysis to investigate the use of a combination of ERP and enterprise 2.0 in a single enterprise with the aim of improving both efficiency and flexibility in operations. The chapter provides an in-depth analysis of the combination of ERP with enterprise 2.0 based on social-technical information systems management theory. The chapter also provides a summary of the benefits of the combination of ERP systems and enterprise 2.0 and how they could contribute to the development of a new generation of business management that combines both formal and informal mechanisms. For example, the multiple-sites or informal communities of an enterprise could collaborate efficiently with a common platform with a certain level of standardization but also have the flexibility in order to provide an agile reaction to internal and external events.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This paper will look at the benefits and limitations of content distribution using Forward Error Correction (FEC) in conjunction with the Transmission Control Protocol (TCP). FEC can be used to reduce the number of retransmissions which would usually result from a lost packet. The requirement for TCP to deal with any losses is then greatly reduced. There are however side-effects to using FEC as a countermeasure to packet loss: an additional requirement for bandwidth. When applications such as real-time video conferencing are needed, delay must be kept to a minimum, and retransmissions are certainly not desirable. A balance, therefore, between additional bandwidth and delay due to retransmissions must be struck. Our results show that the throughput of data can be significantly improved when packet loss occurs using a combination of FEC and TCP, compared to relying solely on TCP for retransmissions. Furthermore, a case study applies the result to demonstrate the achievable improvements in the quality of streaming video perceived by end users.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Acute respiratory infections caused by bacterial or viral pathogens are among the most common reasons for seeking medical care. Despite improvements in pathogen-based diagnostics, most patients receive inappropriate antibiotics. Host response biomarkers offer an alternative diagnostic approach to direct antimicrobial use. This observational cohort study determined whether host gene expression patterns discriminate noninfectious from infectious illness and bacterial from viral causes of acute respiratory infection in the acute care setting. Peripheral whole blood gene expression from 273 subjects with community-onset acute respiratory infection (ARI) or noninfectious illness, as well as 44 healthy controls, was measured using microarrays. Sparse logistic regression was used to develop classifiers for bacterial ARI (71 probes), viral ARI (33 probes), or a noninfectious cause of illness (26 probes). Overall accuracy was 87% (238 of 273 concordant with clinical adjudication), which was more accurate than procalcitonin (78%, P < 0.03) and three published classifiers of bacterial versus viral infection (78 to 83%). The classifiers developed here externally validated in five publicly available data sets (AUC, 0.90 to 0.99). A sixth publicly available data set included 25 patients with co-identification of bacterial and viral pathogens. Applying the ARI classifiers defined four distinct groups: a host response to bacterial ARI, viral ARI, coinfection, and neither a bacterial nor a viral response. These findings create an opportunity to develop and use host gene expression classifiers as diagnostic platforms to combat inappropriate antibiotic use and emerging antibiotic resistance.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Surveys can collect important data that inform policy decisions and drive social science research. Large government surveys collect information from the U.S. population on a wide range of topics, including demographics, education, employment, and lifestyle. Analysis of survey data presents unique challenges. In particular, one needs to account for missing data, for complex sampling designs, and for measurement error. Conceptually, a survey organization could spend lots of resources getting high-quality responses from a simple random sample, resulting in survey data that are easy to analyze. However, this scenario often is not realistic. To address these practical issues, survey organizations can leverage the information available from other sources of data. For example, in longitudinal studies that suffer from attrition, they can use the information from refreshment samples to correct for potential attrition bias. They can use information from known marginal distributions or survey design to improve inferences. They can use information from gold standard sources to correct for measurement error.

This thesis presents novel approaches to combining information from multiple sources that address the three problems described above.

The first method addresses nonignorable unit nonresponse and attrition in a panel survey with a refreshment sample. Panel surveys typically suffer from attrition, which can lead to biased inference when basing analysis only on cases that complete all waves of the panel. Unfortunately, the panel data alone cannot inform the extent of the bias due to attrition, so analysts must make strong and untestable assumptions about the missing data mechanism. Many panel studies also include refreshment samples, which are data collected from a random sample of new

individuals during some later wave of the panel. Refreshment samples offer information that can be utilized to correct for biases induced by nonignorable attrition while reducing reliance on strong assumptions about the attrition process. To date, these bias correction methods have not dealt with two key practical issues in panel studies: unit nonresponse in the initial wave of the panel and in the

refreshment sample itself. As we illustrate, nonignorable unit nonresponse

can significantly compromise the analyst's ability to use the refreshment samples for attrition bias correction. Thus, it is crucial for analysts to assess how sensitive their inferences---corrected for panel attrition---are to different assumptions about the nature of the unit nonresponse. We present an approach that facilitates such sensitivity analyses, both for suspected nonignorable unit nonresponse

in the initial wave and in the refreshment sample. We illustrate the approach using simulation studies and an analysis of data from the 2007-2008 Associated Press/Yahoo News election panel study.

The second method incorporates informative prior beliefs about

marginal probabilities into Bayesian latent class models for categorical data.

The basic idea is to append synthetic observations to the original data such that

(i) the empirical distributions of the desired margins match those of the prior beliefs, and (ii) the values of the remaining variables are left missing. The degree of prior uncertainty is controlled by the number of augmented records. Posterior inferences can be obtained via typical MCMC algorithms for latent class models, tailored to deal efficiently with the missing values in the concatenated data.

We illustrate the approach using a variety of simulations based on data from the American Community Survey, including an example of how augmented records can be used to fit latent class models to data from stratified samples.

The third method leverages the information from a gold standard survey to model reporting error. Survey data are subject to reporting error when respondents misunderstand the question or accidentally select the wrong response. Sometimes survey respondents knowingly select the wrong response, for example, by reporting a higher level of education than they actually have attained. We present an approach that allows an analyst to model reporting error by incorporating information from a gold standard survey. The analyst can specify various reporting error models and assess how sensitive their conclusions are to different assumptions about the reporting error process. We illustrate the approach using simulations based on data from the 1993 National Survey of College Graduates. We use the method to impute error-corrected educational attainments in the 2010 American Community Survey using the 2010 National Survey of College Graduates as the gold standard survey.