137 resultados para Sequential extraction
Resumo:
Currently we are facing an overburdening growth of the number of reliable information sources on the Internet. The quantity of information available to everyone via Internet is dramatically growing each year [15]. At the same time, temporal and cognitive resources of human users are not changing, therefore causing a phenomenon of information overload. World Wide Web is one of the main sources of information for decision makers (reference to my research). However our studies show that, at least in Poland, the decision makers see some important problems when turning to Internet as a source of decision information. One of the most common obstacles raised is distribution of relevant information among many sources, and therefore need to visit different Web sources in order to collect all important content and analyze it. A few research groups have recently turned to the problem of information extraction from the Web [13]. The most effort so far has been directed toward collecting data from dispersed databases accessible via web pages (related to as data extraction or information extraction from the Web) and towards understanding natural language texts by means of fact, entity, and association recognition (related to as information extraction). Data extraction efforts show some interesting results, however proper integration of web databases is still beyond us. Information extraction field has been recently very successful in retrieving information from natural language texts, however it is still lacking abilities to understand more complex information, requiring use of common sense knowledge, discourse analysis and disambiguation techniques.
Resumo:
We present an empirical evaluation and comparison of two content extraction methods in HTML: absolute XPath expressions and relative XPath expressions. We argue that the relative XPath expressions, although not widely used, should be used in preference to absolute XPath expressions in extracting content from human-created Web documents. Evaluation of robustness covers four thousand queries executed on several hundred webpages. We show that in referencing parts of real world dynamic HTML documents, relative XPath expressions are on average significantly more robust than absolute XPath ones.
Resumo:
Introduction Clinically, the Cobb angle method measures the overall scoliotic curve in the coronal plane but does not measure individual vertebra and disc wedging. The contributions of the vertebrae and discs in the growing scoliotic spine were measured to investigate coronal plane deformity progression with growth. Methods A 0.49mm isotropic 3D MRI technique was developed to investigate the level-by-level changes that occur in the growing spine of a group of Adolescent Idiopathic Scoliosis (AIS) patients, who received two to four sequential scans (spaced 3-12 months apart). The coronal plane wedge angles of each vertebra and disc in the major curve were measured to capture any changes that occurred during their adolescent growth phase. Results Seventeen patients had at least two scans. Mean patient age was 12.9 years (SD 1.5 years). Sixteen were classified as right-sided major thoracic Lenke Type 1 (one left sided). Mean standing Cobb angle at initial presentation was 31° (SD 12°). Six received two scans, nine three scans and two four scans, with 65% showing a Cobb angle progression of 5° or more between scans. Overall, there was no clear pattern of deformity progression of individual vertebrae and discs, nor between patients who progressed and those who didn’t. There were measurable changes in the wedging of the vertebrae and discs in all patients. In sequential scans, change in direction of wedging was also seen. In several patients there was reverse wedging in the discs that counteracted increased wedging of the vertebrae such that no change in overall Cobb angle was seen. Conclusion Sequential MRI data showed complex patterns of deformity progression. Changes to the wedging of individual vertebrae and discs may occur in patients who have no increase in Cobb angle measure; the Cobb method alone may be insufficient to capture the complex mechanisms of deformity progression.
Resumo:
INTRODUCTION. Clinically, the Cobb angle method measures the overall scoliotic curve in the coronal plane but does not measure individual vertebra and disc wedging. The contributions of the vertebrae and discs in the growing scoliotic spine were measured to investigate coronal plane deformity progression with growth. METHODS. A 0.49mm isotropic 3D MRI technique was developed to investigate the level-by-level changes that occur in the growing spine of a group of Adolescent Idiopathic Scoliosis (AIS) patients, who received two to four sequential scans (spaced 3-12 months apart). The coronal plane wedge angles of each vertebra and disc in the major curve were measured to capture any changes that occurred during their adolescent growth phase. RESULTS. Seventeen patients had at least two scans. Mean patient age was 12.9 years (SD 1.5 years). Sixteen were classified as right-sided major thoracic Lenke Type 1 (one left sided). Mean standing Cobb angle at initial presentation was 31° (SD 12°). Six received two scans, nine three scans and two four scans, with 65% showing a Cobb angle progression of 5° or more between scans. Overall, there was no clear pattern of deformity progression of individual vertebrae and discs, nor between patients who progressed and those who didn’t. There were measurable changes in the wedging of the vertebrae and discs in all patients. In sequential scans, change in direction of wedging was also seen. In several patients there was reverse wedging in the discs that counteracted increased wedging of the vertebrae such that no change in overall Cobb angle was seen. CONCLUSION. Sequential MRI data showed complex patterns of deformity progression. Changes to the wedging of individual vertebrae and discs may occur in patients who have no increase in Cobb angle measure; the Cobb method alone may be insufficient to capture the complex mechanisms of deformity progression.
Resumo:
Clinically, the Cobb angle method measures the overall scoliotic curve in the coronal plane but does not measure individual vertebra and disc wedging. The contributions of the vertebrae and discs in the growing scoliotic spine were measured to investigate coronal plane deformity progression with growth. Sequential MRI data in this project showed complex patterns of deformity progression. Changes to the wedging of individual vertebrae and discs may occur in patients who have no increase in Cobb angle measure; the Cobb method alone may be insufficient to capture the complex mechanisms of deformity progression.
Resumo:
Scientists have injected endotoxin into animals to investigate and understand various pathologies and novel therapies for several decades. Recent observations have shown that there is selective susceptibility to Escherichia coli lipopolysaccharide (LPS) endotoxin in sheep, despite having similar breed characteristics. The reason behind this difference is unknown, and has prompted studies aiming to explain the variation by proteogenomic characterisation of circulating acute phase biomarkers. It is hypothesised that genetic trait, biochemical, immunological and inflammation marker patterns contribute in defining and predicting mammalian response to LPS. This review discusses the effects of endotoxin and host responses, genetic basis of innate defences, activation of the acute phase response (APR) following experimental LPS challenge, and the current approaches employed in detecting novel biomarkers including acute phase proteins (APP) and micro-ribonucleic acids (miRNAs) in serum or plasma. miRNAs are novel targets for elucidating molecular mechanisms of disease because of their differential expression during pathological, and in healthy states. Changes in miRNA profiles during a disease challenge may be reflected in plasma. Studies show that gel-based two-dimensional electrophoresis (2-DE) coupled with either matrix-assisted laser desorption/ionisation time-of-flight mass spectrometry (MALDI-TOF MS) or liquid chromatography-mass spectrometry (LC-MS/MS) are currently the most used methods for proteome characterisation. Further evidence suggests that proteomic investigations are preferentially shifting from 2-DE to non-gel based LC-MS/MS coupled with data extraction by sequential window acquisition of all theoretical fragment-ion spectra (SWATH) approaches that are able to identify a wider range of proteins. Enzyme-linked immunosorbent assay (ELISA), quantitative real-time polymerase chain reaction (qRT-PCR), and most recently proteomic methods have been used to quantify low abundance proteins such as cytokines. qRT-PCR and next generation sequencing (NGS) are used for the characterisation of miRNA. Proteogenomic approaches for detecting APP and novel miRNA profiling are essential in understanding the selective resistance to endotoxin in sheep. The results of these methods could help in understanding similar pathology in humans. It might also be helpful in the development of physiological and diagnostic screening assays for determining experimental inclusion and endpoints, and in clinical trials in future
Resumo:
A method for determination of tricyclazole in water using solid phase extraction and high performance liquid chromatography (HPLC) with UV detection at 230nm and a mobile phase of acetonitrile:water (20:80, v/v) was developed. A performance comparison between two types of solid phase sorbents, the C18 sorbent of Supelclean ENVI-18 cartridge and the styrene-divinyl benzene copolymer sorbent of Sep-Pak PS2-Plus cartridge was conducted. The Sep-Pak PS2-Plus cartridges were found more suitable for extracting tricyclazole from water samples than the Supelclean ENVI-18 cartridges. For this cartridge, both methanol and ethyl acetate produced good results. The method was validated with good linearity and with a limit of detection of 0.008gL-1 for a 500-fold concentration through the SPE procedure. The recoveries of the method were stable at 80% and the precision was from 1.1-6.0% within the range of fortified concentrations. The validated method was also applied to measure the concentrations of tricyclazole in real paddy water.
Resumo:
Frog protection has become increasingly essential due to the rapid decline of its biodiversity. Therefore, it is valuable to develop new methods for studying this biodiversity. In this paper, a novel feature extraction method is proposed based on perceptual wavelet packet decomposition for classifying frog calls in noisy environments. Pre-processing and syllable segmentation are first applied to the frog call. Then, a spectral peak track is extracted from each syllable if possible. Track duration, dominant frequency and oscillation rate are directly extracted from the track. With k-means clustering algorithm, the calculated dominant frequency of all frog species is clustered into k parts, which produce a frequency scale for wavelet packet decomposition. Based on the adaptive frequency scale, wavelet packet decomposition is applied to the frog calls. Using the wavelet packet decomposition coefficients, a new feature set named perceptual wavelet packet decomposition sub-band cepstral coefficients is extracted. Finally, a k-nearest neighbour (k-NN) classifier is used for the classification. The experiment results show that the proposed features can achieve an average classification accuracy of 97.45% which outperforms syllable features (86.87%) and Mel-frequency cepstral coefficients (MFCCs) feature (90.80%).
Resumo:
Background: Haemodialysis nurses form long term relationships with patients in a technologically complex work environment. Previous studies have highlighted that haemodialysis nurses face stressors related to the nature of their work and also their work environments leading to reported high levels of burnout. Using Kanters (1997) Structural Empowerment Theory as a guiding framework, the aim of this study was to explore the factors contributing to satisfaction with the work environment, job satisfaction, job stress and burnout in haemodialysis nurses. Methods: Using a sequential mixed-methods design, the first phase involved an on-line survey comprising demographic and work characteristics, Brisbane Practice Environment Measure (B-PEM), Index of Work Satisfaction(IWS), Nursing Stress Scale (NSS) and the Maslach Burnout Inventory (MBI). The second phase involved conducting eight semi-structured interviews with data thematically analyzed. Results: From the 417 nurses surveyed the majority were female (90.9 %), aged over 41 years of age (74.3 %), and 47.4 % had worked in haemodialysis for more than 10 years. Overall the work environment was perceived positively and there was a moderate level of job satisfaction. However levels of stress and emotional exhaustion (burnout) were high. Two themes, ability to care and feeling successful as a nurse, provided clarity to the level of job satisfaction found in phase 1. While two further themes, patients as quasi-family and intense working teams, explained why working as a haemodialysis nurse was both satisfying and stressful. Conclusions: Nurse managers can use these results to identify issues being experienced by haemodialysis nurses working in the unit they are supervising.
Resumo:
This study investigates the use of unsupervised features derived from word embedding approaches and novel sequence representation approaches for improving clinical information extraction systems. Our results corroborate previous findings that indicate that the use of word embeddings significantly improve the effectiveness of concept extraction models; however, we further determine the influence that the corpora used to generate such features have. We also demonstrate the promise of sequence-based unsupervised features for further improving concept extraction.
Resumo:
Summary. Interim analysis is important in a large clinical trial for ethical and cost considerations. Sometimes, an interim analysis needs to be performed at an earlier than planned time point. In that case, methods using stochastic curtailment are useful in examining the data for early stopping while controlling the inflation of type I and type II errors. We consider a three-arm randomized study of treatments to reduce perioperative blood loss following major surgery. Owing to slow accrual, an unplanned interim analysis was required by the study team to determine whether the study should be continued. We distinguish two different cases: when all treatments are under direct comparison and when one of the treatments is a control. We used simulations to study the operating characteristics of five different stochastic curtailment methods. We also considered the influence of timing of the interim analyses on the type I error and power of the test. We found that the type I error and power between the different methods can be quite different. The analysis for the perioperative blood loss trial was carried out at approximately a quarter of the planned sample size. We found that there is little evidence that the active treatments are better than a placebo and recommended closure of the trial.
Resumo:
Suppose two treatments with binary responses are available for patients with some disease and that each patient will receive one of the two treatments. In this paper we consider the interests of patients both within and outside a trial using a Bayesian bandit approach and conclude that equal allocation is not appropriate for either group of patients. It is suggested that Gittins indices should be used (using an approach called dynamic discounting by choosing the discount rate based on the number of future patients in the trial) if the disease is rare, and the least failures rule if the disease is common. Some analytical and simulation results are provided.
Resumo:
The growth of APIs and Web services on the Internet, especially through larger enterprise systems increasingly being leveraged for Cloud and software-as-a-service opportuni- ties, poses challenges to improving the efficiency of integration with these services. Interfaces of enterprise systems are typically larger, more complex and overloaded, with single operation having multiple data entities and parameter sets, supporting varying requests, and reflecting versioning across different system releases, compared to fine-grained operations of contemporary interfaces. We propose a technique to support the refactoring of service interfaces by deriving business entities and their relationships. In this paper, we focus on the behavioural aspects of service interfaces, aiming to discover the sequential dependencies of operations (otherwise known as protocol extraction) based on the entities and relationships derived. Specifically, we propose heuristics according to these relationships, and in turn, deriving permissible orders in which operations are invoked. As a result of this, service operations can be refactored on business entity CRUD lines, with explicit behavioural protocols as part of an interface definition. This supports flexible service discovery, composition and integration. A prototypical implementation and analysis of existing Web services, including those of commercial logistic systems (Fedex), are used to validate the algorithms proposed through the paper.
Resumo:
A spatial sampling design that uses pair-copulas is presented that aims to reduce prediction uncertainty by selecting additional sampling locations based on both the spatial configuration of existing locations and the values of the observations at those locations. The novelty of the approach arises in the use of pair-copulas to estimate uncertainty at unsampled locations. Spatial pair-copulas are able to more accurately capture spatial dependence compared to other types of spatial copula models. Additionally, unlike traditional kriging variance, uncertainty estimates from the pair-copula account for influence from measurement values and not just the configuration of observations. This feature is beneficial, for example, for more accurate identification of soil contamination zones where high contamination measurements are located near measurements of varying contamination. The proposed design methodology is applied to a soil contamination example from the Swiss Jura region. A partial redesign of the original sampling configuration demonstrates the potential of the proposed methodology.
Resumo:
Organochlorine pesticides (OCPs) are ubiquitous environmental contaminants with adverse impacts on aquatic biota, wildlife and human health even at low concentrations. However, conventional methods for their determination in river sediments are resource intensive. This paper presents an approach that is rapid and also reliable for the detection of OCPs. Accelerated Solvent Extraction (ASE) with in-cell silica gel clean-up followed by Triple Quadrupole Gas Chromatograph Mass Spectrometry (GCMS/MS) was used to recover OCPs from sediment samples. Variables such as temperature, solvent ratio, adsorbent mass and extraction cycle were evaluated and optimised for the extraction. With the exception of Aldrin, which was unaffected by any of the variables evaluated, the recovery of OCPs from sediment samples was largely influenced by solvent ratio and adsorbent mass and, to some extent, the number of cycles and temperature. The optimised conditions for OCPs extraction in sediment with good recoveries were determined to be 4 cycles, 4.5 g of silica gel, 105 ᴼC, and 4:3 v/v DCM: hexane mixture. With the exception of two compounds (α-BHC and Aldrin) whose recoveries were low (59.73 and 47.66 % respectively), the recovery of the other pesticides were in the range 85.35 – 117.97% with precision < 10 % RSD. The method developed significantly reduces sample preparation time, the amount of solvent used, matrix interference, and is highly sensitive and selective.