365 resultados para false positive rates

em Queensland University of Technology - ePrints Archive


Relevância:

100.00% 100.00%

Publicador:

Resumo:

The objective of the study was to determine, through meta-analysis, the rate of confirmed false reports of sexual assault to police. The meta-analysis initially involved a search for relevant articles. The search revealed seven studies where researchers or their trained helpers evaluated reported sexual assault cases to determine the rate of confirmed false reports. The meta-analysis calculated an overall rate and tested for possible moderators of effect size. The meta-analytic rate of false reports of sexual assault was .052 (95% CIs .030, .089). The rates for the individual studies were heterogeneous, suggesting the possibility of moderators of rate. However, the four possible moderators examined, year of publication, whether the data set used had information in addition to police reports, whether the study was completed in the U.S. or elsewhere, and whether inter-rater reliabilities were reported, were all not significant. The meta-analysis of seven relevant studies shows that confirmed false allegations of sexual assault made to police occur at a significant rate. The total false reporting rate, including both confirmed and equivocal cases, would be greater than the 5 percent rate found here.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

- Background In the UK, women aged 50–73 years are invited for screening by mammography every 3 years. In 2009–10, more than 2.24 million women in this age group in England were invited to take part in the programme, of whom 73% attended a screening clinic. Of these, 64,104 women were recalled for assessment. Of those recalled, 81% did not have breast cancer; these women are described as having a false-positive mammogram. - Objective The aim of this systematic review was to identify the psychological impact on women of false-positive screening mammograms and any evidence for the effectiveness of interventions designed to reduce this impact. We were also looking for evidence of effects in subgroups of women. - Data sources MEDLINE, MEDLINE In-Process & Other Non-Indexed Citations, EMBASE, Health Management Information Consortium, Cochrane Central Register for Controlled Trials, Cochrane Database of Systematic Reviews, Centre for Reviews and Dissemination (CRD) Database of Abstracts of Reviews of Effects, CRD Health Technology Assessment (HTA), Cochrane Methodology, Web of Science, Science Citation Index, Social Sciences Citation Index, Conference Proceedings Citation Index-Science, Conference Proceeding Citation Index-Social Science and Humanities, PsycINFO, Cumulative Index to Nursing and Allied Health Literature, Sociological Abstracts, the International Bibliography of the Social Sciences, the British Library's Electronic Table of Contents and others. Initial searches were carried out between 8 October 2010 and 25 January 2011. Update searches were carried out on 26 October 2011 and 23 March 2012. - Review methods Based on the inclusion criteria, titles and abstracts were screened independently by two reviewers. Retrieved papers were reviewed and selected using the same independent process. Data were extracted by one reviewer and checked by another. Each included study was assessed for risk of bias. - Results Eleven studies were found from 4423 titles and abstracts. Studies that used disease-specific measures found a negative psychological impact lasting up to 3 years. Distress increased with the level of invasiveness of the assessment procedure. Studies using instruments designed to detect clinical levels of morbidity did not find this effect. Women with false-positive mammograms were less likely to return for the next round of screening [relative risk (RR) 0.97; 95% confidence interval (CI) 0.96 to 0.98] than those with normal mammograms, were more likely to have interval cancer [odds ratio (OR) 3.19 (95% CI 2.34 to 4.35)] and were more likely to have cancer detected at the next screening round [OR 2.15 (95% CI 1.55 to 2.98)]. - Limitations This study was limited to UK research and by the robustness of the included studies, which frequently failed to report quality indicators, for example failure to consider the risk of bias or confounding, or failure to report participants' demographic characteristics. - Conclusions We conclude that the experience of having a false-positive screening mammogram can cause breast cancer-specific psychological distress that may endure for up to 3 years, and reduce the likelihood that women will return for their next round of mammography screening. These results should be treated cautiously owing to inherent weakness of observational designs and weaknesses in reporting. Future research should include a qualitative interview study and observational studies that compare generic and disease-specific measures, collect demographic data and include women from different social and ethnic groups.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

- Objectives To identify the psychological effects of false-positive screening mammograms in the UK. - Methods Systematic review of all controlled studies and qualitative studies of women with a false-positive screening mammogram. The control group participants had normal mammograms. All psychological outcomes including returning for routine screening were permitted. All studies had a narrative synthesis. - Results The searches returned seven includable studies (7/4423). Heterogeneity was such that meta-analysis was not possible. Studies using disease-specific measures found that, compared to normal results, there could be enduring psychological distress that lasted up to 3 years; the level of distress was related to the degree of invasiveness of the assessment. At 3 years the relative risks were, further mammography, 1.28 (95% CI 0.82 to 2.00), fine needle aspiration 1.80 (95% CI 1.17 to 2.77), biopsy 2.07 (95% CI 1.22 to 3.52) and early recall 1.82 (95% CI 1.22 to 2.72). Studies that used generic measures of anxiety and depression found no such impact up to 3 months after screening. Evidence suggests that women with false-positive mammograms have an increased likelihood of failing to reattend for routine screening, relative risk 0.97 (95% CI 0.96 to 0.98) compared with women with normal mammograms. - Conclusions Having a false-positive screening mammogram can cause breast cancer-specific distress for up to 3 years. The degree of distress is related to the invasiveness of the assessment. Women with false-positive mammograms are less likely to return for routine assessment than those with normal ones.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Exponential growth of genomic data in the last two decades has made manual analyses impractical for all but trial studies. As genomic analyses have become more sophisticated, and move toward comparisons across large datasets, computational approaches have become essential. One of the most important biological questions is to understand the mechanisms underlying gene regulation. Genetic regulation is commonly investigated and modelled through the use of transcriptional regulatory network (TRN) structures. These model the regulatory interactions between two key components: transcription factors (TFs) and the target genes (TGs) they regulate. Transcriptional regulatory networks have proven to be invaluable scientific tools in Bioinformatics. When used in conjunction with comparative genomics, they have provided substantial insights into the evolution of regulatory interactions. Current approaches to regulatory network inference, however, omit two additional key entities: promoters and transcription factor binding sites (TFBSs). In this study, we attempted to explore the relationships among these regulatory components in bacteria. Our primary goal was to identify relationships that can assist in reducing the high false positive rates associated with transcription factor binding site predictions and thereupon enhance the reliability of the inferred transcription regulatory networks. In our preliminary exploration of relationships between the key regulatory components in Escherichia coli transcription, we discovered a number of potentially useful features. The combination of location score and sequence dissimilarity scores increased de novo binding site prediction accuracy by 13.6%. Another important observation made was with regards to the relationship between transcription factors grouped by their regulatory role and corresponding promoter strength. Our study of E.coli ��70 promoters, found support at the 0.1 significance level for our hypothesis | that weak promoters are preferentially associated with activator binding sites to enhance gene expression, whilst strong promoters have more repressor binding sites to repress or inhibit gene transcription. Although the observations were specific to �70, they nevertheless strongly encourage additional investigations when more experimentally confirmed data are available. In our preliminary exploration of relationships between the key regulatory components in E.coli transcription, we discovered a number of potentially useful features { some of which proved successful in reducing the number of false positives when applied to re-evaluate binding site predictions. Of chief interest was the relationship observed between promoter strength and TFs with respect to their regulatory role. Based on the common assumption, where promoter homology positively correlates with transcription rate, we hypothesised that weak promoters would have more transcription factors that enhance gene expression, whilst strong promoters would have more repressor binding sites. The t-tests assessed for E.coli �70 promoters returned a p-value of 0.072, which at 0.1 significance level suggested support for our (alternative) hypothesis; albeit this trend may only be present for promoters where corresponding TFBSs are either all repressors or all activators. Nevertheless, such suggestive results strongly encourage additional investigations when more experimentally confirmed data will become available. Much of the remainder of the thesis concerns a machine learning study of binding site prediction, using the SVM and kernel methods, principally the spectrum kernel. Spectrum kernels have been successfully applied in previous studies of protein classification [91, 92], as well as the related problem of promoter predictions [59], and we have here successfully applied the technique to refining TFBS predictions. The advantages provided by the SVM classifier were best seen in `moderately'-conserved transcription factor binding sites as represented by our E.coli CRP case study. Inclusion of additional position feature attributes further increased accuracy by 9.1% but more notable was the considerable decrease in false positive rate from 0.8 to 0.5 while retaining 0.9 sensitivity. Improved prediction of transcription factor binding sites is in turn extremely valuable in improving inference of regulatory relationships, a problem notoriously prone to false positive predictions. Here, the number of false regulatory interactions inferred using the conventional two-component model was substantially reduced when we integrated de novo transcription factor binding site predictions as an additional criterion for acceptance in a case study of inference in the Fur regulon. This initial work was extended to a comparative study of the iron regulatory system across 20 Yersinia strains. This work revealed interesting, strain-specific difierences, especially between pathogenic and non-pathogenic strains. Such difierences were made clear through interactive visualisations using the TRNDifi software developed as part of this work, and would have remained undetected using conventional methods. This approach led to the nomination of the Yfe iron-uptake system as a candidate for further wet-lab experimentation due to its potential active functionality in non-pathogens and its known participation in full virulence of the bubonic plague strain. Building on this work, we introduced novel structures we have labelled as `regulatory trees', inspired by the phylogenetic tree concept. Instead of using gene or protein sequence similarity, the regulatory trees were constructed based on the number of similar regulatory interactions. While the common phylogentic trees convey information regarding changes in gene repertoire, which we might regard being analogous to `hardware', the regulatory tree informs us of the changes in regulatory circuitry, in some respects analogous to `software'. In this context, we explored the `pan-regulatory network' for the Fur system, the entire set of regulatory interactions found for the Fur transcription factor across a group of genomes. In the pan-regulatory network, emphasis is placed on how the regulatory network for each target genome is inferred from multiple sources instead of a single source, as is the common approach. The benefit of using multiple reference networks, is a more comprehensive survey of the relationships, and increased confidence in the regulatory interactions predicted. In the present study, we distinguish between relationships found across the full set of genomes as the `core-regulatory-set', and interactions found only in a subset of genomes explored as the `sub-regulatory-set'. We found nine Fur target gene clusters present across the four genomes studied, this core set potentially identifying basic regulatory processes essential for survival. Species level difierences are seen at the sub-regulatory-set level; for example the known virulence factors, YbtA and PchR were found in Y.pestis and P.aerguinosa respectively, but were not present in both E.coli and B.subtilis. Such factors and the iron-uptake systems they regulate, are ideal candidates for wet-lab investigation to determine whether or not they are pathogenic specific. In this study, we employed a broad range of approaches to address our goals and assessed these methods using the Fur regulon as our initial case study. We identified a set of promising feature attributes; demonstrated their success in increasing transcription factor binding site prediction specificity while retaining sensitivity, and showed the importance of binding site predictions in enhancing the reliability of regulatory interaction inferences. Most importantly, these outcomes led to the introduction of a range of visualisations and techniques, which are applicable across the entire bacterial spectrum and can be utilised in studies beyond the understanding of transcriptional regulatory networks.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

A major challenge in neuroscience is finding which genes affect brain integrity, connectivity, and intellectual function. Discovering influential genes holds vast promise for neuroscience, but typical genome-wide searches assess approximately one million genetic variants one-by-one, leading to intractable false positive rates, even with vast samples of subjects. Even more intractable is the question of which genes interact and how they work together to affect brain connectivity. Here, we report a novel approach that discovers which genes contribute to brain wiring and fiber integrity at all pairs of points in a brain scan. We studied genetic correlations between thousands of points in human brain images from 472 twins and their nontwin siblings (mean age: 23.7 2.1 SD years; 193 male/279 female).Wecombined clustering with genome-wide scanning to find brain systems withcommongenetic determination.Wethen filtered the image in a new way to boost power to find causal genes. Using network analysis, we found a network of genes that affect brain wiring in healthy young adults. Our new strategy makes it computationally more tractable to discover genes that affect brain integrity. The gene network showed small-world and scale-free topologies, suggesting efficiency in genetic interactions and resilience to network disruption. Genetic variants at hubs of the network influence intellectual performance by modulating associations between performance intelligence quotient and the integrity of major white matter tracts, such as the callosal genu and splenium, cingulum, optic radiations, and the superior longitudinal fasciculus.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Objective: To describe patient participation and clinical performance in a colorectal cancer (CRC) screening program utilising faecal occult blood test (FOBT). Methods: A community-based intervention was conducted in a small, rural community in north Queensland, 2000/01. One of two FOBT kits – guaiac (Hemoccult-ll) or immunochemical (Inform) – was assigned by general practice and mailed to participants (3,358 patients aged 50–74 years listed with the local practices). Results: Overall participation in FOBT screening was 36.3%. Participation was higher with the immunochemical kit than the guaiac kit (OR=1.9, 95% Cl 1.6-2.2). Women were more likely to comply with testing than men (OR=1.4, 95% Cl 1.2-1.7), and people in their 60s were less likely to participate than those 70–74 years (OR=0.8, 95% Cl 0.6-0.9). The positivity rate was higher for the immunochemical (9.5%) than the guaiac (3.9%) test (χ2=9.2, p=0.002), with positive predictive values for cancer or adenoma of advanced pathology of 37.8% (95% Cl 28.1–48.6) for !nform and 40.0% (95% Cl 16.8–68.7) for Hemoccult-ll. Colonoscopy follow-up was 94.8% with a medical complication rate of 2–3%. Conclusions: An immunochemical FOBT enhanced participation. Higher positivity rates for this kit did not translate into higher false-positive rates, and both test types resulted in a high yield of neoplasia. Implications: In addition to type of FOBT, the ultimate success of a population-based screening program for CRC using FOBT will depend on appropriate education of health professionals and the public as well as significant investment in medical infrastructure for colonoscopy follow-up.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Objective: To compare the effectiveness of the STRATIFY falls tool with nurses’ clinical judgments in predicting patient falls. Study Design and Setting: A prospective cohort study was conducted among the inpatients of an acute tertiary hospital. Participants were patients over 65 years of age admitted to any hospital unit. Sensitivity, specificity, and positive predictive value (PPV) and negative predictive values (NPV) of the instrument and nurses’ clinical judgments in predicting falls were calculated. Results: Seven hundred and eighty-eight patients were screened and followed up during the study period. The fall prevalence was 9.2%. Of the 335 patients classified as being ‘‘at risk’’ for falling using the STRATIFY tool, 59 (17.6%) did sustain a fall (sensitivity50.82, specificity50.61, PPV50.18, NPV50.97). Nurses judged that 501 patients were at risk of falling and, of these, 60 (12.0%) fell (sensitivity50.84, specificity50.38, PPV50.12, NPV50.96). The STRATIFY tool correctly identified significantly more patients as either fallers or nonfallers than the nurses (P50.027). Conclusion: Considering the poor specificity and high rates of false-positive results for both the STRATIFY tool and nurses’ clinical judgments, we conclude that neither of these approaches are useful for screening of falls in acute hospital settings.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Acoustic sensors provide an effective means of monitoring biodiversity at large spatial and temporal scales. They can continuously and passively record large volumes of data over extended periods, however these data must be analysed to detect the presence of vocal species. Automated analysis of acoustic data for large numbers of species is complex and can be subject to high levels of false positive and false negative results. Manual analysis by experienced users can produce accurate results, however the time and effort required to process even small volumes of data can make manual analysis prohibitive. Our research examined the use of sampling methods to reduce the cost of analysing large volumes of acoustic sensor data, while retaining high levels of species detection accuracy. Utilising five days of manually analysed acoustic sensor data from four sites, we examined a range of sampling rates and methods including random, stratified and biologically informed. Our findings indicate that randomly selecting 120, one-minute samples from the three hours immediately following dawn provided the most effective sampling method. This method detected, on average 62% of total species after 120 one-minute samples were analysed, compared to 34% of total species from traditional point counts. Our results demonstrate that targeted sampling methods can provide an effective means for analysing large volumes of acoustic sensor data efficiently and accurately.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Background Detection of outbreaks is an important part of disease surveillance. Although many algorithms have been designed for detecting outbreaks, few have been specifically assessed against diseases that have distinct seasonal incidence patterns, such as those caused by vector-borne pathogens. Methods We applied five previously reported outbreak detection algorithms to Ross River virus (RRV) disease data (1991-2007) for the four local government areas (LGAs) of Brisbane, Emerald, Redland and Townsville in Queensland, Australia. The methods used were the Early Aberration Reporting System (EARS) C1, C2 and C3 methods, negative binomial cusum (NBC), historical limits method (HLM), Poisson outbreak detection (POD) method and the purely temporal SaTScan analysis. Seasonally-adjusted variants of the NBC and SaTScan methods were developed. Some of the algorithms were applied using a range of parameter values, resulting in 17 variants of the five algorithms. Results The 9,188 RRV disease notifications that occurred in the four selected regions over the study period showed marked seasonality, which adversely affected the performance of some of the outbreak detection algorithms. Most of the methods examined were able to detect the same major events. The exception was the seasonally-adjusted NBC methods that detected an excess of short signals. The NBC, POD and temporal SaTScan algorithms were the only methods that consistently had high true positive rates and low false positive and false negative rates across the four study areas. The timeliness of outbreak signals generated by each method was also compared but there was no consistency across outbreaks and LGAs. Conclusions This study has highlighted several issues associated with applying outbreak detection algorithms to seasonal disease data. In lieu of a true gold standard, a quantitative comparison is difficult and caution should be taken when interpreting the true positives, false positives, sensitivity and specificity.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Yao, Begg, and Livingston (1996, Biometrics 52, 992-1001) considered the optimal group size for testing a series of potentially therapeutic agents to identify a promising one as soon as possible for given error rates. The number of patients to be tested with each agent was fixed as the group size. We consider a sequential design that allows early acceptance and rejection, and we provide an optimal strategy to minimize the sample sizes (patients) required using Markov decision processes. The minimization is under the constraints of the two types (false positive and false negative) of error probabilities, with the Lagrangian multipliers corresponding to the cost parameters for the two types of errors. Numerical studies indicate that there can be a substantial reduction in the number of patients required.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Species distribution modelling (SDM) typically analyses species’ presence together with some form of absence information. Ideally absences comprise observations or are inferred from comprehensive sampling. When such information is not available, then pseudo-absences are often generated from the background locations within the study region of interest containing the presences, or else absence is implied through the comparison of presences to the whole study region, e.g. as is the case in Maximum Entropy (MaxEnt) or Poisson point process modelling. However, the choice of which absence information to include can be both challenging and highly influential on SDM predictions (e.g. Oksanen and Minchin, 2002). In practice, the use of pseudo- or implied absences often leads to an imbalance where absences far outnumber presences. This leaves analysis highly susceptible to ‘naughty-noughts’: absences that occur beyond the envelope of the species, which can exert strong influence on the model and its predictions (Austin and Meyers, 1996). Also known as ‘excess zeros’, naughty noughts can be estimated via an overall proportion in simple hurdle or mixture models (Martin et al., 2005). However, absences, especially those that occur beyond the species envelope, can often be more diverse than presences. Here we consider an extension to excess zero models. The two-staged approach first exploits the compartmentalisation provided by classification trees (CTs) (as in O’Leary, 2008) to identify multiple sources of naughty noughts and simultaneously delineate several species envelopes. Then SDMs can be fit separately within each envelope, and for this stage, we examine both CTs (as in Falk et al., 2014) and the popular MaxEnt (Elith et al., 2006). We introduce a wider range of model performance measures to improve treatment of naughty noughts in SDM. We retain an overall measure of model performance, the area under the curve (AUC) of the Receiver-Operating Curve (ROC), but focus on its constituent measures of false negative rate (FNR) and false positive rate (FPR), and how these relate to the threshold in the predicted probability of presence that delimits predicted presence from absence. We also propose error rates more relevant to users of predictions: false omission rate (FOR), the chance that a predicted absence corresponds to (and hence wastes) an observed presence, and the false discovery rate (FDR), reflecting those predicted (or potential) presences that correspond to absence. A high FDR may be desirable since it could help target future search efforts, whereas zero or low FOR is desirable since it indicates none of the (often valuable) presences have been ignored in the SDM. For illustration, we chose Bradypus variegatus, a species that has previously been published as an exemplar species for MaxEnt, proposed by Phillips et al. (2006). We used CTs to increasingly refine the species envelope, starting with the whole study region (E0), eliminating more and more potential naughty noughts (E1–E3). When combined with an SDM fit within the species envelope, the best CT SDM had similar AUC and FPR to the best MaxEnt SDM, but otherwise performed better. The FNR and FOR were greatly reduced, suggesting that CTs handle absences better. Interestingly, MaxEnt predictions showed low discriminatory performance, with the most common predicted probability of presence being in the same range (0.00-0.20) for both true absences and presences. In summary, this example shows that SDMs can be improved by introducing an initial hurdle to identify naughty noughts and partition the envelope before applying SDMs. This improvement was barely detectable via AUC and FPR yet visible in FOR, FNR, and the comparison of predicted probability of presence distribution for pres/absence.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

This paper investigates the use of the FAB-MAP appearance-only SLAM algorithm as a method for performing visual data association for RatSLAM, a semi-metric full SLAM system. While both systems have shown the ability to map large (60-70km) outdoor locations of approximately the same scale, for either larger areas or across longer time periods both algorithms encounter difficulties with false positive matches. By combining these algorithms using a mapping between appearance and pose space, both false positives and false negatives generated by FAB-MAP are significantly reduced during outdoor mapping using a forward-facing camera. The hybrid FAB-MAP-RatSLAM system developed demonstrates the potential for successful SLAM over large periods of time.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Keyword Spotting is the task of detecting keywords of interest within continu- ous speech. The applications of this technology range from call centre dialogue systems to covert speech surveillance devices. Keyword spotting is particularly well suited to data mining tasks such as real-time keyword monitoring and unre- stricted vocabulary audio document indexing. However, to date, many keyword spotting approaches have su®ered from poor detection rates, high false alarm rates, or slow execution times, thus reducing their commercial viability. This work investigates the application of keyword spotting to data mining tasks. The thesis makes a number of major contributions to the ¯eld of keyword spotting. The ¯rst major contribution is the development of a novel keyword veri¯cation method named Cohort Word Veri¯cation. This method combines high level lin- guistic information with cohort-based veri¯cation techniques to obtain dramatic improvements in veri¯cation performance, in particular for the problematic short duration target word class. The second major contribution is the development of a novel audio document indexing technique named Dynamic Match Lattice Spotting. This technique aug- ments lattice-based audio indexing principles with dynamic sequence matching techniques to provide robustness to erroneous lattice realisations. The resulting algorithm obtains signi¯cant improvement in detection rate over lattice-based audio document indexing while still maintaining extremely fast search speeds. The third major contribution is the study of multiple veri¯er fusion for the task of keyword veri¯cation. The reported experiments demonstrate that substantial improvements in veri¯cation performance can be obtained through the fusion of multiple keyword veri¯ers. The research focuses on combinations of speech background model based veri¯ers and cohort word veri¯ers. The ¯nal major contribution is a comprehensive study of the e®ects of limited training data for keyword spotting. This study is performed with consideration as to how these e®ects impact the immediate development and deployment of speech technologies for non-English languages.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

We describe research into the identification of anomalous events and event patterns as manifested in computer system logs. Prototype software has been developed with a capability that identifies anomalous events based on usage patterns or user profiles, and alerts administrators when such events are identified. To reduce the number of false positive alerts we have investigated the use of different user profile training techniques and introduce the use of abstractions to group together applications which are related. Our results suggest that the number of false alerts that are generated is significantly reduced when a growing time window is used for user profile training and when abstraction into groups of applications is used.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Data flow analysis techniques can be used to help assess threats to data confidentiality and integrity in security critical program code. However, a fundamental weakness of static analysis techniques is that they overestimate the ways in which data may propagate at run time. Discounting large numbers of these false-positive data flow paths wastes an information security evaluator's time and effort. Here we show how to automatically eliminate some false-positive data flow paths by precisely modelling how classified data is blocked by certain expressions in embedded C code. We present a library of detailed data flow models of individual expression elements and an algorithm for introducing these components into conventional data flow graphs. The resulting models can be used to accurately trace byte-level or even bit-level data flow through expressions that are normally treated as atomic. This allows us to identify expressions that safely downgrade their classified inputs and thereby eliminate false-positive data flow paths from the security evaluation process. To validate the approach we have implemented and tested it in an existing data flow analysis toolkit.