968 resultados para Screening-programs


Relevância:

20.00% 20.00%

Publicador:

Resumo:

Saxitoxin (STX) is a low molecular weight neurotoxin mainly produced by certain marine dinoflagellates that, along with its family of similarly related paralytic shellfish toxins, may cause the potentially fatal intoxication known as paralytic shellfish poisoning. Illness and fatality rates are low due to the effective monitoring programs that determine when toxins exceed the established regulatory action level and effectuate shellfish harvesting closures accordingly. Such monitoring programs rely on the ability to rapidly screen large volumes of samples. Many of the screening assays currently available employ antibodies or live animals. This research focused on developing an analytical recognition element that would eliminate the challenges associated with the limited availability of antibodies and the use of animals. Here we report the discovery of a DNA aptamer that targets STX. Concentration-dependent and selective binding of the aptamer to STX was determined using a surface plasmon resonance sensor. Not only does this work represent the first reported aptamer to STX, but also the first aptamer to any marine biotoxin. A novel strategy of using a toxin-protein conjugate for DNA aptamer selection was successfully implemented to overcome the challenges associated with aptamer selection to small molecules. Taking advantage of such an approach could lead to increased diversity and accessibility of aptamers to low molecular weight toxins, which could then be incorporated as analytical recognition elements in diagnostic assays for foodborne toxin detection. The selected STX aptamer sequence is provided here, making it available to any investigator for use in assay development for the detection of STX.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Ecotoxicological screening of dust sampled throughout a Kenyan tannery was conducted using a luminescence (lux)-based bacterial biosensor for both solid and liquid assays. This was complemented by chemical analysis in an attempt to identify possible causative toxic components. The biosensor results showed a highly significant (p <0.001) difference in both solid and liquid phase toxicity in samples collected from various identified sampling points in the tannery. A positive correlation was observed between results of the solid and liquid phase techniques, for most of the sampling points indicating that the toxic contaminants were bioavailable both in the solid and liquid state. However, the results generally indicated toxicity associated with liquid phase except certain areas in solid phase such as chemical handling, buffing area and weighing. The most toxic tannery area identified was the weighing area (p <0.001), showing the lowest bioluminescence for both the solid (0.38 +/- 2.21) and liquid phases (0.01 +/- 0.001). Chromium was the metal present in the highest concentration indicating levels higher than the stipulated regulatory requirement of 0.5 mg Cr/m3 for total Cr (highest Cr concentration was at chemical handling at 209.24 mg l(-1)) in all dust samples. The weighing area had the highest Ni concentration (1.87 mg l(-1)) and the chemical handling area showed the highest Zn concentration (31.9 mg l(-1)). These results raise environmental health concerns, as occupational exposure to dust samples from this site has been shown to give rise to elevated concentrations (above the stipulated levels) of chromium in blood, urine and some body tissues, with inhalation being the main route. Health and Safety Executive (HSE), UK, and American Conference of Governmental Industrial Hygienist (ACGIH) and National Institute for Occupational Safety and Health (NIOSH), USA stipulates an occupational exposure limit of 0.5 mg Cr/m3 (8 h TWA) for total chromium. However, schedule 1 of Controls of substances hazardous to health (COSHH) regulations developed by HSE, indicate 0.05 mg m3 (8 h TWA reference periods) to be the limit for Cr (VI) exposure. The exposure limit for individual (e.g., Cr, Zn, Ni etc.) contaminants (homogeneity) was not exceeded, but potential impact of heterogeneity (multi-element synergistic effect) on toxicity requires application of the precautionary principle.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Most parallel computing applications in highperformance computing use the Message Passing Interface (MPI) API. Given the fundamental importance of parallel computing to science and engineering research, application correctness is paramount. MPI was originally developed around 1993 by the MPI Forum, a group of vendors, parallel programming researchers, and computational scientists. However, the document defining the standard is not issued by an official standards organization but has become a de facto standard © 2011 ACM.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

We propose a dynamic verification approach for large-scale message passing programs to locate correctness bugs caused by unforeseen nondeterministic interactions. This approach hinges on an efficient protocol to track the causality between nondeterministic message receive operations and potentially matching send operations. We show that causality tracking protocols that rely solely on logical clocks fail to capture all nuances of MPI program behavior, including the variety of ways in which nonblocking calls can complete. Our approach is hinged on formally defining the matches-before relation underlying the MPI standard, and devising lazy update logical clock based algorithms that can correctly discover all potential outcomes of nondeterministic receives in practice. can achieve the same coverage as a vector clock based algorithm while maintaining good scalability. LLCP allows us to analyze realistic MPI programs involving a thousand MPI processes, incurring only modest overheads in terms of communication bandwidth, latency, and memory consumption. © 2011 IEEE.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Consideration of the ethical, social, and policy implications of research has become increasingly important to scientists and scholars whose work focuses on brain and mind, but limited empirical data exist on the education in ethics available to them. We examined the current landscape of ethics training in neuroscience programs, beginning with the Canadian context specifically, to elucidate the perceived needs of mentors and trainees and offer recommendations for resource development to meet those needs. We surveyed neuroscientists at all training levels and interviewed directors of neuroscience programs and training grants. A total of 88% of survey respondents reported general interest in ethics, and 96% indicated a desire for more ethics content as it applies to brain research and clinical translation. Expert interviews revealed formal ethics education in over half of programs and in 90% of grants-based programs. Lack of time, resources, and expertise, however, are major barriers to expanding ethics content in neuroscience education. We conclude with an initial set of recommendations to address these barriers which includes the development of flexible, tailored ethics education tools, increased financial support for ethics training, and strategies for fostering collaboration between ethics experts, neuroscience program directors, and funding agencies. © 2010 the Authors. Journal Compilation © 2010 International Mind, Brain, and Education Society and Blackwell Publishing, Inc.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The aim of this study was to compare time-domain waveform analysis of second-trimester uterine artery Doppler using the resistance index (RI) with waveform analysis using a mathematical tool known as wavelet transform for the prediction of pre-eclampsia (PE). This was a retrospective, nested case-cohort study of 336 women, 37 of whom subsequently developed PE. Uterine artery Doppler waveforms were analysed using both RI and waveform analysis. The utility of these indices in screening for PE was then evaluated using receiver operating characteristic curves. There were significant differences in uterine artery RI between the PE women and those with normal pregnancy outcome. After wavelet analysis, significant difference in the mean amplitude in wavelet frequency band 4 was noted between the 2 groups. The sensitivity for both Doppler RI and frequency band 4 for the detection of PE at a 10% false-positive rate was 45%. This small study demonstrates the application of wavelet transform analysis of uterine artery Doppler waveforms in screening for PE. Further prospective studies are needed in order to clearly define if this analytical approach to waveform analysis may have the potential to improve the detection of PE by uterine artery Doppler screening.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

PURPOSE: To assess the comparative accuracy of potential screening tests for open angle glaucoma (OAG).

METHODS: Medline, Embase, Biosis (to November 2005), Science Citation Index (to December 2005), and The Cochrane Library (Issue 4, 2005) were searched. Studies assessing candidate screening tests for detecting OAG in persons older than 40 years that reported true and false positives and negatives were included. Meta-analysis was undertaken using the hierarchical summary receiver operating characteristic model.

RESULTS: Forty studies enrolling over 48,000 people reported nine tests. Most tests were reported by only a few studies. Frequency-doubling technology (FDT; C-20-1) was significantly more sensitive than ophthalmoscopy (30, 95% credible interval [CrI] 0-62) and Goldmann applanation tonometry (GAT; 45, 95% CrI 17-68), whereas threshold standard automated perimetry (SAP) and Heidelberg Retinal Tomograph (HRT II) were both more sensitive than GAT (41, 95% CrI 14-64 and 39, 95% CrI 3-64, respectively). GAT was more specific than both FDT C-20-5 (19, 95% CrI 0-53) and threshold SAP (14, 95% CrI 1-37). Judging performance by diagnostic odds ratio, FDT, oculokinetic perimetry, and HRT II are promising tests. Ophthalmoscopy, SAP, retinal photography, and GAT had relatively poor performance as single tests. These findings are based on heterogeneous data of limited quality and as such are associated with considerable uncertainty.

CONCLUSIONS: No test or group of tests was clearly superior for glaucoma screening. Further research is needed to evaluate the comparative accuracy of the most promising tests.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

BACKGROUND: To compare the ability of Glaucoma Progression Analysis (GPA) and Threshold Noiseless Trend (TNT) programs to detect visual-field deterioration.

METHODS: Patients with open-angle glaucoma followed for a minimum of 2 years and a minimum of seven reliable visual fields were included. Progression was assessed subjectively by four masked glaucoma experts, and compared with GPA and TNT results. Each case was judged to be stable, deteriorated or suspicious of deterioration

RESULTS: A total of 56 eyes of 42 patients were followed with a mean of 7.8 (SD 1.0) tests over an average of 5.5 (1.04) years. Interobserver agreement to detect progression was good (mean kappa = 0.57). Progression was detected in 10-19 eyes by the experts, in six by GPA and in 24 by TNT. Using the consensus expert opinion as the gold standard (four clinicians detected progression), the GPA sensitivity and specificity were 75% and 83%, respectively, while the TNT sensitivity and specificity was 100% and 77%, respectively.

CONCLUSION: TNT showed greater concordance with the experts than GPA in the detection of visual-field deterioration. GPA showed a high specificity but lower sensitivity, mainly detecting cases of high focality and pronounced mean defect slopes.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Familial hypercholesterolemia (FH) is a common single gene disorder, which predisposes to coronary artery disease. In a previous study, we have shown that in patients with definite FH around 20% had no identifiable gene defect after screening the entire exon coding area of the low density lipoprotein receptor (LDLR) and testing for the common Apolipoprotein B (ApoB) R3500Q mutation. In this study, we have extended the screen to additional families and have included the non-coding intron splice regions of the gene. In families with definite FH (tendon xanthoma present, n = 68) the improved genetic screening protocol increased the detection rate of mutations to 87%. This high detection rate greatly enhances the potential value of this test as part of a clinical screening program for FH. In contrast, the use of a limited screen in patients with possible FH (n = 130) resulted in a detection rate of 26%, but this is still of significant benefit in diagnosis of this genetic condition. We have also shown that 14% of LDLR defects are due to splice site mutations and that the most frequent splice mutation in our series (c.1845 + 11 c > g) is expressed at the RNA level. In addition, DNA samples from the patients in whom no LDLR or ApoB gene mutations were found, were sequenced for the NARC-1 gene. No mutations were identified which suggests that the role of NARC-1 in causing FH is minor. In a small proportion of families (

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The aim of this study was to develop a mutation screening protocol for familial hypercholesterolaemia (FH) patients and to assess genotype/phenotype effects in terms of pre-treatment lipid profiles and presentation of tendon xanthomata (TX). A total of 158 families with clinical definitions of possible (120) or definite (38) FH were studied using a tiered screening protocol. Mutations were identified in 52 families, 44 families showing 23 different LDLR gene defects and eight families showing the common Apo B100 gene defect R3500Q. LDLR defects were detected in various regions of the gene with 56% in the LDL binding domain (exons 2-6) and 37% in the EGF precursor homology domain (exons 7-14). The most common mutations were D461N(7), C210X(5), 932delA(5), and C163Y(4). Frameshift mutations accounted for 20% with nonsense 13%, mis-sense 35%, splice 3%, Apo B 13% and 2% large deletion, 13% of clinically definite FH remained undefined. In conclusion, DNA based diagnosis is possible in 79% (30/38) of clinically definite FH families and of the 120 possible FH families at the start of the screening program, 18% (22/120) now have defined mutations. Overall 60 families from the original 158 meet the clinical and/or genetic criteria for definite FH. Tendon xanthomata were present in only 58% (30/52) of genetically defined FH families, thus limiting its use as a strict diagnostic criteria. Families with low density lipoprotein receptor (LDLR) defects present with higher total and LDL cholesterol levels and a higher incidence of TX than do those with the common Apo B variant, and frameshift mutations appear to have the most severe presentation. Copyright (C) 1999 Elsevier Science Ireland Ltd.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Abstract
BACKGROUND:
Glaucoma is a leading cause of blindness. Early detection is advocated but there is insufficient evidence from randomized controlled trials (RCTs) to inform health policy on population screening. Primarily, there is no agreed screening intervention. For a screening programme, agreement is required on the screening tests to be used, either individually or in combination, the person to deliver the test and the location where testing should take place. This study aimed to use ophthalmologists (who were experienced glaucoma subspecialists), optometrists, ophthalmic nurses and patients to develop a reduced set of potential screening tests and testing arrangements that could then be explored in depth in a further study of their feasibility for evaluation in a glaucoma screening RCT.
METHODS:
A two-round Delphi survey involving 38 participants was conducted. Materials were developed from a prior evidence synthesis. For round one, after some initial priming questions in four domains, specialists were asked to nominate three screening interventions, the intervention being a combination of the four domains; target population, (age and higher risk groups), site, screening test and test operator (provider). More than 250 screening interventions were identified. For round two, responses were condensed into 72 interventions and each was rated by participants on a 0-10 scale in terms of feasibility.
RESULTS:
Using a cut-off of a median rating of feasibility of =5.5 as evidence of agreement of intervention feasibility, six interventions were identified from round 2. These were initiating screening at age 50, with a combination of two or three screening tests (varying combinations of tonometry/measures of visual function/optic nerve damage) organized in a community setting with an ophthalmic trained technical assistant delivering the tests. An alternative intervention was a 'glaucoma risk score' ascertained by questionnaire. The advisory panel recommended that further exploration of the feasibility of screening higher risk populations and detailed specification of the screening tests was required.
CONCLUSIONS:
With systematic use of expert opinions, a shortlist of potential screening interventions was identified. Views of users, service providers and cost-effectiveness modeling are now required to identify a feasible intervention to evaluate in a future glaucoma screening trial.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Objectives: To assess whether open angle glaucoma (OAG) screening meets the UK National Screening Committee criteria, to compare screening strategies with case finding, to estimate test parameters, to model estimates of cost and cost-effectiveness, and to identify areas for future research. Data sources: Major electronic databases were searched up to December 2005. Review methods: Screening strategies were developed by wide consultation. Markov submodels were developed to represent screening strategies. Parameter estimates were determined by systematic reviews of epidemiology, economic evaluations of screening, and effectiveness (test accuracy, screening and treatment). Tailored highly sensitive electronic searches were undertaken. Results: Most potential screening tests reviewed had an estimated specificity of 85% or higher. No test was clearly most accurate, with only a few, heterogeneous studies for each test. No randomised controlled trials (RCTs) of screening were identified. Based on two treatment RCTs, early treatment reduces the risk of progression. Extrapolating from this, and assuming accelerated progression with advancing disease severity, without treatment the mean time to blindness in at least one eye was approximately 23 years, compared to 35 years with treatment. Prevalence would have to be about 3-4% in 40 year olds with a screening interval of 10 years to approach cost-effectiveness. It is predicted that screening might be cost-effective in a 50-year-old cohort at a prevalence of 4% with a 10-year screening interval. General population screening at any age, thus, appears not to be cost-effective. Selective screening of groups with higher prevalence (family history, black ethnicity) might be worthwhile, although this would only cover 6% of the population. Extension to include other at-risk cohorts (e.g. myopia and diabetes) would include 37% of the general population, but the prevalence is then too low for screening to be considered cost-effective. Screening using a test with initial automated classification followed by assessment by a specialised optometrist, for test positives, was more cost-effective than initial specialised optometric assessment. The cost-effectiveness of the screening programme was highly sensitive to the perspective on costs (NHS or societal). In the base-case model, the NHS costs of visual impairment were estimated as £669. If annual societal costs were £8800, then screening might be considered cost-effective for a 40-year-old cohort with 1% OAG prevalence assuming a willingness to pay of £30,000 per quality-adjusted life-year. Of lesser importance were changes to estimates of attendance for sight tests, incidence of OAG, rate of progression and utility values for each stage of OAG severity. Cost-effectiveness was not particularly sensitive to the accuracy of screening tests within the ranges observed. However, a highly specific test is required to reduce large numbers of false-positive referrals. The findings that population screening is unlikely to be cost-effective are based on an economic model whose parameter estimates have considerable uncertainty, in particular, if rate of progression and/or costs of visual impairment are higher than estimated then screening could be cost-effective. Conclusions: While population screening is not cost-effective, the targeted screening of high-risk groups may be. Procedures for identifying those at risk, for quality assuring the programme, as well as adequate service provision for those screened positive would all be needed. Glaucoma detection can be improved by increasing attendance for eye examination, and improving the performance of current testing by either refining practice or adding in a technology-based first assessment, the latter being the more cost-effective option. This has implications for any future organisational changes in community eye-care services. Further research should aim to develop and provide quality data to populate the economic model, by conducting a feasibility study of interventions to improve detection, by obtaining further data on costs of blindness, risk of progression and health outcomes, and by conducting an RCT of interventions to improve the uptake of glaucoma testing. © Queen's Printer and Controller of HMSO 2007. All rights reserved.