540 resultados para refining
Resumo:
Objectives: To assess whether open angle glaucoma (OAG) screening meets the UK National Screening Committee criteria, to compare screening strategies with case finding, to estimate test parameters, to model estimates of cost and cost-effectiveness, and to identify areas for future research. Data sources: Major electronic databases were searched up to December 2005. Review methods: Screening strategies were developed by wide consultation. Markov submodels were developed to represent screening strategies. Parameter estimates were determined by systematic reviews of epidemiology, economic evaluations of screening, and effectiveness (test accuracy, screening and treatment). Tailored highly sensitive electronic searches were undertaken. Results: Most potential screening tests reviewed had an estimated specificity of 85% or higher. No test was clearly most accurate, with only a few, heterogeneous studies for each test. No randomised controlled trials (RCTs) of screening were identified. Based on two treatment RCTs, early treatment reduces the risk of progression. Extrapolating from this, and assuming accelerated progression with advancing disease severity, without treatment the mean time to blindness in at least one eye was approximately 23 years, compared to 35 years with treatment. Prevalence would have to be about 3-4% in 40 year olds with a screening interval of 10 years to approach cost-effectiveness. It is predicted that screening might be cost-effective in a 50-year-old cohort at a prevalence of 4% with a 10-year screening interval. General population screening at any age, thus, appears not to be cost-effective. Selective screening of groups with higher prevalence (family history, black ethnicity) might be worthwhile, although this would only cover 6% of the population. Extension to include other at-risk cohorts (e.g. myopia and diabetes) would include 37% of the general population, but the prevalence is then too low for screening to be considered cost-effective. Screening using a test with initial automated classification followed by assessment by a specialised optometrist, for test positives, was more cost-effective than initial specialised optometric assessment. The cost-effectiveness of the screening programme was highly sensitive to the perspective on costs (NHS or societal). In the base-case model, the NHS costs of visual impairment were estimated as £669. If annual societal costs were £8800, then screening might be considered cost-effective for a 40-year-old cohort with 1% OAG prevalence assuming a willingness to pay of £30,000 per quality-adjusted life-year. Of lesser importance were changes to estimates of attendance for sight tests, incidence of OAG, rate of progression and utility values for each stage of OAG severity. Cost-effectiveness was not particularly sensitive to the accuracy of screening tests within the ranges observed. However, a highly specific test is required to reduce large numbers of false-positive referrals. The findings that population screening is unlikely to be cost-effective are based on an economic model whose parameter estimates have considerable uncertainty, in particular, if rate of progression and/or costs of visual impairment are higher than estimated then screening could be cost-effective. Conclusions: While population screening is not cost-effective, the targeted screening of high-risk groups may be. Procedures for identifying those at risk, for quality assuring the programme, as well as adequate service provision for those screened positive would all be needed. Glaucoma detection can be improved by increasing attendance for eye examination, and improving the performance of current testing by either refining practice or adding in a technology-based first assessment, the latter being the more cost-effective option. This has implications for any future organisational changes in community eye-care services. Further research should aim to develop and provide quality data to populate the economic model, by conducting a feasibility study of interventions to improve detection, by obtaining further data on costs of blindness, risk of progression and health outcomes, and by conducting an RCT of interventions to improve the uptake of glaucoma testing. © Queen's Printer and Controller of HMSO 2007. All rights reserved.
Resumo:
This work presents the application of reduced rank regression to the field of systems biology. A computational approach is used to investigate the mechanisms of the janus-associated kinases/signal transducers and transcription factors (JAK/STAT) and mitogen activated protein kinases (MAPK) signal transduction pathways in hepatic cells stimulated by interleukin-6. The results obtained identify the contribution of individual reactions to the dynamics of the model. These findings are compared to previously available results from sensitivity analysis of the model which focused on the parameters involved and their effect. This application of reduced rank regression allows for an understanding of the individual reaction terms involved in the modelled signal transduction pathways and has the benefit of being computationally inexpensive. The obtained results complement existing findings and also confirm the importance of several protein complexes in the MAPK pathway which hints at benefits that can be achieved by further refining the model.
Resumo:
Aim
It is widely acknowledged that species distributions result from a variety of biotic and abiotic factors operating at different spatial scales. Here, we aimed to (1) determine the extent to which global climate niche models (CNMs) can be improved by the addition of fine-scale regional data; (2) examine climatic and environmental factors influencing the range of 15 invasive aquatic plant species; and (3) provide a case study for the use of such models in invasion management on an island.
Location
Global, with a case study of species invasions in Ireland.
Methods
Climate niche models of global extent (including climate only) and regional environmental niche models (with additional factors such as human influence, land use and soil characteristics) were generated using maxent for 15 invasive aquatic plants. The performance of these models within the invaded range of the study species in Ireland was assessed, and potential hotspots of invasion suitability were determined. Models were projected forward up to 2080 based on two climate scenarios.
Results
While climate variables are important in defining the global range of species, factors related to land use and nutrient level were of greater importance in regional projections. Global climatic models were significantly improved at the island scale by the addition of fine-scale environmental variables (area under the curve values increased by 0.18 and true skill statistic values by 0.36), and projected ranges decreased from an average of 86% to 36% of the island.
Main conclusions
Refining CNMs with regional data on land use, human influence and landscape may have a substantial impact on predictive capacity, providing greater value for prioritization of conservation management at subregional or local scales.
Resumo:
A new process for the preparation and surface modification of submicron YAl2 intermetallic particles was proposed to control the agglomeration of ultrafine YAl2 particles and interface in the fabrication of YAl2p/MgLiAl composites. The morphological and structural evolution during mechanical milling of YAl2 powders (< 30 μm) with magnesium particles (~ 100 μm) has been characterized by scanning electron microscopy, transmission electron microscopy, and X-ray diffraction. The results show that YAl2 particles are refined to submicron scale and separately cladded in magnesium coatings after mixed milling with magnesium particles for 20 h. Mechanical and metallurgical bonds have been found in YAl2/Mg interfaces without any interface reactions. Both the refining and mechanical activation efficiencies for YAl2 particles are enhanced, which may be related to the addition of magnesium particles leading to atomic solid solution and playing a role as “dispersion stabilizer”.
Resumo:
This study combined high resolution mass spectrometry (HRMS), advanced chemometrics and pathway enrichment analysis to analyse the blood metabolome of patients attending the memory clinic: cases of mild cognitive impairment (MCI; n = 16), cases of MCI who upon subsequent follow-up developed Alzheimer's disease (MCI_AD; n = 19), and healthy age-matched controls (Ctrl; n = 37). Plasma was extracted in acetonitrile and applied to an Acquity UPLC HILIC (1.7μm x 2.1 x 100 mm) column coupled to a Xevo G2 QTof mass spectrometer using a previously optimised method. Data comprising 6751 spectral features were used to build an OPLS-DA statistical model capable of accurately distinguishing Ctrl, MCI and MCI_AD. The model accurately distinguished (R2 = 99.1%; Q2 = 97%) those MCI patients who later went on to develop AD. S-plots were used to shortlist ions of interest which were responsible for explaining the maximum amount of variation between patient groups. Metabolite database searching and pathway enrichment analysis indicated disturbances in 22 biochemical pathways, and excitingly it discovered two interlinked areas of metabolism (polyamine metabolism and L-Arginine metabolism) were differentially disrupted in this well-defined clinical cohort. The optimised untargeted HRMS methods described herein not only demonstrate that it is possible to distinguish these pathologies in human blood but also that MCI patients 'at risk' from AD could be predicted up to 2 years earlier than conventional clinical diagnosis. Blood-based metabolite profiling of plasma from memory clinic patients is a novel and feasible approach in improving MCI and AD diagnosis and, refining clinical trials through better patient stratification.
Resumo:
BACKGROUND: Tumorigenesis is characterised by changes in transcriptional control. Extensive transcript expression data have been acquired over the last decade and used to classify prostate cancers. Prostate cancer is, however, a heterogeneous multifocal cancer and this poses challenges in identifying robust transcript biomarkers.
METHODS: In this study, we have undertaken a meta-analysis of publicly available transcriptomic data spanning datasets and technologies from the last decade and encompassing laser capture microdissected and macrodissected sample sets.
RESULTS: We identified a 33 gene signature that can discriminate between benign tissue controls and localised prostate cancers irrespective of detection platform or dissection status. These genes were significantly overexpressed in localised prostate cancer versus benign tissue in at least three datasets within the Oncomine Compendium of Expression Array Data. In addition, they were also overexpressed in a recent exon-array dataset as well a prostate cancer RNA-seq dataset generated as part of the The Cancer Genomics Atlas (TCGA) initiative. Biologically, glycosylation was the single enriched process associated with this 33 gene signature, encompassing four glycosylating enzymes. We went on to evaluate the performance of this signature against three individual markers of prostate cancer, v-ets avian erythroblastosis virus E26 oncogene homolog (ERG) expression, prostate specific antigen (PSA) expression and androgen receptor (AR) expression in an additional independent dataset. Our signature had greater discriminatory power than these markers both for localised cancer and metastatic disease relative to benign tissue, or in the case of metastasis, also localised prostate cancer.
CONCLUSION: In conclusion, robust transcript biomarkers are present within datasets assembled over many years and cohorts and our study provides both examples and a strategy for refining and comparing datasets to obtain additional markers as more data are generated.
Resumo:
The endosomal system provides a route whereby nutrients, viruses, and receptors are internalized. During the course of endocytosis, activated receptors can accumulate within endosomal structures and certain signal-transducing molecules can be recruited to endosomal membranes. In the context of signaling and cancer, they provide platforms within the cell from which signals can be potentiated or attenuated. Regulation of the duration of receptor signaling is a pivotal means of refining growth responses in cells. In cancers, this is often considered in terms of mutations that affect receptor tyrosine kinases and maintain them in hyperactivated states of dimerization and/or phosphorylation. However, disruption to the regulatory control exerted by the assembly of protein complexes within the endosomal network can also contribute to disease among which oncogenesis is characterized in part by dysregulated growth, enhanced cell survival, and changes in the expression of markers of differentiation. In this chapter, we will discuss the role of proteins that regulate in endocytosis as tumor suppressors or oncogenes and how changing the fate of internalized receptors and concomitant endosomal signaling can contribute to cancer.
Resumo:
Statement of purpose The purpose of this concurrent session is to present the main findings and recommendations from a five year study evaluating the implementation of Early Warning Systems (EWS) and the Acute Life-threatening Events: Recognition and Treatment (ALERT) course in Northern Ireland. The presentation will provide delegates with an understanding of those factors that enable and constrain successful implementation of EWS and ALERT in practice in order to provide an impetus for change. Methods The research design was a multiple case study approach of four wards in two hospitals in Northern Ireland. It followed the principles of realist evaluation research which allowed empirical data to be gathered to test and refine RRS programme theory [1]. The stages included identifying the programme theories underpinning EWS and ALERT, generating hypotheses, gathering empirical evidence and refining the programme theories. This approach used a variety of mixed methods including individual and focus group interviews, observation and documentary analysis of EWS compliance data and ALERT training records. A within and across case comparison facilitated the development of mid-range theories from the research evidence. Results The official RRS theories developed from the realist synthesis were critically evaluated and compared with the study findings to develop a mid-range theory to explain what works, for whom in what circumstances. The findings of what works suggests that clinical experience, established working relationships, flexible implementation of protocols, ongoing experiential learning, empowerment and pre-emptive management are key to the success of EWS and ALERT implementation. Each concept is presented as ‘context, mechanism and outcome configurations’ to provide an understanding of how the context impacts on individual reasoning or behaviour to produce certain outcomes. Conclusion These findings highlight the combination of factors that can improve the implementation and sustainability of EWS and ALERT and in light of this evidence several recommendations are made to provide policymakers with guidance and direction for future policy development. References: 1. Pawson R and Tilley N. (1997) Realistic Evaluation. Sage Publications; London Type of submission: Concurrent session Source of funding: Sandra Ryan Fellowship funded by the School of Nursing & Midwifery, Queen’s University of Belfast
Resumo:
O consumo de energia a nível mundial aumenta a cada dia, de forma inversa aos recursos fósseis que decrescem de dia para dia. O sector dos transportes é o maior consumidor deste recurso. Face ao actual cenário urge encontrar uma solução renovável e sustentável que permita não só, diminuir a nossa dependência de combustíveis fósseis mas fundamentalmente promover a sua substituição por energias de fontes renováveis. O biodiesel apresenta-se na vanguarda das alternativas aos combustiveis derivados do petróleo, para o sector dos transportes, sendo considerado uma importante opção a curto prazo, uma vez que o seu preço pode ser competitivo com o diesel convencional, e para a sua utilização o motor de combustão não necessita de alterações. O biodiesel é uma mistura líquida, não tóxica, biodegradável de ésteres de ácidos gordos, sem teor de enxofre ou compostos aromáticos, apresenta boa lubricidade, alto número de cetano, e origina emissões gasosas mais limpas. O presente trabalho contribui para um melhor conhecimento da dependência das propriedades termofisicas do biodiesel com a sua composição. A publicação de novos dados permitirá o desenvolvimento de modelos mais fiáveis na previsão do comportamento do biodiesel. As propriedades densidade e viscosidade são o espelho da composição do biodiesel, uma vez que dependem directamente da matéria prima que lhe deu origem, mais do que do processo de produção. Neste trabalho os dados medidos de densidade e viscosidade de biodiesel foram testados com vários modelos e inclusivamente foram propostos novos modelos ajustados para esta família de compostos. Os dados medidos abrangem uma ampla gama de temperaturas e no caso da densidade também foram medidos dados a alta pressão de biodiesel e de alguns ésteres metilico puros. Neste trabalho também são apresentados dados experimentais para o equilíbrio de fases sólido-liquído de biodiesel e equlibrio de fases líquidolíquido de alguns sistemas importantes para a produção de biodiesel. Ambos os tipos de equilíbrio foram descritos por modelos desenvolvidos no nosso laboratório. Uma importância especial é dado aqui a propriedades que dependem do perfil de ácidos gordos da matéria-prima além de densidade e viscosidade; o índice de iodo e temperature limite de filtrabalidade são aqui avaliados com base nas considerações das normas. Os ácidos gordos livres são um sub-produto de refinação de óleo alimentar, que são removidos na desodoração, no processo de purificação do óleo. A catálise enzimática é aqui abordada como alternativa para a conversão destes ácidos gordos livres em biodiesel. Estudou-se a capacidade da lipase da Candida antartica (Novozym 435) para promover a esterificação de ácidos gordos livres com metanol ou etanol, utilizando metodologia de superfície de resposta com planeamento experimental. Avaliou-se a influência de diversas variáveis no rendimento da reacção.
Resumo:
O objeto principal desta tese é o estudo de algoritmos de processamento e representação automáticos de dados, em particular de informação obtida por sensores montados a bordo de veículos (2D e 3D), com aplicação em contexto de sistemas de apoio à condução. O trabalho foca alguns dos problemas que, quer os sistemas de condução automática (AD), quer os sistemas avançados de apoio à condução (ADAS), enfrentam hoje em dia. O documento é composto por duas partes. A primeira descreve o projeto, construção e desenvolvimento de três protótipos robóticos, incluindo pormenores associados aos sensores montados a bordo dos robôs, algoritmos e arquitecturas de software. Estes robôs foram utilizados como plataformas de ensaios para testar e validar as técnicas propostas. Para além disso, participaram em várias competições de condução autónoma tendo obtido muito bons resultados. A segunda parte deste documento apresenta vários algoritmos empregues na geração de representações intermédias de dados sensoriais. Estes podem ser utilizados para melhorar técnicas já existentes de reconhecimento de padrões, deteção ou navegação, e por este meio contribuir para futuras aplicações no âmbito dos AD ou ADAS. Dado que os veículos autónomos contêm uma grande quantidade de sensores de diferentes naturezas, representações intermédias são particularmente adequadas, pois podem lidar com problemas relacionados com as diversas naturezas dos dados (2D, 3D, fotométrica, etc.), com o carácter assíncrono dos dados (multiplos sensores a enviar dados a diferentes frequências), ou com o alinhamento dos dados (problemas de calibração, diferentes sensores a disponibilizar diferentes medições para um mesmo objeto). Neste âmbito, são propostas novas técnicas para a computação de uma representação multi-câmara multi-modal de transformação de perspectiva inversa, para a execução de correcção de côr entre imagens de forma a obter mosaicos de qualidade, ou para a geração de uma representação de cena baseada em primitivas poligonais, capaz de lidar com grandes quantidades de dados 3D e 2D, tendo inclusivamente a capacidade de refinar a representação à medida que novos dados sensoriais são recebidos.
Resumo:
Desulfurization is one of the most important processes in the refining industry. Due to a growing concern about the risks to human health and environment, associated with the emissions of sulfur compounds, legislation has become more stringent, requiring a drastic reduction in the sulfur content of fuel to levels close to zero (< 10 ppm S). However, conventional desulfurization processes are inefficient and have high operating costs. This scenario stimulates the improvement of existing processes and the development of new and more efficient technologies. Aiming at overcoming these shortcomings, this work investigates an alternative desulfurization process using ionic liquids for the removal of mercaptans from "jet fuel" streams. The screening and selection of the most suitable ionic liquid were performed based on experimental and COSMO-RS predicted liquid-liquid equilibrium data. A model feed of 1-hexanethiol and n-dodecane was selected to represent a jet-fuel stream. High selectivities were determined, as a result of the low mutual solubility between the ionic liquid and the hydrocarbon matrix, proving the potential use of the ionic liquid, which prevents the loss of fuel for the solvent. The distribution ratios of mercaptans towards the ionic liquids were not as favorable, making the traditional liquid-liquid extraction processes not suitable for the removal of aliphatic S-compounds due to the high volume of extractant required. This work explores alternative methods and proposes the use of ionic liquids in a separation process assisted by membranes. In the process proposed the ionic liquid is used as extracting solvent of the sulfur species, in a hollow fiber membrane contactor, without co-extracting the other jet-fuel compound. In a second contactor, the ionic liquid is regenerated applying a sweep gas stripping, which allows for its reuse in a closed loop between the two membrane contactors. This integrated extraction/regeneration process of desulfurization produced a jet-fuel model with sulfur content lower than 2 ppm of S, as envisaged by legislation for the use of ultra-low sulfur jet-fuel. This result confirms the high potential for development of ultra-deep desulfurization application.
Resumo:
Desde tempos históricos que diferentes tipos de lama são utilizados para aplicações externas no corpo humano, para fins terapêuticos e cosméticos. As lamas cuja beneficiação e caracterização físico-química são apresentadas nesta dissertação são formadas no ambiente hipersalino que existe nas salinas da Troncalhada e de São Tiago da Fonte, localizadas no estuário do rio Vouga, em Aveiro, Portugal. As salinas são constituídas por uma sequência de tanques onde, no período de verão, da água do mar e por evaporação natural, se produz sal marinho que precipita no último tanque (o cristalizador) do qual é extraído. Na base dos tanques ocorrem lamas que correspondem a sedimentos constituídos por material argiloso de cor preto-cinzento e que incorporam argila, silte, areia, bioclastos, sal, matéria orgânica e gás. A componente inorgânica da lama extraída do cristalizador foi estudada por Difracção de Raios-X (DRX) e Fluorescência de Raios-X (FRX), enquanto a componente orgânica da mesma lama foi estudada por Cromatografia de Gás- Espectrometria de Massa (GC-MS). Os estudos efectuados incidiram sobre amostras representativas de lama, obtidas antes e após refinação e beneficiação a que foi submetida a lama tal-qual colhida nas salinas. Foram utilizados métodos geofísicos para caracterizar e distinguir as lamas depositadas na época de safra e no período de interregno. Para o efeito, foram cravados tubos amostradores no sedimento que reveste o fundo dos tanques, tubos que seguidamente foram transportados para o laboratório para medição da condutividade eléctrica do topo até à base da coluna de sedimento amostrado. A refinação foi efectuada por elutriação de suspensões aquosas de lama utilizando um equipamento desenvolvido para o efeito e que permitiu concentrar a lama fina no overflow. Após floculação, sifonagem da água sobrenadante, dessalinização e centrifugação a lama refinada e beneficiada pôde ser incorporada em formulações com objectivos terapêuticos e cosméticos. O estudo microbiológico efectuado nas amostras de lama com sal e sem sal e na água das salinas permitiu identificar diversos tipos de bactérias e colónias presentes na lama e avaliar também os processos de esterilização testados. Concluiu-se que a lama hipersalina ou dessalinizada resultante do processamento a que foram submetidas, não deve ser aplicada ou incorporada em formulações tal-qual, pelo facto de em ambas terem sido identificadas bactérias como é o caso de Clostridium perfringens. Não obstante, se submetidas a esterilização utilizando autoclave a lama salina refinada e beneficiada poderá ser aplicada como peloide extemporâneo. Assim sendo foram desenvolvidas formulações dermoterapêuticas e dermocosméticas contendo lama beneficiada e dessalinizada e esterilizada termicamente.
Resumo:
Tese de doutoramento, Ciências do Mar ( Processos de Ecossistemas Marinhos), Faculdade de Ciências e Tecnologia, Universidade do Algarve, 2012
Resumo:
Tese de mestrado integrado em Engenharia da Energia e do Ambiente, apresentada à Universidade de Lisboa, através da Faculdade de Ciências, 2016