913 resultados para Methods validation
Resumo:
Tetrodotoxin (TTX) is one of the most potent marine neurotoxins reported. The global distribution of this toxin is spreading with the European Atlantic coastline now being affected. Climate change and increasing pollution have been suggested as underlying causes for this. In the present study, two different sample preparation techniques were used to extract TTX from Trumpet shells and pufferfish samples. Both extraction procedures (accelerated solvent extraction (ASE) and a simple solvent extraction) were shown to provide good recoveries (80-92%). A UPLC-MS/MS method was developed for the analysis of TTX and validated following the guidelines contained in the Commission Decision 2002/657/EC for chemical contaminant analysis. The performance of this procedure was demonstrated to be fit for purpose. This study is the first report on the use of ASE as a mean for TTX extraction, the use of UPLC-MS/MS for TTX analysis, and the validation of this method for TTX in gastropods.
Resumo:
This research validates a computerized dietary selection task (Food-Linked Virtual Response or FLVR) for use in studies of food consumption. In two studies, FLVR task responses were compared with measures of health consciousness, mood, body mass index, personality, cognitive restraint toward food, and actual food selections from a buffet table. The FLVR task was associated with variables which typically predict healthy decision-making and was unrelated to mood or body mass index. Furthermore, the FLVR task predicted participants' unhealthy selections from the buffet, but not overall amount of food. The FLVR task is an inexpensive, valid, and easily administered option for assessing momentary dietary decisions.
Resumo:
Fundação de Amparo à Pesquisa do Estado de São Paulo (FAPESP)
Resumo:
A validação dos métodos de análises em situações fora da rotina, em P&D pesquisa e desenvolvimento, pode tornar-se um processo de difícil solução, em oposição àquelas de rotina, onde as concentrações são conhecidas. Nesta última situação, o método analítico é validado e aprovado com simplicidade. Apesar dos elementos de qualidade básicos para análises em P&D serem os mesmos adotados nas análises de rotina, fatores como o comportamento imprevisível da amostra, a composição desconhecida da matriz e interdependência com o analito, mesmo a incerteza a respeito da escolha do método instrumental, requer uma atenção renovada nos elementos de qualidade existentes. A razão pode ser atribuída à imprevisibilidade do procedimento analítico e a extensão do esforço adicional dos elementos de qualidade. As incertezas das análises em P&D requerem uma cuidadosa consideração do problema, assim como a disponibilidade de técnicas analíticas. Ao mesmo tempo, devem-se observar o planejamento e organização do trabalho, a supervisão do desenvolvimento analítico e a adequação dos aspectos técnico e organizacional - no trabalho de P&D, a definição de responsabilidades e a competência do corpo técnico. A garantia da qualidade nas indústrias farmacêuticas é estabelecida pelo desenvolvimento de especificações apropriadas para matérias-primas, produtos intermediários e produtos finais. A importância de especificações adequadas foi estabelecida e exemplos foram dados no desenvolvimento de especificação típica para produtos farmacêuticos, tais como a vacina e matéria prima. Isto incluiu uma discussão dos tipos de parâmetros que precisam ser controlados e uma indicação dos valores numéricos ou faixas esperadas para muitos materiais utilizados. A metrologia analítica que necessita ser aplicada para demonstrar a observância com as especificações foi discutida, com as implicações de desempenho do método para diferenciar variações na qualidade do produto e a simples variabilidade analítica. Esse trabalho foi escrito para laboratórios, com a finalidade de dar um suporte na implementação da validação de métodos analíticos e introduzir o conceito de incerteza da medida nas análises de rotina. O objetivo foi fornecer de uma maneira compreensível e prática a metodologia dos cálculos das medidas de incerteza baseados principalmente em dados de controle de qualidade e de validação já existentes. Exemplos práticos retirados diretamente do Laboratório de Metrologia e Validação, FIOCRUZ / Bio-Manguinhos, foram apresentados. Entretanto, o tratamento usado neste trabalho é genérico e pode ser aplicável a outros laboratórios químicos
Resumo:
Trabalho Final de mestrado para obtenção do grau de Mestre em Engenharia Química e Biológica
Resumo:
Le traitement de l’épilepsie chez le jeune enfant représente un enjeu majeur pour le développement de ce dernier. Chez la grande majorité des enfants atteints de spasmes infantiles et chez plusieurs atteints de crises partielles complexes réfractaires, le vigabatrin (VGB) représente un traitement incontournable. Cette médication, ayant démontré un haut taux d’efficacité chez cette population, semble toutefois mener à une atteinte du champ visuel périphérique souvent asymptomatique. L’évaluation clinique des champs visuels avec la périmétrie chez les patients de moins de neuf ans d’âge développemental est toutefois très difficile, voire impossible. Les études électrophysiologiques classiques menées auprès de la population épileptique pédiatrique suggèrent l’atteinte des structures liées aux cônes de la rétine. Les protocoles standards ne sont toutefois pas spécifiques aux champs visuels et les déficits soulignés ne concordent pas avec l’atteinte périphérique observée. Cette thèse vise donc à élaborer une tâche adaptée à l’évaluation des champs visuels chez les enfants en utilisant un protocole objectif, rapide et spécifique aux champs visuels à partir des potentiels évoqués visuels (PEVs) et à évaluer, à l’aide de cette méthode, les effets neurotoxiques à long terme du VGB chez des enfants épileptiques exposés en bas âge. La validation de la méthode est présentée dans le premier article. La stimulation est constituée de deux cercles concentriques faits de damiers à renversement de phase alternant à différentes fréquences temporelles. La passation de la tâche chez l’adulte permet de constater qu’une seule électrode corticale (Oz) est nécessaire à l’enregistrement simultané des réponses du champ visuel central et périphérique et qu’il est possible de recueillir les réponses électrophysiologiques très rapidement grâces l’utilisation de l’état-stationnaire (steady-state). La comparaison des données d’enfants et d’adultes normaux permet de constater que les réponses recueillies au sein des deux régions visuelles ne dépendent ni de l’âge ni du sexe. Les réponses centrales sont aussi corrélées à l’acuité visuelle. De plus, la validité de cette méthode est corroborée auprès d’adolescents ayant reçu un diagnostic clinique d’un déficit visuel central ou périphérique. En somme, la méthode validée permet d’évaluer adéquatement les champs visuels corticaux central et périphérique simultanément et rapidement, tant chez les adultes que chez les enfants. Le second article de cette thèse porte sur l’évaluation des champs visuels, grâce à la méthode préalablement validée, d’enfants épileptiques exposés au VGB en jeune âge en comparaison avec des enfants épileptiques exposés à d’autres antiépileptiques et à des enfants neurologiquement sains. La méthode a été bonifiée grâce à la variation du contraste et à l’enregistrement simultané de la réponse rétinienne. On trouve que la réponse corticale centrale est diminuée à haut et à moyen contrastes chez les enfants exposés au VGB et à haut contraste chez les enfants exposés à d’autres antiépileptiques. Le gain de contraste est altéré au sein des deux groupes d’enfants épileptiques. Par contre, l’absence de différences entre les deux groupes neurologiquement atteints ne permet pas de faire la distinction entre l’effet de la médication et celui de la maladie. De plus, la réponse rétinienne périphérique est atteinte chez les enfants épileptiques exposés au Sabril® en comparaison avec les enfants neurologiquement sains. La réponse rétinienne périphérique semble liée à la durée d’exposition à la médication. Ces résultats corroborent ceux rapportés dans la littérature. En somme, les résultats de cette thèse offrent une méthode complémentaire, rapide, fiable, objective à celles connues pour l’évaluation des champs visuels chez les enfants. Ils apportent aussi un éclairage nouveau sur les impacts à long terme possibles chez les enfants exposés au VGB dans la petite enfance.
Resumo:
Coordenação de Aperfeiçoamento de Pessoal de Nível Superior (CAPES)
Resumo:
BACKGROUND: Cathepsin S has been implicated in a variety of malignancies with genetic ablation studies demonstrating a key role in tumor invasion and neo-angiogenesis. Thus, the application of cathepsin S inhibitors may have clinical utility in the treatment of cancer. In this investigation, we applied a cell-permeable dipeptidyl nitrile inhibitor of cathepsin S, originally developed to target cathepsin S in inflammatory diseases, in both in vitro and in vivo tumor models.
METHODS: Validation of cathepsin S selectivity was carried out by assaying fluorogenic substrate turnover using recombinant cathepsin protease. Complete kinetic analysis was carried out and true K i values calculated. Abrogation of tumour invasion using murine MC38 and human MCF7 cell lines were carried out in vitro using a transwell migration assay. Effect on endothelial tube formation was evaluated using primary HUVEC cells. The effect of inhibitor in vivo on MC38 and MCF7 tumor progression was evaluated using cells propagated in C57BL/6 and BALB/c mice respectively. Subsequent immunohistochemical staining of proliferation (Ki67) and apoptosis (TUNEL) was carried out on MCF7 tumors.
RESULTS: We confirmed that this inhibitor was able to selectively target cathepsin S over family members K, V, L and B. The inhibitor also significantly reduced MC38 and MCF7 cell invasion and furthermore, significantly reduced HUVEC endothelial tubule formation in vitro. In vivo analysis revealed that the compound could significantly reduce tumor volume in murine MC38 syngeneic and MCF7 xenograft models. Immunohistochemical analysis of MCF7 tumors revealed cathepsin S inhibitor treatment significantly reduced proliferation and increased apoptosis.
CONCLUSIONS: In summary, these results highlight the characterisation of this nitrile cathepsin S inhibitor using in vitro and in vivo tumor models, presenting a compound which may be used to further dissect the role of cathepsin S in cancer progression and may hold therapeutic potential.
Resumo:
Microbial pollution in water periodically affects human health in Australia, particularly in times of drought and flood. There is an increasing need for the control of waterborn microbial pathogens. Methods, allowing the determination of the origin of faecal contamination in water, are generally referred to as Microbial Source Tracking (MST). Various approaches have been evaluated as indicatorsof microbial pathogens in water samples, including detection of different microorganisms and various host-specific markers. However, until today there have been no universal MST methods that could reliably determine the source (human or animal) of faecal contamination. Therefore, the use of multiple approaches is frequently advised. MST is currently recognised as a research tool, rather than something to be included in routine practices. The main focus of this research was to develop novel and universally applicable methods to meet the demands for MST methods in routine testing of water samples. Escherichia coli was chosen initially as the object organism for our studies as, historically and globally, it is the standard indicator of microbial contamination in water. In this thesis, three approaches are described: single nucleotide polymorphism (SNP) genotyping, clustered regularly interspaced short palindromic repeats (CRISPR) screening using high resolution melt analysis (HRMA) methods and phage detection development based on CRISPR types. The advantage of the combination SNP genotyping and CRISPR genes has been discussed in this study. For the first time, a highly discriminatory single nucleotide polymorphism interrogation of E. coli population was applied to identify the host-specific cluster. Six human and one animal-specific SNP profile were revealed. SNP genotyping was successfully applied in the field investigations of the Coomera watershed, South-East Queensland, Australia. Four human profiles [11], [29], [32] and [45] and animal specific SNP profile [7] were detected in water. Two human-specific profiles [29] and [11] were found to be prevalent in the samples over a time period of years. The rainfall (24 and 72 hours), tide height and time, general land use (rural, suburban), seasons, distance from the river mouth and salinity show a lack of relashionship with the diversity of SNP profiles present in the Coomera watershed (p values > 0.05). Nevertheless, SNP genotyping method is able to identify and distinquish between human- and non-human specific E. coli isolates in water sources within one day. In some samples, only mixed profiles were detected. To further investigate host-specificity in these mixed profiles CRISPR screening protocol was developed, to be used on the set of E. coli, previously analysed for SNP profiles. CRISPR loci, which are the pattern of previous DNA coliphages attacks, were considered to be a promising tool for detecting host-specific markers in E. coli. Spacers in CRISPR loci could also reveal the dynamics of virulence in E. coli as well in other pathogens in water. Despite the fact that host-specificity was not observed in the set of E. coli analysed, CRISPR alleles were shown to be useful in detection of the geographical site of sources. HRMA allows determination of ‘different’ and ‘same’ CRISPR alleles and can be introduced in water monitoring as a cost-effective and rapid method. Overall, we show that the identified human specific SNP profiles [11], [29], [32] and [45] can be useful as marker genotypes globally for identification of human faecal contamination in water. Developed in the current study, the SNP typing approach can be used in water monitoring laboratories as an inexpensive, high-throughput and easy adapted protocol. The unique approach based on E. coli spacers for the search for unknown phage was developed to examine the host-specifity in phage sequences. Preliminary experiments on the recombinant plasmids showed the possibility of using this method for recovering phage sequences. Future studies will determine the host-specificity of DNA phage genotyping as soon as first reliable sequences can be acquired. No doubt, only implication of multiple approaches in MST will allow identification of the character of microbial contamination with higher confidence and readability.
Resumo:
Ambiguity resolution plays a crucial role in real time kinematic GNSS positioning which gives centimetre precision positioning results if all the ambiguities in each epoch are correctly fixed to integers. However, the incorrectly fixed ambiguities can result in large positioning offset up to several meters without notice. Hence, ambiguity validation is essential to control the ambiguity resolution quality. Currently, the most popular ambiguity validation is ratio test. The criterion of ratio test is often empirically determined. Empirically determined criterion can be dangerous, because a fixed criterion cannot fit all scenarios and does not directly control the ambiguity resolution risk. In practice, depending on the underlying model strength, the ratio test criterion can be too conservative for some model and becomes too risky for others. A more rational test method is to determine the criterion according to the underlying model and user requirement. Miss-detected incorrect integers will lead to a hazardous result, which should be strictly controlled. In ambiguity resolution miss-detected rate is often known as failure rate. In this paper, a fixed failure rate ratio test method is presented and applied in analysis of GPS and Compass positioning scenarios. A fixed failure rate approach is derived from the integer aperture estimation theory, which is theoretically rigorous. The criteria table for ratio test is computed based on extensive data simulations in the approach. The real-time users can determine the ratio test criterion by looking up the criteria table. This method has been applied in medium distance GPS ambiguity resolution but multi-constellation and high dimensional scenarios haven't been discussed so far. In this paper, a general ambiguity validation model is derived based on hypothesis test theory, and fixed failure rate approach is introduced, especially the relationship between ratio test threshold and failure rate is examined. In the last, Factors that influence fixed failure rate approach ratio test threshold is discussed according to extensive data simulation. The result shows that fixed failure rate approach is a more reasonable ambiguity validation method with proper stochastic model.
Resumo:
With increasing recognition of the roles RNA molecules and RNA/protein complexes play in an unexpected variety of biological processes, understanding of RNA structure-function relationships is of high current importance. To make clean biological interpretations from three-dimensional structures, it is imperative to have high-quality, accurate RNA crystal structures available, and the community has thoroughly embraced that goal. However, due to the many degrees of freedom inherent in RNA structure (especially for the backbone), it is a significant challenge to succeed in building accurate experimental models for RNA structures. This chapter describes the tools and techniques our research group and our collaborators have developed over the years to help RNA structural biologists both evaluate and achieve better accuracy. Expert analysis of large, high-resolution, quality-conscious RNA datasets provides the fundamental information that enables automated methods for robust and efficient error diagnosis in validating RNA structures at all resolutions. The even more crucial goal of correcting the diagnosed outliers has steadily developed toward highly effective, computationally based techniques. Automation enables solving complex issues in large RNA structures, but cannot circumvent the need for thoughtful examination of local details, and so we also provide some guidance for interpreting and acting on the results of current structure validation for RNA.
Resumo:
Thin-layer and high-performance thin-layer chromatography (TLC/HPTLC) methods for assaying compound(s) in a sample must be validated to ensure that they are fit for their intended purpose and, where applicable, meet the strict regulatory requirements for controlled products. Two validation approaches are identified in the literature, i.e. the classic and the alternative, which is using accuracy profiles.Detailed procedures of the two approaches are discussed based on the validation of methods for pharmaceutical analysis, which is an area considered having more strict requirements. Estimation of the measurement uncertainty from the validation approach using accuracy profiles is also described.Examples of HPTLC methods, developed and validated to assay sulfamethoxazole and trimethoprim on the one hand and lamivudine, stavudine, and nevirapine on the other, in their fixed-dose combination tablets, are further elaborated.