956 resultados para Analytical methods validate


Relevância:

90.00% 90.00%

Publicador:

Resumo:

Microalgae have many applications, such as biodiesel production or food supplement. Depending on the application, the optimization of certain fractions of the biochemical composition (proteins, carbohydrates and lipids) is required. Therefore, samples obtained in different culture conditions must be analyzed in order to compare the content of such fractions. Nevertheless, traditional methods necessitate lengthy analytical procedures with prolonged sample turn-around times. Results of the biochemical composition of Nannochloropsis oculata samples with different protein, carbohydrate and lipid contents obtained by conventional analytical methods have been compared to those obtained by thermogravimetry (TGA) and a Pyroprobe device connected to a gas chromatograph with mass spectrometer detector (Py–GC/MS), showing a clear correlation. These results suggest a potential applicability of these techniques as fast and easy methods to qualitatively compare the biochemical composition of microalgal samples.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Le papier bioactif est obtenu par la modification de substrat du papier avec des biomolécules et des réactifs. Ce type de papier est utilisé dans le développement de nouveaux biocapteurs qui sont portables, jetables et économiques visant à capturer, détecter et dans certains cas, désactiver les agents pathogènes. Généralement les papiers bioactifs sont fabriqués par l’incorporation de biomolécules telles que les enzymes et les anticorps sur la surface du papier. L’immobilisation de ces biomolécules sur les surfaces solides est largement utilisée pour différentes applications de diagnostic comme dans immunocapteurs et immunoessais mais en raison de la nature sensible des enzymes, leur intégration au papier à grande échelle a rencontré plusieurs difficultés surtout dans les conditions industrielles. Pendant ce temps, les microcapsules sont une plate-forme intéressante pour l’immobilisation des enzymes et aussi assez efficace pour permettre à la fonctionnalisation du papier à grande échelle car le papier peut être facilement recouvert avec une couche de telles microcapsules. Dans cette étude, nous avons développé une plate-forme générique utilisant des microcapsules à base d’alginate qui peuvent être appliquées aux procédés usuels de production de papier bioactif et antibactérien avec la capacité de capturer des pathogènes à sa surface et de les désactiver grâce à la production d’un réactif anti-pathogène. La conception de cette plate-forme antibactérienne est basée sur la production constante de peroxyde d’hydrogène en tant qu’agent antibactérien à l’intérieur des microcapsules d’alginate. Cette production de peroxyde d’hydrogène est obtenue par oxydation du glucose catalysée par la glucose oxydase encapsulée à l’intérieur des billes d’alginate. Les différentes étapes de cette étude comprennent le piégeage de la glucose oxydase à l’intérieur des microcapsules d’alginate, l’activation et le renforcement de la surface des microcapsules par ajout d’une couche supplémentaire de chitosan, la vérification de la possibilité d’immobilisation des anticorps (immunoglobulines G humaine comme une modèle d’anticorps) sur la surface des microcapsules et enfin, l’évaluation des propriétés antibactériennes de cette plate-forme vis-à-vis l’Escherichia coli K-12 (E. coli K-12) en tant qu’un représentant des agents pathogènes. Après avoir effectué chaque étape, certaines mesures et observations ont été faites en utilisant diverses méthodes et techniques analytiques telles que la méthode de Bradford pour dosage des protéines, l’électroanalyse d’oxygène, la microscopie optique et confocale à balayage laser (CLSM), la spectrométrie de masse avec désorption laser assistée par matrice- temps de vol (MALDI-TOF-MS), etc. Les essais appropriés ont été effectués pour valider la réussite de modification des microcapsules et pour confirmer à ce fait que la glucose oxydase est toujours active après chaque étape de modification. L’activité enzymatique spécifique de la glucose oxydase après l’encapsulation a été évaluée à 120±30 U/g. Aussi, des efforts ont été faits pour immobiliser la glucose oxydase sur des nanoparticules d’or avec deux tailles différentes de diamètre (10,9 nm et 50 nm) afin d’améliorer l’activité enzymatique et augmenter l’efficacité d’encapsulation. Les résultats obtenus lors de cette étude démontrent les modifications réussies sur les microcapsules d’alginate et aussi une réponse favorable de cette plate-forme antibactérienne concernant la désactivation de E. coli K-12. La concentration efficace de l’activité enzymatique afin de désactivation de cet agent pathogénique modèle a été déterminée à 1.3×10-2 U/ml pour une concentration de 6.7×108 cellules/ml de bactéries. D’autres études sont nécessaires pour évaluer l’efficacité de l’anticorps immobilisé dans la désactivation des agents pathogènes et également intégrer la plate-forme sur le papier et valider l’efficacité du système une fois qu’il est déposé sur papier.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Le papier bioactif est obtenu par la modification de substrat du papier avec des biomolécules et des réactifs. Ce type de papier est utilisé dans le développement de nouveaux biocapteurs qui sont portables, jetables et économiques visant à capturer, détecter et dans certains cas, désactiver les agents pathogènes. Généralement les papiers bioactifs sont fabriqués par l’incorporation de biomolécules telles que les enzymes et les anticorps sur la surface du papier. L’immobilisation de ces biomolécules sur les surfaces solides est largement utilisée pour différentes applications de diagnostic comme dans immunocapteurs et immunoessais mais en raison de la nature sensible des enzymes, leur intégration au papier à grande échelle a rencontré plusieurs difficultés surtout dans les conditions industrielles. Pendant ce temps, les microcapsules sont une plate-forme intéressante pour l’immobilisation des enzymes et aussi assez efficace pour permettre à la fonctionnalisation du papier à grande échelle car le papier peut être facilement recouvert avec une couche de telles microcapsules. Dans cette étude, nous avons développé une plate-forme générique utilisant des microcapsules à base d’alginate qui peuvent être appliquées aux procédés usuels de production de papier bioactif et antibactérien avec la capacité de capturer des pathogènes à sa surface et de les désactiver grâce à la production d’un réactif anti-pathogène. La conception de cette plate-forme antibactérienne est basée sur la production constante de peroxyde d’hydrogène en tant qu’agent antibactérien à l’intérieur des microcapsules d’alginate. Cette production de peroxyde d’hydrogène est obtenue par oxydation du glucose catalysée par la glucose oxydase encapsulée à l’intérieur des billes d’alginate. Les différentes étapes de cette étude comprennent le piégeage de la glucose oxydase à l’intérieur des microcapsules d’alginate, l’activation et le renforcement de la surface des microcapsules par ajout d’une couche supplémentaire de chitosan, la vérification de la possibilité d’immobilisation des anticorps (immunoglobulines G humaine comme une modèle d’anticorps) sur la surface des microcapsules et enfin, l’évaluation des propriétés antibactériennes de cette plate-forme vis-à-vis l’Escherichia coli K-12 (E. coli K-12) en tant qu’un représentant des agents pathogènes. Après avoir effectué chaque étape, certaines mesures et observations ont été faites en utilisant diverses méthodes et techniques analytiques telles que la méthode de Bradford pour dosage des protéines, l’électroanalyse d’oxygène, la microscopie optique et confocale à balayage laser (CLSM), la spectrométrie de masse avec désorption laser assistée par matrice- temps de vol (MALDI-TOF-MS), etc. Les essais appropriés ont été effectués pour valider la réussite de modification des microcapsules et pour confirmer à ce fait que la glucose oxydase est toujours active après chaque étape de modification. L’activité enzymatique spécifique de la glucose oxydase après l’encapsulation a été évaluée à 120±30 U/g. Aussi, des efforts ont été faits pour immobiliser la glucose oxydase sur des nanoparticules d’or avec deux tailles différentes de diamètre (10,9 nm et 50 nm) afin d’améliorer l’activité enzymatique et augmenter l’efficacité d’encapsulation. Les résultats obtenus lors de cette étude démontrent les modifications réussies sur les microcapsules d’alginate et aussi une réponse favorable de cette plate-forme antibactérienne concernant la désactivation de E. coli K-12. La concentration efficace de l’activité enzymatique afin de désactivation de cet agent pathogénique modèle a été déterminée à 1.3×10-2 U/ml pour une concentration de 6.7×108 cellules/ml de bactéries. D’autres études sont nécessaires pour évaluer l’efficacité de l’anticorps immobilisé dans la désactivation des agents pathogènes et également intégrer la plate-forme sur le papier et valider l’efficacité du système une fois qu’il est déposé sur papier.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Elemental analysis can become an important piece of evidence to assist the solution of a case. The work presented in this dissertation aims to evaluate the evidential value of the elemental composition of three particular matrices: ink, paper and glass. In the first part of this study, the analytical performance of LIBS and LA-ICP-MS methods was evaluated for paper, writing inks and printing inks. A total of 350 ink specimens were examined including black and blue gel inks, ballpoint inks, inkjets and toners originating from several manufacturing sources and/or batches. The paper collection set consisted of over 200 paper specimens originating from 20 different paper sources produced by 10 different plants. Micro-homogeneity studies show smaller variation of elemental compositions within a single source (i.e., sheet, pen or cartridge) than the observed variation between different sources (i.e., brands, types, batches). Significant and detectable differences in the elemental profile of the inks and paper were observed between samples originating from different sources (discrimination of 87–100% of samples, depending on the sample set under investigation and the method applied). These results support the use of elemental analysis, using LA-ICP-MS and LIBS, for the examination of documents and provide additional discrimination to the currently used techniques in document examination. In the second part of this study, a direct comparison between four analytical methods (µ-XRF, solution-ICP-MS, LA-ICP-MS and LIBS) was conducted for glass analyses using interlaboratory studies. The data provided by 21 participants were used to assess the performance of the analytical methods in associating glass samples from the same source and differentiating different sources, as well as the use of different match criteria (confidence interval (±6s, ±5s, ±4s, ±3s, ±2s), modified confidence interval, t-test (sequential univariate, p=0.05 and p=0.01), t-test with Bonferroni correction (for multivariate comparisons), range overlap, and Hotelling's T2 tests. Error rates (Type 1 and Type 2) are reported for the use of each of these match criteria and depend on the heterogeneity of the glass sources, the repeatability between analytical measurements, and the number of elements that were measured. The study provided recommendations for analytical performance-based parameters for µ-XRF and LA-ICP-MS as well as the best performing match criteria for both analytical techniques, which can be applied now by forensic glass examiners.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

New designer drugs are constantly emerging onto the illicit drug market and it is often difficult to validate and maintaincomprehensive analytical methods for accurate detection of these compounds. Generally, toxicology laboratories utilize a screening method, such as immunoassay, for the presumptive identification of drugs of abuse. When a positive result occurs, confirmatory methods, such as gas chromatography (GC) or liquid chromatography (LC) coupled with mass spectrometry (MS), are required for more sensitive and specific analyses. In recent years, the need to study the activities of these compounds in screening assays as well as to develop confirmatory techniques to detect them in biological specimens has been recognized. Severe intoxications and fatalities have been encountered with emerging designer drugs, presenting analytical challenges for detection and identification of such novel compounds. The first major task of this research was to evaluate the performance of commercially available immunoassays to determine if designer drugs were cross-reactive. The second major task was to develop and validate a confirmatory method, using LC-MS, to identify and quantify these designer drugs in biological specimens.^ Cross-reactivity towards the cathinone derivatives was found to be minimal. Several other phenethylamines demonstrated cross-reactivity at low concentrations, but results were consistent with those published by the assay manufacturer or as reported in the literature. Current immunoassay-based screening methods may not be ideal for presumptively identifying most designer drugs, including the "bath salts." For this reason, an LC-MS based confirmatory method was developed for 32 compounds, including eight cathinone derivatives, with limits of quantification in the range of 1-10 ng/mL. The method was fully validated for selectivity, matrix effects, stability, recovery, precision, and accuracy. In order to compare the screening and confirmatory techniques, several human specimens were analyzed to demonstrate the importance of using a specific analytical method, such as LC-MS, to detect designer drugs in serum as immunoassays lack cross-reactivity with the novel compounds. Overall, minimal cross-reactivity was observed, highlighting the conclusion that these presumptive screens cannot detect many of the designer drugs and that a confirmatory technique, such as the LC-MS, is required for the comprehensive forensic toxicological analysis of designer drugs.^

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Harmful algal blooms (HABs) are a natural global phenomena emerging in severity and extent. Incidents have many economic, ecological and human health impacts. Monitoring and providing early warning of toxic HABs are critical for protecting public health. Current monitoring programmes include measuring the number of toxic phytoplankton cells in the water and biotoxin levels in shellfish tissue. As these efforts are demanding and labour intensive, methods which improve the efficiency are essential. This study compares the utilisation of a multitoxin surface plasmon resonance (multitoxin SPR) biosensor with enzyme-linked immunosorbent assay (ELISA) and analytical methods such as high performance liquid chromatography with fluorescence detection (HPLC-FLD) and liquid chromatography–tandem mass spectrometry (LC–MS/MS) for toxic HAB monitoring efforts in Europe. Seawater samples (n = 256) from European waters, collected 2009–2011, were analysed for biotoxins: saxitoxin and analogues, okadaic acid and dinophysistoxins 1/2 (DTX1/DTX2) and domoic acid responsible for paralytic shellfish poisoning (PSP), diarrheic shellfish poisoning (DSP) and amnesic shellfish poisoning (ASP), respectively. Biotoxins were detected mainly in samples from Spain and Ireland. France and Norway appeared to have the lowest number of toxic samples. Both the multitoxin SPR biosensor and the RNA microarray were more sensitive at detecting toxic HABs than standard light microscopy phytoplankton monitoring. Correlations between each of the detection methods were performed with the overall agreement, based on statistical 2 × 2 comparison tables, between each testing platform ranging between 32% and 74% for all three toxin families illustrating that one individual testing method may not be an ideal solution. An efficient early warning monitoring system for the detection of toxic HABs could therefore be achieved by combining both the multitoxin SPR biosensor and RNA microarray.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Harmful algal blooms (HABs) are a natural global phenomena emerging in severity and extent. Incidents have many economic, ecological and human health impacts. Monitoring and providing early warning of toxic HABs are critical for protecting public health. Current monitoring programmes include measuring the number of toxic phytoplankton cells in the water and biotoxin levels in shellfish tissue. As these efforts are demanding and labour intensive, methods which improve the efficiency are essential. This study compares the utilisation of a multitoxin surface plasmon resonance (multitoxin SPR) biosensor with enzyme-linked immunosorbent assay (ELISA) and analytical methods such as high performance liquid chromatography with fluorescence detection (HPLC-FLD) and liquid chromatography–tandem mass spectrometry (LC–MS/MS) for toxic HAB monitoring efforts in Europe. Seawater samples (n = 256) from European waters, collected 2009–2011, were analysed for biotoxins: saxitoxin and analogues, okadaic acid and dinophysistoxins 1/2 (DTX1/DTX2) and domoic acid responsible for paralytic shellfish poisoning (PSP), diarrheic shellfish poisoning (DSP) and amnesic shellfish poisoning (ASP), respectively. Biotoxins were detected mainly in samples from Spain and Ireland. France and Norway appeared to have the lowest number of toxic samples. Both the multitoxin SPR biosensor and the RNA microarray were more sensitive at detecting toxic HABs than standard light microscopy phytoplankton monitoring. Correlations between each of the detection methods were performed with the overall agreement, based on statistical 2 × 2 comparison tables, between each testing platform ranging between 32% and 74% for all three toxin families illustrating that one individual testing method may not be an ideal solution. An efficient early warning monitoring system for the detection of toxic HABs could therefore be achieved by combining both the multitoxin SPR biosensor and RNA microarray.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Major food adulteration and contamination events occur with alarming regularity and are known to be episodic, with the question being not if but when another large-scale food safety/integrity incident will occur. Indeed, the challenges of maintaining food security are now internationally recognised. The ever increasing scale and complexity of food supply networks can lead to them becoming significantly more vulnerable to fraud and contamination, and potentially dysfunctional. This can make the task of deciding which analytical methods are more suitable to collect and analyse (bio)chemical data within complex food supply chains, at targeted points of vulnerability, that much more challenging. It is evident that those working within and associated with the food industry are seeking rapid, user-friendly methods to detect food fraud and contamination, and rapid/high-throughput screening methods for the analysis of food in general. In addition to being robust and reproducible, these methods should be portable and ideally handheld and/or remote sensor devices, that can be taken to or be positioned on/at-line at points of vulnerability along complex food supply networks and require a minimum amount of background training to acquire information rich data rapidly (ergo point-and-shoot). Here we briefly discuss a range of spectrometry and spectroscopy based approaches, many of which are commercially available, as well as other methods currently under development. We discuss a future perspective of how this range of detection methods in the growing sensor portfolio, along with developments in computational and information sciences such as predictive computing and the Internet of Things, will together form systems- and technology-based approaches that significantly reduce the areas of vulnerability to food crime within food supply chains. As food fraud is a problem of systems and therefore requires systems level solutions and thinking.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Elemental analysis can become an important piece of evidence to assist the solution of a case. The work presented in this dissertation aims to evaluate the evidential value of the elemental composition of three particular matrices: ink, paper and glass. In the first part of this study, the analytical performance of LIBS and LA-ICP-MS methods was evaluated for paper, writing inks and printing inks. A total of 350 ink specimens were examined including black and blue gel inks, ballpoint inks, inkjets and toners originating from several manufacturing sources and/or batches. The paper collection set consisted of over 200 paper specimens originating from 20 different paper sources produced by 10 different plants. Micro-homogeneity studies show smaller variation of elemental compositions within a single source (i.e., sheet, pen or cartridge) than the observed variation between different sources (i.e., brands, types, batches). Significant and detectable differences in the elemental profile of the inks and paper were observed between samples originating from different sources (discrimination of 87 – 100% of samples, depending on the sample set under investigation and the method applied). These results support the use of elemental analysis, using LA-ICP-MS and LIBS, for the examination of documents and provide additional discrimination to the currently used techniques in document examination. In the second part of this study, a direct comparison between four analytical methods (µ-XRF, solution-ICP-MS, LA-ICP-MS and LIBS) was conducted for glass analyses using interlaboratory studies. The data provided by 21 participants were used to assess the performance of the analytical methods in associating glass samples from the same source and differentiating different sources, as well as the use of different match criteria (confidence interval (±6s, ±5s, ±4s, ±3s, ±2s), modified confidence interval, t-test (sequential univariate, p=0.05 and p=0.01), t-test with Bonferroni correction (for multivariate comparisons), range overlap, and Hotelling’s T2 tests. Error rates (Type 1 and Type 2) are reported for the use of each of these match criteria and depend on the heterogeneity of the glass sources, the repeatability between analytical measurements, and the number of elements that were measured. The study provided recommendations for analytical performance-based parameters for µ-XRF and LA-ICP-MS as well as the best performing match criteria for both analytical techniques, which can be applied now by forensic glass examiners.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Espécies forrageiras adaptadas às condições semiáridas são uma alternativa para reduzir os impactos negativos na cadeia produtiva de ruminantes da região Nordeste brasileira devido à sazonalidade na oferta de forragem, além de reduzir custo com o fornecimento de alimentos concentrados. Dentre as espécies, a vagem de algaroba (Prosopis juliflora SW D.C.) e palma forrageira (Opuntia e Nopalea) ganham destaque por tolerarem o déficit hídrico e produzirem em períodos onde a oferta de forragem está reduzida, além de apresentam bom valor nutricional e serem bem aceitas pelos animais. Porém, devido à variação na sua composição, seu uso na alimentação animal exige o conhecimento profundo da sua composição para a elaboração de dietas balanceadas. No entanto, devido ao custo e tempo para análise, os produtores não fazem uso da prática de análise da composição químico-bromatológica dos alimentos. Por isto, a espectroscopia de reflectância no infravermelho próximo (NIRS) representa uma importante alternativa aos métodos tradicionais. Objetivou-se com este estudo desenvolver e validar modelos de predição da composição bromatológica de vagem de algaroba e palma forrageira baseados em espectroscopia NIRS, escaneadas em dois modelos de equipamentos e com diferentes processamentos da amostra. Foram coletadas amostras de vagem de algaroba nos estados do Ceará, Bahia, Paraíba e Pernambuco, e amostras de palma forrageira nos estados do Ceará, Paraíba e Pernambuco, frescas (in natura) ou pré-secas e moídas. Para obtenção dos espectros utilizaram-se dois equipamentos NIR, Perten DA 7250 e FOSS 5000. Inicialmente os alimentos foram escaneados in natura em aparelho do modelo Perten, e, com o auxílio do software The Unscrambler 10.2 foi selecionado um grupo de amostras para o banco de calibração. As amostras selecionadas foram secas e moídas, e escaneadas novamente em equipamentos Perten e FOSS. Os valores dos parâmetros de referência foram obtidos por meio de metodologias tradicionalmente aplicadas em laboratório de nutrição animal para matéria seca (MS), matéria mineral (MM), matéria orgânica (MO), proteína bruta (PB), estrato etéreo (EE), fibra solúvel em detergente neutro (FDN), fibra solúvel em detergente ácido (FDA), hemicelulose (HEM) e digestibilidade in vitro da matéria seca (DIVMS). O desempenho dos modelos foi avaliado de acordo com os erros médios de calibração (RMSEC) e validação (RMSECV), coeficiente de determinação (R2 ) e da relação de desempenho de desvio dos modelos (RPD). A análise exploratória dos dados, por meio de tratamentos espectrais e análise de componentes principais (PCA), demonstraram que os bancos de dados eram similares entre si, dando segurança de desenvolver os modelos com todas as amostras selecionadas em um único modelo para cada alimento, algaroba e palma. Na avaliação dos resultados de referência, observou-se que a variação dos resultados para cada parâmetro corroboraram com os descritos na literatura. No desempenho dos modelos, aqueles desenvolvidos com pré-processamento da amostra (pré-secagem e moagem) se mostraram mais robustos do que aqueles construídos com amostras in natura. O aparelho NIRS Perten apresentou desempenho semelhante ao equipamento FOSS, apesar desse último cobrir uma faixa espectral maior e com intervalos de leituras menores. A técnica NIR, associada ao método de calibração multivariada de regressão por meio de quadrados mínimos (PLS), mostrou-se confiável para prever a composição químico-bromatológica de vagem de algaroba e da palma forrageira. Abstract: Forage species adapted to semi-arid conditions are an alternative to reduce the negative impacts in the feed supply for ruminants in the Brazilian Northeast region, due to seasonality in forage availability, as well as in the reducing of cost by providing concentrated feedstuffs. Among the species, mesquite pods (Prosopis juliflora SW DC) and spineless cactus (Opuntia and Nopalea) are highlighted for tolerating the drought and producion in periods where the forage is scarce, and have high nutritional value and also are well accepted by the animals. However, its use in animal diets requires a knowledge about its composition to prepare balanced diets. However, farmers usually do not use feed composition analysis, because their high cost and time-consuming. Thus, the Near Infrared Reflectance Spectroscopy in the (NIRS) is an important alternative to traditional methods. The objective of this study to develop and validate predictive models of the chemical composition of mesquite pods and spineless cactus-based NIRS spectroscopy, scanned in two different spectrometers and sample processing. Mesquite pods samples were collected in the states of Ceará, Bahia, Paraiba and Pernambuco, and samples of forage cactus in the states of Ceará, Paraíba and Pernambuco. In order to obtain the spectra, it was used two NIR equipment: Perten DA 7250 and FOSS 5000. sSpectra of samples were initially obtained fresh (as received) using Perten instrument, and with The Unscrambler software 10.2, a group of subsamples was selected to model development, keeping out redundant ones. The selected samples were dried and ground, and scanned again in both Perten and FOSS instruments. The values of the reference analysis were obtained by methods traditionally applied in animal nutrition laboratory to dry matter (DM), mineral matter (MM), organic matter (OM), crude protein (CP), ether extract (EE), soluble neutral detergent fiber (NDF), soluble acid detergent fiber (ADF), hemicellulose ( HEM) and in vitro digestibility of dry matter (DIVDM). The performance of the models was evaluated according to the Root Mean Square Error of Calibration (RMSEC) and cross-validation (RMSECV), coefficient of determination (R2 ) and the deviation of Ratio of performance Deviation of the models (RPD). Exploratory data analysis through spectral treatments and principal component analysis (PCA), showed that the databases were similar to each other, and may be treated asa single model for each feed - mesquite pods and cactus. Evaluating the reference results, it was observed that the variation were similar to those reported in the literature. Comparing the preprocessing of samples, the performance ofthose developed with preprocessing (dried and ground) of the sample were more robust than those built with fresh samples. The NIRS Perten device performance similar to FOSS equipment, although the latter cover a larger spectral range and with lower readings intervals. NIR technology associate do multivariate techniques is reliable to predict the bromatological composition of mesquite pods and cactus.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Scott A. Shane is the 2009 winner of the Global Award for Entrepreneurship Research. In this article we discuss and analyze Shane’s most important contributions to the field of entrepreneurship. His contribution is extraordinarily broad in scope, which makes it difficult to pinpoint one or a few specifics that we associate with Shane’s scholarship. Instead, they can be summarized in the following three points. First, he has influenced what we view as central aspects of entrepreneurship. Shane has been a leading figure in redirecting the focus on entrepreneurship research itself. Second, he has influenced how we view entrepreneurship. Shane’s research is arguably theory driven and it applies and develops theoretical lenses that greatly improve our understanding of entrepreneurship. Third, he has contributed to how we conduct entrepreneurship research. Shane has been a forerunner in examining relevant units of analysis that are difficult to sample; research designs and databases specifically designed for studying entrepreneurial processes; and sophisticated analytical methods. This has contributed to advancing the methodological rigor of the field. Summing them up, the contributions are very impressive indeed.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Effective focusing of electromagnetic (EM) energy to nanoscale regions is one of the major challenges in nano-photonics and plasmonics. The strong localization of the optical energy into regions much smaller than allowed by the diffraction limit, also called nanofocusing, offers promising applications in nano-sensor technology, nanofabrication, near-field optics or spectroscopy. One of the most promising solutions to the problem of efficient nanofocusing is related to surface plasmon propagation in metallic structures. Metallic tapered rods, commonly used as probes in near field microscopy and spectroscopy, are of a particular interest. They can provide very strong EM field enhancement at the tip due to surface plasmons (SP’s) propagating towards the tip of the tapered metal rod. A large number of studies have been devoted to the manufacturing process of tapered rods or tapered fibers coated by a metal film. On the other hand, structures such as metallic V-grooves or metal wedges can also provide strong electric field enhancements but manufacturing of these structures is still a challenge. It has been shown, however, that the attainable electric field enhancement at the apex in the V-groove is higher than at the tip of a metal tapered rod when the dissipation level in the metal is strong. Metallic V-grooves also have very promising characteristics as plasmonic waveguides. This thesis will present a thorough theoretical and numerical investigation of nanofocusing during plasmon propagation along a metal tapered rod and into a metallic V-groove. Optimal structural parameters including optimal taper angle, taper length and shape of the taper are determined in order to achieve maximum field enhancement factors at the tip of the nanofocusing structure. An analytical investigation of plasmon nanofocusing by metal tapered rods is carried out by means of the geometric optics approximation (GOA), which is also called adiabatic nanofocusing. However, GOA is applicable only for analysing tapered structures with small taper angles and without considering a terminating tip structure in order to neglect reflections. Rigorous numerical methods are employed for analysing non-adiabatic nanofocusing, by tapered rod and V-grooves with larger taper angles and with a rounded tip. These structures cannot be studied by analytical methods due to the presence of reflected waves from the taper section, the tip and also from (artificial) computational boundaries. A new method is introduced to combine the advantages of GOA and rigorous numerical methods in order to reduce significantly the use of computational resources and yet achieve accurate results for the analysis of large tapered structures, within reasonable calculation time. Detailed comparison between GOA and rigorous numerical methods will be carried out in order to find the critical taper angle of the tapered structures at which GOA is still applicable. It will be demonstrated that optimal taper angles, at which maximum field enhancements occur, coincide with the critical angles, at which GOA is still applicable. It will be shown that the applicability of GOA can be substantially expanded to include structures which could be analysed previously by numerical methods only. The influence of the rounded tip, the taper angle and the role of dissipation onto the plasmon field distribution along the tapered rod and near the tip will be analysed analytically and numerically in detail. It will be demonstrated that electric field enhancement factors of up to ~ 2500 within nanoscale regions are predicted. These are sufficient, for instance, to detect single molecules using surface enhanced Raman spectroscopy (SERS) with the tip of a tapered rod, an approach also known as tip enhanced Raman spectroscopy or TERS. The results obtained in this project will be important for applications for which strong local field enhancement factors are crucial for the performance of devices such as near field microscopes or spectroscopy. The optimal design of nanofocusing structures, at which the delivery of electromagnetic energy to the nanometer region is most efficient, will lead to new applications in near field sensors, near field measuring technology, or generation of nanometer sized energy sources. This includes: applications in tip enhanced Raman spectroscopy (TERS); manipulation of nanoparticles and molecules; efficient coupling of optical energy into and out of plasmonic circuits; second harmonic generation in non-linear optics; or delivery of energy to quantum dots, for instance, for quantum computations.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

A national-level safety analysis tool is needed to complement existing analytical tools for assessment of the safety impacts of roadway design alternatives. FHWA has sponsored the development of the Interactive Highway Safety Design Model (IHSDM), which is roadway design and redesign software that estimates the safety effects of alternative designs. Considering the importance of IHSDM in shaping the future of safety-related transportation investment decisions, FHWA justifiably sponsored research with the sole intent of independently validating some of the statistical models and algorithms in IHSDM. Statistical model validation aims to accomplish many important tasks, including (a) assessment of the logical defensibility of proposed models, (b) assessment of the transferability of models over future time periods and across different geographic locations, and (c) identification of areas in which future model improvements should be made. These three activities are reported for five proposed types of rural intersection crash prediction models. The internal validation of the model revealed that the crash models potentially suffer from omitted variables that affect safety, site selection and countermeasure selection bias, poorly measured and surrogate variables, and misspecification of model functional forms. The external validation indicated the inability of models to perform on par with model estimation performance. Recommendations for improving the state of the practice from this research include the systematic conduct of carefully designed before-and-after studies, improvements in data standardization and collection practices, and the development of analytical methods to combine the results of before-and-after studies with cross-sectional studies in a meaningful and useful way.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Advances in safety research—trying to improve the collective understanding of motor vehicle crash causation—rests upon the pursuit of numerous lines of inquiry. The research community has focused on analytical methods development (negative binomial specifications, simultaneous equations, etc.), on better experimental designs (before-after studies, comparison sites, etc.), on improving exposure measures, and on model specification improvements (additive terms, non-linear relations, etc.). One might think of different lines of inquiry in terms of ‘low lying fruit’—areas of inquiry that might provide significant improvements in understanding crash causation. It is the contention of this research that omitted variable bias caused by the exclusion of important variables is an important line of inquiry in safety research. In particular, spatially related variables are often difficult to collect and omitted from crash models—but offer significant ability to better understand contributing factors to crashes. This study—believed to represent a unique contribution to the safety literature—develops and examines the role of a sizeable set of spatial variables in intersection crash occurrence. In addition to commonly considered traffic and geometric variables, examined spatial factors include local influences of weather, sun glare, proximity to drinking establishments, and proximity to schools. The results indicate that inclusion of these factors results in significant improvement in model explanatory power, and the results also generally agree with expectation. The research illuminates the importance of spatial variables in safety research and also the negative consequences of their omissions.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Impedance cardiography is an application of bioimpedance analysis primarily used in a research setting to determine cardiac output. It is a non invasive technique that measures the change in the impedance of the thorax which is attributed to the ejection of a volume of blood from the heart. The cardiac output is calculated from the measured impedance using the parallel conductor theory and a constant value for the resistivity of blood. However, the resistivity of blood has been shown to be velocity dependent due to changes in the orientation of red blood cells induced by changing shear forces during flow. The overall goal of this thesis was to study the effect that flow deviations have on the electrical impedance of blood, both experimentally and theoretically, and to apply the results to a clinical setting. The resistivity of stationary blood is isotropic as the red blood cells are randomly orientated due to Brownian motion. In the case of blood flowing through rigid tubes, the resistivity is anisotropic due to the biconcave discoidal shape and orientation of the cells. The generation of shear forces across the width of the tube during flow causes the cells to align with the minimal cross sectional area facing the direction of flow. This is in order to minimise the shear stress experienced by the cells. This in turn results in a larger cross sectional area of plasma and a reduction in the resistivity of the blood as the flow increases. Understanding the contribution of this effect on the thoracic impedance change is a vital step in achieving clinical acceptance of impedance cardiography. Published literature investigates the resistivity variations for constant blood flow. In this case, the shear forces are constant and the impedance remains constant during flow at a magnitude which is less than that for stationary blood. The research presented in this thesis, however, investigates the variations in resistivity of blood during pulsataile flow through rigid tubes and the relationship between impedance, velocity and acceleration. Using rigid tubes isolates the impedance change to variations associated with changes in cell orientation only. The implications of red blood cell orientation changes for clinical impedance cardiography were also explored. This was achieved through measurement and analysis of the experimental impedance of pulsatile blood flowing through rigid tubes in a mock circulatory system. A novel theoretical model including cell orientation dynamics was developed for the impedance of pulsatile blood through rigid tubes. The impedance of flowing blood was theoretically calculated using analytical methods for flow through straight tubes and the numerical Lattice Boltzmann method for flow through complex geometries such as aortic valve stenosis. The result of the analytical theoretical model was compared to the experimental impedance measurements through rigid tubes. The impedance calculated for flow through a stenosis using the Lattice Boltzmann method provides results for comparison with impedance cardiography measurements collected as part of a pilot clinical trial to assess the suitability of using bioimpedance techniques to assess the presence of aortic stenosis. The experimental and theoretical impedance of blood was shown to inversely follow the blood velocity during pulsatile flow with a correlation of -0.72 and -0.74 respectively. The results for both the experimental and theoretical investigations demonstrate that the acceleration of the blood is an important factor in determining the impedance, in addition to the velocity. During acceleration, the relationship between impedance and velocity is linear (r2 = 0.98, experimental and r2 = 0.94, theoretical). The relationship between the impedance and velocity during the deceleration phase is characterised by a time decay constant, ô , ranging from 10 to 50 s. The high level of agreement between the experimental and theoretically modelled impedance demonstrates the accuracy of the model developed here. An increase in the haematocrit of the blood resulted in an increase in the magnitude of the impedance change due to changes in the orientation of red blood cells. The time decay constant was shown to decrease linearly with the haematocrit for both experimental and theoretical results, although the slope of this decrease was larger in the experimental case. The radius of the tube influences the experimental and theoretical impedance given the same velocity of flow. However, when the velocity was divided by the radius of the tube (labelled the reduced average velocity) the impedance response was the same for two experimental tubes with equivalent reduced average velocity but with different radii. The temperature of the blood was also shown to affect the impedance with the impedance decreasing as the temperature increased. These results are the first published for the impedance of pulsatile blood. The experimental impedance change measured orthogonal to the direction of flow is in the opposite direction to that measured in the direction of flow. These results indicate that the impedance of blood flowing through rigid cylindrical tubes is axisymmetric along the radius. This has not previously been verified experimentally. Time frequency analysis of the experimental results demonstrated that the measured impedance contains the same frequency components occuring at the same time point in the cycle as the velocity signal contains. This suggests that the impedance contains many of the fluctuations of the velocity signal. Application of a theoretical steady flow model to pulsatile flow presented here has verified that the steady flow model is not adequate in calculating the impedance of pulsatile blood flow. The success of the new theoretical model over the steady flow model demonstrates that the velocity profile is important in determining the impedance of pulsatile blood. The clinical application of the impedance of blood flow through a stenosis was theoretically modelled using the Lattice Boltzman method (LBM) for fluid flow through complex geometeries. The impedance of blood exiting a narrow orifice was calculated for varying degrees of stenosis. Clincial impedance cardiography measurements were also recorded for both aortic valvular stenosis patients (n = 4) and control subjects (n = 4) with structurally normal hearts. This pilot trial was used to corroborate the results of the LBM. Results from both investigations showed that the decay time constant for impedance has potential in the assessment of aortic valve stenosis. In the theoretically modelled case (LBM results), the decay time constant increased with an increase in the degree of stenosis. The clinical results also showed a statistically significant difference in time decay constant between control and test subjects (P = 0.03). The time decay constant calculated for test subjects (ô = 180 - 250 s) is consistently larger than that determined for control subjects (ô = 50 - 130 s). This difference is thought to be due to difference in the orientation response of the cells as blood flows through the stenosis. Such a non-invasive technique using the time decay constant for screening of aortic stenosis provides additional information to that currently given by impedance cardiography techniques and improves the value of the device to practitioners. However, the results still need to be verified in a larger study. While impedance cardiography has not been widely adopted clinically, it is research such as this that will enable future acceptance of the method.