34 resultados para Deterministic walkers

em Université de Lausanne, Switzerland


Relevância:

20.00% 20.00%

Publicador:

Resumo:

General introductionThe Human Immunodeficiency/Acquired Immunodeficiency Syndrome (HIV/AIDS) epidemic, despite recent encouraging announcements by the World Health Organization (WHO) is still today one of the world's major health care challenges.The present work lies in the field of health care management, in particular, we aim to evaluate the behavioural and non-behavioural interventions against HIV/AIDS in developing countries through a deterministic simulation model, both in human and economic terms. We will focus on assessing the effectiveness of the antiretroviral therapies (ART) in heterosexual populations living in lesser developed countries where the epidemic has generalized (formerly defined by the WHO as type II countries). The model is calibrated using Botswana as a case study, however our model can be adapted to other countries with similar transmission dynamics.The first part of this thesis consists of reviewing the main mathematical concepts describing the transmission of infectious agents in general but with a focus on human immunodeficiency virus (HIV) transmission. We also review deterministic models assessing HIV interventions with a focus on models aimed at African countries. This review helps us to recognize the need for a generic model and allows us to define a typical structure of such a generic deterministic model.The second part describes the main feed-back loops underlying the dynamics of HIV transmission. These loops represent the foundation of our model. This part also provides a detailed description of the model, including the various infected and non-infected population groups, the type of sexual relationships, the infection matrices, important factors impacting HIV transmission such as condom use, other sexually transmitted diseases (STD) and male circumcision. We also included in the model a dynamic life expectancy calculator which, to our knowledge, is a unique feature allowing more realistic cost-efficiency calculations. Various intervention scenarios are evaluated using the model, each of them including ART in combination with other interventions, namely: circumcision, campaigns aimed at behavioral change (Abstain, Be faithful or use Condoms also named ABC campaigns), and treatment of other STD. A cost efficiency analysis (CEA) is performed for each scenario. The CEA consists of measuring the cost per disability-adjusted life year (DALY) averted. This part also describes the model calibration and validation, including a sensitivity analysis.The third part reports the results and discusses the model limitations. In particular, we argue that the combination of ART and ABC campaigns and ART and treatment of other STDs are the most cost-efficient interventions through 2020. The main model limitations include modeling the complexity of sexual relationships, omission of international migration and ignoring variability in infectiousness according to the AIDS stage.The fourth part reviews the major contributions of the thesis and discusses model generalizability and flexibility. Finally, we conclude that by selecting the adequate interventions mix, policy makers can significantly reduce the adult prevalence in Botswana in the coming twenty years providing the country and its donors can bear the cost involved.Part I: Context and literature reviewIn this section, after a brief introduction to the general literature we focus in section two on the key mathematical concepts describing the transmission of infectious agents in general with a focus on HIV transmission. Section three provides a description of HIV policy models, with a focus on deterministic models. This leads us in section four to envision the need for a generic deterministic HIV policy model and briefly describe the structure of such a generic model applicable to countries with generalized HIV/AIDS epidemic, also defined as pattern II countries by the WHO.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The study investigates the possibility to incorporate fracture intensity and block geometry as spatially continuous parameters in GIS-based systems. For this purpose, a deterministic method has been implemented to estimate block size (Bloc3D) and joint frequency (COLTOP). In addition to measuring the block size, the Bloc3D Method provides a 3D representation of the shape of individual blocks. These two methods were applied using field measurements (joint set orientation and spacing) performed over a large field area, in the Swiss Alps. This area is characterized by a complex geology, a number of different rock masses and varying degrees of metamorphism. The spatial variability of the parameters was evaluated with regard to lithology and major faults. A model incorporating these measurements and observations into a GIS system to assess the risk associated with rock falls is proposed. The analysis concludes with a discussion on the feasibility of such an application in regularly and irregularly jointed rock masses, with persistent and impersistent discontinuities.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

There are far-reaching conceptual similarities between bi-static surface georadar and post-stack, "zero-offset" seismic reflection data, which is expressed in largely identical processing flows. One important difference is, however, that standard deconvolution algorithms routinely used to enhance the vertical resolution of seismic data are notoriously problematic or even detrimental to the overall signal quality when applied to surface georadar data. We have explored various options for alleviating this problem and have tested them on a geologically well-constrained surface georadar dataset. Standard stochastic and direct deterministic deconvolution approaches proved to be largely unsatisfactory. While least-squares-type deterministic deconvolution showed some promise, the inherent uncertainties involved in estimating the source wavelet introduced some artificial "ringiness". In contrast, we found spectral balancing approaches to be effective, practical and robust means for enhancing the vertical resolution of surface georadar data, particularly, but not exclusively, in the uppermost part of the georadar section, which is notoriously plagued by the interference of the direct air- and groundwaves. For the data considered in this study, it can be argued that band-limited spectral blueing may provide somewhat better results than standard band-limited spectral whitening, particularly in the uppermost part of the section affected by the interference of the air- and groundwaves. Interestingly, this finding is consistent with the fact that the amplitude spectrum resulting from least-squares-type deterministic deconvolution is characterized by a systematic enhancement of higher frequencies at the expense of lower frequencies and hence is blue rather than white. It is also consistent with increasing evidence that spectral "blueness" is a seemingly universal, albeit enigmatic, property of the distribution of reflection coefficients in the Earth. Our results therefore indicate that spectral balancing techniques in general and spectral blueing in particular represent simple, yet effective means of enhancing the vertical resolution of surface georadar data and, in many cases, could turn out to be a preferable alternative to standard deconvolution approaches.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

AbstractBreast cancer is one of the most common cancers affecting one in eight women during their lives. Survival rates have increased steadily thanks to early diagnosis with mammography screening and more efficient treatment strategies. Post-operative radiation therapy is a standard of care in the management of breast cancer and has been shown to reduce efficiently both local recurrence rate and breast cancer mortality. Radiation therapy is however associated with some late effects for long-term survivors. Radiation-induced secondary cancer is a relatively rare but severe late effect of radiation therapy. Currently, radiotherapy plans are essentially optimized to maximize tumor control and minimize late deterministic effects (tissue reactions) that are mainly associated with high doses (» 1 Gy). With improved cure rates and new radiation therapy technologies, it is also important to evaluate and minimize secondary cancer risks for different treatment techniques. This is a particularly challenging task due to the large uncertainties in the dose-response relationship.In contrast with late deterministic effects, secondary cancers may be associated with much lower doses and therefore out-of-field doses (also called peripheral doses) that are typically inferior to 1 Gy need to be determined accurately. Out-of-field doses result from patient scatter and head scatter from the treatment unit. These doses are particularly challenging to compute and we characterized it by Monte Carlo (MC) calculation. A detailed MC model of the Siemens Primus linear accelerator has been thoroughly validated with measurements. We investigated the accuracy of such a model for retrospective dosimetry in epidemiological studies on secondary cancers. Considering that patients in such large studies could be treated on a variety of machines, we assessed the uncertainty in reconstructed peripheral dose due to the variability of peripheral dose among various linac geometries. For large open fields (> 10x10 cm2), the uncertainty would be less than 50%, but for small fields and wedged fields the uncertainty in reconstructed dose could rise up to a factor of 10. It was concluded that such a model could be used for conventional treatments using large open fields only.The MC model of the Siemens Primus linac was then used to compare out-of-field doses for different treatment techniques in a female whole-body CT-based phantom. Current techniques such as conformai wedged-based radiotherapy and hybrid IMRT were investigated and compared to older two-dimensional radiotherapy techniques. MC doses were also compared to those of a commercial Treatment Planning System (TPS). While the TPS is routinely used to determine the dose to the contralateral breast and the ipsilateral lung which are mostly out of the treatment fields, we have shown that these doses may be highly inaccurate depending on the treatment technique investigated. MC shows that hybrid IMRT is dosimetrically similar to three-dimensional wedge-based radiotherapy within the field, but offers substantially reduced doses to out-of-field healthy organs.Finally, many different approaches to risk estimations extracted from the literature were applied to the calculated MC dose distribution. Absolute risks varied substantially as did the ratio of risk between two treatment techniques, reflecting the large uncertainties involved with current risk models. Despite all these uncertainties, the hybrid IMRT investigated resulted in systematically lower cancer risks than any of the other treatment techniques. More epidemiological studies with accurate dosimetry are required in the future to construct robust risk models. In the meantime, any treatment strategy that reduces out-of-field doses to healthy organs should be investigated. Electron radiotherapy might offer interesting possibilities with this regard.RésuméLe cancer du sein affecte une femme sur huit au cours de sa vie. Grâce au dépistage précoce et à des thérapies de plus en plus efficaces, le taux de guérison a augmenté au cours du temps. La radiothérapie postopératoire joue un rôle important dans le traitement du cancer du sein en réduisant le taux de récidive et la mortalité. Malheureusement, la radiothérapie peut aussi induire des toxicités tardives chez les patients guéris. En particulier, les cancers secondaires radio-induits sont une complication rare mais sévère de la radiothérapie. En routine clinique, les plans de radiothérapie sont essentiellement optimisées pour un contrôle local le plus élevé possible tout en minimisant les réactions tissulaires tardives qui sont essentiellement associées avec des hautes doses (» 1 Gy). Toutefois, avec l'introduction de différentes nouvelles techniques et avec l'augmentation des taux de survie, il devient impératif d'évaluer et de minimiser les risques de cancer secondaire pour différentes techniques de traitement. Une telle évaluation du risque est une tâche ardue étant donné les nombreuses incertitudes liées à la relation dose-risque.Contrairement aux effets tissulaires, les cancers secondaires peuvent aussi être induits par des basses doses dans des organes qui se trouvent hors des champs d'irradiation. Ces organes reçoivent des doses périphériques typiquement inférieures à 1 Gy qui résultent du diffusé du patient et du diffusé de l'accélérateur. Ces doses sont difficiles à calculer précisément, mais les algorithmes Monte Carlo (MC) permettent de les estimer avec une bonne précision. Un modèle MC détaillé de l'accélérateur Primus de Siemens a été élaboré et validé avec des mesures. La précision de ce modèle a également été déterminée pour la reconstruction de dose en épidémiologie. Si on considère que les patients inclus dans de larges cohortes sont traités sur une variété de machines, l'incertitude dans la reconstruction de dose périphérique a été étudiée en fonction de la variabilité de la dose périphérique pour différents types d'accélérateurs. Pour de grands champs (> 10x10 cm ), l'incertitude est inférieure à 50%, mais pour de petits champs et des champs filtrés, l'incertitude de la dose peut monter jusqu'à un facteur 10. En conclusion, un tel modèle ne peut être utilisé que pour les traitements conventionnels utilisant des grands champs.Le modèle MC de l'accélérateur Primus a été utilisé ensuite pour déterminer la dose périphérique pour différentes techniques dans un fantôme corps entier basé sur des coupes CT d'une patiente. Les techniques actuelles utilisant des champs filtrés ou encore l'IMRT hybride ont été étudiées et comparées par rapport aux techniques plus anciennes. Les doses calculées par MC ont été comparées à celles obtenues d'un logiciel de planification commercial (TPS). Alors que le TPS est utilisé en routine pour déterminer la dose au sein contralatéral et au poumon ipsilatéral qui sont principalement hors des faisceaux, nous avons montré que ces doses peuvent être plus ou moins précises selon la technTque étudiée. Les calculs MC montrent que la technique IMRT est dosimétriquement équivalente à celle basée sur des champs filtrés à l'intérieur des champs de traitement, mais offre une réduction importante de la dose aux organes périphériques.Finalement différents modèles de risque ont été étudiés sur la base des distributions de dose calculées par MC. Les risques absolus et le rapport des risques entre deux techniques de traitement varient grandement, ce qui reflète les grandes incertitudes liées aux différents modèles de risque. Malgré ces incertitudes, on a pu montrer que la technique IMRT offrait une réduction du risque systématique par rapport aux autres techniques. En attendant des données épidémiologiques supplémentaires sur la relation dose-risque, toute technique offrant une réduction des doses périphériques aux organes sains mérite d'être étudiée. La radiothérapie avec des électrons offre à ce titre des possibilités intéressantes.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The 22q11.2 deletion syndrome (22q11DS) is a widely recognized genetic model allowing the study of neuroanatomical biomarkers that underlie the risk for developing schizophrenia. Recent advances in magnetic resonance image analyses enable the examination of structural connectivity integrity, scarcely used in the 22q11DS field. This framework potentially provides evidence for the disconnectivity hypothesis of schizophrenia in this high-risk population. In the present study, we quantify the whole brain white matter connections in 22q11DS using deterministic tractography. Diffusion Tensor Imaging was acquired in 30 affected patients and 30 age- and gender-matched healthy participants. The Human Connectome technique was applied to register white matter streamlines with cortical anatomy. The number of fibers (streamlines) was used as a measure of connectivity for comparison between groups at the global, lobar and regional level. All statistics were corrected for age and gender. Results showed a 10% reduction of the total number of fibers in patients compared to controls. After correcting for this global reduction, preserved connectivity was found within the right frontal and right parietal lobes. The relative increase in the number of fibers was located mainly in the right hemisphere. Conversely, an excessive reduction of connectivity was observed within and between limbic structures. Finally, a disproportionate reduction was shown at the level of fibers connecting the left fronto-temporal regions. We could therefore speculate that the observed disruption to fronto-temporal connectivity in individuals at risk of schizophrenia implies that fronto-temporal disconnectivity, frequently implicated in the pathogenesis of schizophrenia, could precede the onset of symptoms and, as such, constitutes a biomarker of the vulnerability to develop psychosis. On the contrary, connectivity alterations in the limbic lobe play a role in a wide range of psychiatric disorders and therefore seem to be less specific in defining schizophrenia.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Résumé : Depuis la fin de la perestroïka s'est mis en place en Russie un discours identitaire qui, en linguistique, prend des formes extrêmes, reposant sur un strict déterminisme de la pensée par la langue. Les organismes de financement de la recherche scientifique soutiennent des projets qui étudient le rapport entre la grammaire russe et le « caractère national russe ». Des objets nouveaux apparaissent : « l'image linguistique russe du monde », « la personnalité linguistique », la « linguoculturologie ». Cet ensemble discursif construit dans l'imaginaire une identité collective rassurante, reposant sur l'idée que 1) tous les gens qui parlent la même langue pensent de la même façon; 2) les langues, donc les pensées collectives, sont imperméables entre elles, et donc intraduisibles. Cette tendance néo-humboldtienne dans la linguistique russe actuelle se déploie en toute méconnaissance de ses origines historiques : le Romantisme allemand dans son opposition à la philosophie des Lumières, le positivisme évolutionnisme d'Auguste Comte et la linguistique déterministe de l'Allemagne des années 1930.AbstractSince the end of perestroika, in linguistics in Russia, a new form of discourse has taken place, which stresses a very tight determinism of thought by language. The funding organizations of scientific research back up projects studying the relationship between Russiangrammar and the « Russian national character ». New objects of knowledge come to light : « the Russian linguistic image of the world », « linguistic personnality », « culturology ». This kind of discourse builds up an imaginary comforting collective identity, which relies on the principle that 1) all the people who speak the same language think the same way; 2) languages, hence collective kinds of thought, are hermetically closed to each other, and untranslatable. This neo-humboldtian trend in contemporary Russian linguistics has no knowledge of its historical origins : German Romanticism in its Anti-Enlightenment trend, evolutionnist positivism of Auguste Comte, and deterministic linguistics in Germany in the 1930s.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

MOTIVATION: Understanding gene regulation in biological processes and modeling the robustness of underlying regulatory networks is an important problem that is currently being addressed by computational systems biologists. Lately, there has been a renewed interest in Boolean modeling techniques for gene regulatory networks (GRNs). However, due to their deterministic nature, it is often difficult to identify whether these modeling approaches are robust to the addition of stochastic noise that is widespread in gene regulatory processes. Stochasticity in Boolean models of GRNs has been addressed relatively sparingly in the past, mainly by flipping the expression of genes between different expression levels with a predefined probability. This stochasticity in nodes (SIN) model leads to over representation of noise in GRNs and hence non-correspondence with biological observations. RESULTS: In this article, we introduce the stochasticity in functions (SIF) model for simulating stochasticity in Boolean models of GRNs. By providing biological motivation behind the use of the SIF model and applying it to the T-helper and T-cell activation networks, we show that the SIF model provides more biologically robust results than the existing SIN model of stochasticity in GRNs. AVAILABILITY: Algorithms are made available under our Boolean modeling toolbox, GenYsis. The software binaries can be downloaded from http://si2.epfl.ch/ approximately garg/genysis.html.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The observation of non-random phylogenetic distribution of traits in communities provides evidence for niche-based community assembly. Environment may influence the phylogenetic structure of communities because traits determining how species respond to prevailing conditions can be phylogenetically conserved. In this study, we investigate the variation of butterfly species richness and of phylogenetic - and -diversities along temperature and plant species richness gradients. Our study indicates that butterfly richness is independently positively correlated to temperature and plant species richness in the study area. However, the variation of phylogenetic - and -diversities is only correlated to temperature. The significant phylogenetic clustering at high elevation suggests that cold temperature filters butterfly lineages, leading to communities mostly composed of closely related species adapted to those climatic conditions. These results suggest that in colder and more severe conditions at high elevations deterministic processes and not purely stochastic events drive the assemblage of butterfly communities.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

In the forensic examination of DNA mixtures, the question of how to set the total number of contributors (N) presents a topic of ongoing interest. Part of the discussion gravitates around issues of bias, in particular when assessments of the number of contributors are not made prior to considering the genotypic configuration of potential donors. Further complication may stem from the observation that, in some cases, there may be numbers of contributors that are incompatible with the set of alleles seen in the profile of a mixed crime stain, given the genotype of a potential contributor. In such situations, procedures that take a single and fixed number contributors as their output can lead to inferential impasses. Assessing the number of contributors within a probabilistic framework can help avoiding such complication. Using elements of decision theory, this paper analyses two strategies for inference on the number of contributors. One procedure is deterministic and focuses on the minimum number of contributors required to 'explain' an observed set of alleles. The other procedure is probabilistic using Bayes' theorem and provides a probability distribution for a set of numbers of contributors, based on the set of observed alleles as well as their respective rates of occurrence. The discussion concentrates on mixed stains of varying quality (i.e., different numbers of loci for which genotyping information is available). A so-called qualitative interpretation is pursued since quantitative information such as peak area and height data are not taken into account. The competing procedures are compared using a standard scoring rule that penalizes the degree of divergence between a given agreed value for N, that is the number of contributors, and the actual value taken by N. Using only modest assumptions and a discussion with reference to a casework example, this paper reports on analyses using simulation techniques and graphical models (i.e., Bayesian networks) to point out that setting the number of contributors to a mixed crime stain in probabilistic terms is, for the conditions assumed in this study, preferable to a decision policy that uses categoric assumptions about N.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Les écosystèmes fournissent de nombreuses ressources et services écologiques qui sont utiles à la population humaine. La biodiversité est une composante essentielle des écosystèmes et maintient de nombreux services. Afin d'assurer la permanence des services écosystémiques, des mesures doivent être prises pour conserver la biodiversité. Dans ce but, l'acquisition d'informations détaillées sur la distribution de la biodiversité dans l'espace est essentielle. Les modèles de distribution d'espèces (SDMs) sont des modèles empiriques qui mettent en lien des observations de terrain (présences ou absences d'une espèce) avec des descripteurs de l'environnement, selon des courbes de réponses statistiques qui décrive la niche réalisée des espèces. Ces modèles fournissent des projections spatiales indiquant les lieux les plus favorables pour les espèces considérées. Le principal objectif de cette thèse est de fournir des projections plus réalistes de la distribution des espèces et des communautés en montagne pour le climat présent et futur en considérant non-seulement des variables abiotiques mais aussi biotiques. Les régions de montagne et l'écosystème alpin sont très sensibles aux changements globaux et en même temps assurent de nombreux services écosystémiques. Cette thèse est séparée en trois parties : (i) fournir une meilleure compréhension du rôle des interactions biotiques dans la distribution des espèces et l'assemblage des communautés en montagne (ouest des Alpes Suisses), (ii) permettre le développement d'une nouvelle approche pour modéliser la distribution spatiale de la biodiversité, (iii) fournir des projections plus réalistes de la distribution future des espèces ainsi que de la composition des communautés. En me focalisant sur les papillons, bourdons et plantes vasculaires, j'ai détecté des interactions biotiques importantes qui lient les espèces entre elles. J'ai également identifié la signature du filtre de l'environnement sur les communautés en haute altitude confirmant l'utilité des SDMs pour reproduire ce type de processus. A partir de ces études, j'ai contribué à l'amélioration méthodologique des SDMs dans le but de prédire les communautés en incluant les interactions biotiques et également les processus non-déterministes par une approche probabiliste. Cette approche permet de prédire non-seulement la distribution d'espèces individuelles, mais également celle de communautés dans leur entier en empilant les projections (S-SDMs). Finalement, j'ai utilisé cet outil pour prédire la distribution d'espèces et de communautés dans le passé et le futur. En particulier, j'ai modélisé la migration post-glaciaire de Trollius europaeus qui est à l'origine de la structure génétique intra-spécifique chez cette espèce et évalué les risques de perte face au changement climatique. Finalement, j'ai simulé la distribution des communautés de bourdons pour le 21e siècle afin d'évaluer les changements probables dans ce groupe important de pollinisateurs. La diversité fonctionnelle des bourdons va être altérée par la perte d'espèces spécialistes de haute altitude et ceci va influencer la pollinisation des plantes en haute altitude. - Ecosystems provide a multitude of resources and ecological services, which are useful to human. Biodiversity is an essential component of those ecosystems and guarantee many services. To assure the permanence of ecosystem services for future generation, measure should be applied to conserve biodiversity. For this purpose, the acquisition of detailed information on how biodiversity implicated in ecosystem function is distributed in space is essential. Species distribution models (SDMs) are empirical models relating field observations to environmental predictors based on statistically-derived response surfaces that fit the realized niche. These models result in spatial predictions indicating locations of the most suitable environment for the species and may potentially be applied to predict composition of communities and their functional properties. The main objective of this thesis was to provide more accurate projections of species and communities distribution under current and future climate in mountains by considering not solely abiotic but also biotic drivers of species distribution. Mountain areas and alpine ecosystems are considered as particularly sensitive to global changes and are also sources of essential ecosystem services. This thesis had three main goals: (i) a better ecological understanding of biotic interactions and how they shape the distribution of species and communities, (ii) the development of a novel approach to the spatial modeling of biodiversity, that can account for biotic interactions, and (iii) ecologically more realistic projections of future species distributions, of future composition and structure of communities. Focusing on butterfly and bumblebees in interaction with the vegetation, I detected important biotic interactions for species distribution and community composition of both plant and insects along environmental gradients. I identified the signature of environmental filtering processes at high elevation confirming the suitability of SDMs for reproducing patterns of filtering. Using those case-studies, I improved SDMs by incorporating biotic interaction and accounting for non-deterministic processes and uncertainty using a probabilistic based approach. I used improved modeling to forecast the distribution of species through the past and future climate changes. SDMs hindcasting allowed a better understanding of the spatial range dynamic of Trollius europaeus in Europe at the origin of the species intra-specific genetic diversity and identified the risk of loss of this genetic diversity caused by climate change. By simulating the future distribution of all bumblebee species in the western Swiss Alps under nine climate change scenarios for the 21st century, I found that the functional diversity of this pollinator guild will be largely affected by climate change through the loss of high elevation specialists. In turn, this will have important consequences on alpine plant pollination.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

This paper questions the practitioners' deterministic approach(es) in forensic identification and notes the limits of their conclusions in order to encourage a discussion to question current practices. With this end in view, a hypothetical discussion between an expert in dentistry and an enthusiastic member of a jury, eager to understand the scientific principles of evidence interpretation, is presented. This discussion will lead us to regard any argument aiming at identification as probabilistic.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Purpose: To set local dose reference levels (DRL) that allow radiologists to control stochastic and deterministic effects. Methods and materials: Dose indicators for cerebral angiographies and hepatic embolizations were collected during 4 months and analyzed in our hospital. The data were compared when an image amplifier was used instead of a flat panel detector. The Mann and Whitney test was used. Results: For the 40 cerebral angiographies performed the DRL for DAP, fluoroscopy time and number of images were respectively: 166 Gy.cm2, 19 min, 600. The maximum DAP was 490 Gy.cm2 (fluoroscopy time: 84 min). No significant difference for fluoroscopy time and DAP for image amplifier and flat panel detector (p = 0.88) was observed. The number of images was larger for flat panel detector (p = 0.004). The values obtained were slightly over the present proposed DRL: 150 Gy.cm2, 15 min, 400. Concerning the 13 hepatic embolizations the DRL for DAP fluoroscopy time and number of images were: 315 Gy.cm2, 25 min, 370. The maximum DAP delivered was 845 Gy.cm2 (fluoroscopy time of 48 min). No significant difference between image amplifier and flat panel detector was observed (p = 0.005). The values obtained were also slightly over the present proposed DRL: 300 Gy.cm2, 20 min, 200. Conclusion: These results show that the introduction of flat panel detector did not lead to an increase in patient dose. A DRL concerning the cumulative dose (that allow to control the deterministic effect) should be introduced to allow radiologists to have full control on the risks associated with ionizing radiations. Results of this on going study will be presented.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

T cells belong to two mutually exclusive lineages expressing either alpha beta or gamma delta T-cell receptors (TCR). Although alpha beta and gamma delta cells are known to share a common precursor the role of TCR rearrangement and specificity in the lineage commitment process is controversial. Instructive lineage commitment models endow the alpha beta or gamma delta TCR with a deterministic role in lineage choice, whereas separate lineage models invoke TCR-independent lineage commitment followed by TCR-dependent selection and maturation of alpha beta and gamma delta cells. Here we review the published data pertaining to the role of the TCR in alpha beta/gamma delta lineage commitment and provide some additional information obtained from recent intracellular TCR staining studies. We conclude that a variant of the separate lineage model is best able to accommodate all of the available experimental results.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Data mining can be defined as the extraction of previously unknown and potentially useful information from large datasets. The main principle is to devise computer programs that run through databases and automatically seek deterministic patterns. It is applied in different fields of application, e.g., remote sensing, biometry, speech recognition, but has seldom been applied to forensic case data. The intrinsic difficulty related to the use of such data lies in its heterogeneity, which comes from the many different sources of information. The aim of this study is to highlight potential uses of pattern recognition that would provide relevant results from a criminal intelligence point of view. The role of data mining within a global crime analysis methodology is to detect all types of structures in a dataset. Once filtered and interpreted, those structures can point to previously unseen criminal activities. The interpretation of patterns for intelligence purposes is the final stage of the process. It allows the researcher to validate the whole methodology and to refine each step if necessary. An application to cutting agents found in illicit drug seizures was performed. A combinatorial approach was done, using the presence and the absence of products. Methods coming from the graph theory field were used to extract patterns in data constituted by links between products and place and date of seizure. A data mining process completed using graphing techniques is called ``graph mining''. Patterns were detected that had to be interpreted and compared with preliminary knowledge to establish their relevancy. The illicit drug profiling process is actually an intelligence process that uses preliminary illicit drug classes to classify new samples. Methods proposed in this study could be used \textit{a priori} to compare structures from preliminary and post-detection patterns. This new knowledge of a repeated structure may provide valuable complementary information to profiling and become a source of intelligence.