14 resultados para Government sale of real property

em Université de Lausanne, Switzerland


Relevância:

100.00% 100.00%

Publicador:

Resumo:

Traditional culture-dependent methods to quantify and identify airborne microorganisms are limited by factors such as short-duration sampling times and inability to count nonculturableor non-viable bacteria. Consequently, the quantitative assessment of bioaerosols is often underestimated. Use of the real-time quantitative polymerase chain reaction (Q-PCR) to quantify bacteria in environmental samples presents an alternative method, which should overcome this problem. The aim of this study was to evaluate the performance of a real-time Q-PCR assay as a simple and reliable way to quantify the airborne bacterial load within poultry houses and sewage treatment plants, in comparison with epifluorescencemicroscopy and culture-dependent methods. The estimates of bacterial load that we obtained from real-time PCR and epifluorescence methods, are comparable, however, our analysis of sewage treatment plants indicate these methods give values 270-290 fold greater than those obtained by the ''impaction on nutrient agar'' method. The culture-dependent method of air impaction on nutrient agar was also inadequate in poultry houses, as was the impinger-culture method, which gave a bacterial load estimate 32-fold lower than obtained by Q-PCR. Real-time quantitative PCR thus proves to be a reliable, discerning, and simple method that could be used to estimate airborne bacterial load in a broad variety of other environments expected to carry high numbers of airborne bacteria. [Authors]

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Previous studies have demonstrated that poultry house workers are exposed to very high levels of organic dust and consequently have an increased prevalence of adverse respiratory symptoms. However, the influence of the age of broilers on bioaerosol concentrations has not been investigated. To evaluate the evolution of bioaerosol concentration during the fattening period, bioaerosol parameters (inhalable dust, endotoxin and bacteria) were measured in 12 poultry confinement buildings in Switzerland, at three different stages of the birds' growth; samples of air taken from within the breathing zones of individual poultry house employees as they caught the chickens ready to be transported for slaughter were also analysed. Quantitative polymerase chain reaction (Q-PCR) was used to assess the quantity of total airborne bacteria and total airborne Staphylococcus species. Bioaerosol levels increased significantly during the fattening period of the chickens. During the task of catching mature birds, the mean inhalable dust concentration for a worker was 26 +/- 1.9 mg m(-3) and endotoxin concentration was 6198 +/- 2.3 EU m(-3) air, >6-fold higher than the Swiss occupational recommended value (1000 EU m(-3)). The mean exposure level of bird catchers to total bacteria and Staphylococcus species measured by Q-PCR is also very high, respectively, reaching values of 53 (+/-2.6) x 10(7) cells m(-3) air and 62 (+/-1.9) x 10(6) m(-3) air. It was concluded that in the absence of wearing protective breathing apparatus, chicken catchers in Switzerland risk exposure beyond recommended limits for all measured bioaerosol parameters. Moreover, the use of Q-PCR to estimate total and specific numbers of airborne bacteria is a promising tool for evaluating any modifications intended to improve the safety of current working practices

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Epigenetic silencing of the DNA repair protein O(6)-methylguanine-DNA methyltransferase (MGMT) by promoter methylation predicts successful alkylating agent therapy, such as with temozolomide, in glioblastoma patients. Stratified therapy assignment of patients in prospective clinical trials according to tumor MGMT status requires a standardized diagnostic test, suitable for high-throughput analysis of small amounts of formalin-fixed, paraffin-embedded tumor tissue. A direct, real-time methylation-specific PCR (MSP) assay was developed to determine methylation status of the MGMT gene promoter. Assay specificity was obtained by selective amplification of methylated DNA sequences of sodium bisulfite-modified DNA. The copy number of the methylated MGMT promoter, normalized to the beta-actin gene, provides a quantitative test result. We analyzed 134 clinical glioma samples, comparing the new test with the previously validated nested gel-based MSP assay, which yields a binary readout. A cut-off value for the MGMT methylation status was suggested by fitting a bimodal normal mixture model to the real-time results, supporting the hypothesis that there are two distinct populations within the test samples. Comparison of the tests showed high concordance of the results (82/91 [90%]; Cohen's kappa = 0.80; 95% confidence interval, 0.82-0.95). The direct, real-time MSP assay was highly reproducible (Pearson correlation 0.996) and showed valid test results for 93% (125/134) of samples compared with 75% (94/125) for the nested, gel-based MSP assay. This high-throughput test provides an important pharmacogenomic tool for individualized management of alkylating agent chemotherapy.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Previous studies have demonstrated that poultry-house workers are exposed to very high levels of organic dust and consequently have an increased prevalence of adverse respiratory symptoms. However, the influence of the age of broilers, on bioaerosol concentrations has not been investigated. To evaluate the evolution of bioaerosol concentration during the fattening period, bioaerosol parameters (inhalable dust, endotoxin and bacteria) were measured in 12 poultry confinement buildings in Switzerland, at 3 different stages of the birds' growth; Samples of air taken from within the breathing zones of individual poultry-house employees as they caught the chickens ready to be transported for slaughter, were also analysed. Quantitative PCR (Q-PCR) was used to assess the quantity of total airborne bacteria and total airborne Staphylococcus species. Bioaerosol levels increased significantly during the fattening period of the chickens. During the task of catching mature birds, the mean inhalable dust concentration for a worker was 31 ± 4.7 mg/m3, and endotoxin concentration was 11'080 ± 3436 UE/m3 air, more than ten-fold higher than the Swiss occupational recommended value (1000 UE/m3). The mean exposure level of bird catchers to total bacteria and Staphylococcus species measured by Q-PCR is also very high, respectively reaching values of 72 (± 11) x107 cells/m3 air and 70 (± 16) x106/m3 air. It was concluded that in the absence of wearing protective breathing apparatus, chicken catchers in Switzerland risk exposure beyond recommended limits for all measured bioaerosol parameters. Moreover, the use of Q-PCR to estimate total and specific numbers of airborne bacteria is a promising tool for evaluating any modifications intended to improve the safety of current working practices.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

PURPOSE: Adequate empirical antibiotic dose selection for critically ill burn patients is difficult due to extreme variability in drug pharmacokinetics. Therapeutic drug monitoring (TDM) may aid antibiotic prescription and implementation of initial empirical antimicrobial dosage recommendations. This study evaluated how gradual TDM introduction altered empirical dosages of meropenem and imipenem/cilastatin in our burn ICU. METHODS: Imipenem/cilastatin and meropenem use and daily empirical dosage at a five-bed burn ICU were analyzed retrospectively. Data for all burn admissions between 2001 and 2011 were extracted from the hospital's computerized information system. For each patient receiving a carbapenem, episodes of infection were reviewed and scored according to predefined criteria. Carbapenem trough serum levels were characterized. Prior to May 2007, TDM was available only by special request. Real-time carbapenem TDM was introduced in June 2007; it was initially available weekly and has been available 4 days a week since 2010. RESULTS: Of 365 patients, 229 (63%) received antibiotics (109 received carbapenems). Of 23 TDM determinations for imipenem/cilastatin, none exceeded the predefined upper limit and 11 (47.8%) were insufficient; the number of TDM requests was correlated with daily dose (r=0.7). Similar numbers of inappropriate meropenem trough levels (30.4%) were below and above the upper limit. Real-time TDM introduction increased the empirical dose of imipenem/cilastatin, but not meropenem. CONCLUSIONS: Real-time carbapenem TDM availability significantly altered the empirical daily dosage of imipenem/cilastatin at our burn ICU. Further studies are needed to evaluate the individual impact of TDM-based antibiotic adjustment on infection outcomes in these patients.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Introduction In my thesis I argue that economic policy is all about economics and politics. Consequently, analysing and understanding economic policy ideally has at least two parts. The economics part, which is centered around the expected impact of a specific policy on the real economy both in terms of efficiency and equity. The insights of this part point into which direction the fine-tuning of economic policies should go. However, fine-tuning of economic policies will be most likely subject to political constraints. That is why, in the politics part, a much better understanding can be gained by taking into account how the incentives of politicians and special interest groups as well as the role played by different institutional features affect the formation of economic policies. The first part and chapter of my thesis concentrates on the efficiency-related impact of economic policies: how does corporate income taxation in general, and corporate income tax progressivity in specific, affect the creation of new firms? Reduced progressivity and flat-rate taxes are in vogue. By 2009, 22 countries are operating flat-rate income tax systems, as do 7 US states and 14 Swiss cantons (for corporate income only). Tax reform proposals in the spirit of the "flat tax" model typically aim to reduce three parameters: the average tax burden, the progressivity of the tax schedule, and the complexity of the tax code. In joint work, Marius Brülhart and I explore the implications of changes in these three parameters on entrepreneurial activity, measured by counts of firm births in a panel of Swiss municipalities. Our results show that lower average tax rates and reduced complexity of the tax code promote firm births. Controlling for these effects, reduced progressivity inhibits firm births. Our reading of these results is that tax progressivity has an insurance effect that facilitates entrepreneurial risk taking. The positive effects of lower tax levels and reduced complexity are estimated to be significantly stronger than the negative effect of reduced progressivity. To the extent that firm births reflect desirable entrepreneurial dynamism, it is not the flattening of tax schedules that is key to successful tax reforms, but the lowering of average tax burdens and the simplification of tax codes. Flatness per se is of secondary importance and even appears to be detrimental to firm births. The second part of my thesis, which corresponds to the second and third chapter, concentrates on how economic policies are formed. By the nature of the analysis, these two chapters draw on a broader literature than the first chapter. Both economists and political scientists have done extensive research on how economic policies are formed. Thereby, researchers in both disciplines have recognised the importance of special interest groups trying to influence policy-making through various channels. In general, economists base their analysis on a formal and microeconomically founded approach, while abstracting from institutional details. In contrast, political scientists' frameworks are generally richer in terms of institutional features but lack the theoretical rigour of economists' approaches. I start from the economist's point of view. However, I try to borrow as much as possible from the findings of political science to gain a better understanding of how economic policies are formed in reality. In the second chapter, I take a theoretical approach and focus on the institutional policy framework to explore how interactions between different political institutions affect the outcome of trade policy in presence of special interest groups' lobbying. Standard political economy theory treats the government as a single institutional actor which sets tariffs by trading off social welfare against contributions from special interest groups seeking industry-specific protection from imports. However, these models lack important (institutional) features of reality. That is why, in my model, I split up the government into a legislative and executive branch which can both be lobbied by special interest groups. Furthermore, the legislative has the option to delegate its trade policy authority to the executive. I allow the executive to compensate the legislative in exchange for delegation. Despite ample anecdotal evidence, bargaining over delegation of trade policy authority has not yet been formally modelled in the literature. I show that delegation has an impact on policy formation in that it leads to lower equilibrium tariffs compared to a standard model without delegation. I also show that delegation will only take place if the lobby is not strong enough to prevent it. Furthermore, the option to delegate increases the bargaining power of the legislative at the expense of the lobbies. Therefore, the findings of this model can shed a light on why the U.S. Congress often practices delegation to the executive. In the final chapter of my thesis, my coauthor, Antonio Fidalgo, and I take a narrower approach and focus on the individual politician level of policy-making to explore how connections to private firms and networks within parliament affect individual politicians' decision-making. Theories in the spirit of the model of the second chapter show how campaign contributions from lobbies to politicians can influence economic policies. There exists an abundant empirical literature that analyses ties between firms and politicians based on campaign contributions. However, the evidence on the impact of campaign contributions is mixed, at best. In our paper, we analyse an alternative channel of influence in the shape of personal connections between politicians and firms through board membership. We identify a direct effect of board membership on individual politicians' voting behaviour and an indirect leverage effect when politicians with board connections influence non-connected peers. We assess the importance of these two effects using a vote in the Swiss parliament on a government bailout of the national airline, Swissair, in 2001, which serves as a natural experiment. We find that both the direct effect of connections to firms and the indirect leverage effect had a strong and positive impact on the probability that a politician supported the government bailout.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Different interferometric techniques were developed last decade to obtain full field, quantitative, and absolute phase imaging, such as phase-shifting, Fourier phase microscopy, Hilbert phase microscopy or digital holographic microscopy (DHM). Although, these techniques are very similar, DHM combines several advantages. In contrast, to phase shifting, DHM is indeed capable of single-shot hologram recording allowing a real-time absolute phase imaging. On the other hand, unlike to Fourier phase or Hilbert phase microscopy, DHM does not require to record in focus images of the specimen on the digital detector (CCD or CMOS camera), because a numerical focalization adjustment can be performed by a numerical wavefront propagation. Consequently, the depth of view of high NA microscope objectives is numerically extended. For example, two different biological cells, floating at different depths in a liquid, can be focalized numerically from the same digital hologram. Moreover, the numerical propagation associated to digital optics and automatic fitting procedures, permits vibrations insensitive full- field phase imaging and the complete compensation for a priori any image distortion or/and phase aberrations introduced for example by imperfections of holders or perfusion chamber. Examples of real-time full-field phase images of biological cells have been demonstrated. ©2008 COPYRIGHT SPIE

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Most airborne microorganisms are natural components of our ecosystem. Soil, vegetation and animals, including humans, are sources for aerial release of these living or dead cells. In the past, assessment of airborne microorganisms was mainly restricted to occupational health concerns. Indeed, in several occupations, exposure to very high concentrations of non-infectious airborne bacteria and fungi, result in allergenic, toxic or irritant reactions. Recently, the threat of bioterrorism and pandemics have highlighted the urgent need to increase knowledge of bioaerosol ecology. More fundamentally, airborne bacterial and fungal communities begin to draw much more consideration from environmental microbiologists, who have neglected this area for a long time. This increased interest of scientists is to a great part due to the development and use of real-time PCR techniques to identify and quantify airborne microorganisms. Even if the advantages of the PCR technology are obvious, researchers are confronted with new problems. This review describes the methodological state of the art in bioaerosols field and emphasizes the future challenges and perspectives of the real-time PCR-based methods for airborne microorganism studies.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Abstract The object of game theory lies in the analysis of situations where different social actors have conflicting requirements and where their individual decisions will all influence the global outcome. In this framework, several games have been invented to capture the essence of various dilemmas encountered in many common important socio-economic situations. Even though these games often succeed in helping us understand human or animal behavior in interactive settings, some experiments have shown that people tend to cooperate with each other in situations for which classical game theory strongly recommends them to do the exact opposite. Several mechanisms have been invoked to try to explain the emergence of this unexpected cooperative attitude. Among them, repeated interaction, reputation, and belonging to a recognizable group have often been mentioned. However, the work of Nowak and May (1992) showed that the simple fact of arranging the players according to a spatial structure and only allowing them to interact with their immediate neighbors is sufficient to sustain a certain amount of cooperation even when the game is played anonymously and without repetition. Nowak and May's study and much of the following work was based on regular structures such as two-dimensional grids. Axelrod et al. (2002) showed that by randomizing the choice of neighbors, i.e. by actually giving up a strictly local geographical structure, cooperation can still emerge, provided that the interaction patterns remain stable in time. This is a first step towards a social network structure. However, following pioneering work by sociologists in the sixties such as that of Milgram (1967), in the last few years it has become apparent that many social and biological interaction networks, and even some technological networks, have particular, and partly unexpected, properties that set them apart from regular or random graphs. Among other things, they usually display broad degree distributions, and show small-world topological structure. Roughly speaking, a small-world graph is a network where any individual is relatively close, in terms of social ties, to any other individual, a property also found in random graphs but not in regular lattices. However, in contrast with random graphs, small-world networks also have a certain amount of local structure, as measured, for instance, by a quantity called the clustering coefficient. In the same vein, many real conflicting situations in economy and sociology are not well described neither by a fixed geographical position of the individuals in a regular lattice, nor by a random graph. Furthermore, it is a known fact that network structure can highly influence dynamical phenomena such as the way diseases spread across a population and ideas or information get transmitted. Therefore, in the last decade, research attention has naturally shifted from random and regular graphs towards better models of social interaction structures. The primary goal of this work is to discover whether or not the underlying graph structure of real social networks could give explanations as to why one finds higher levels of cooperation in populations of human beings or animals than what is prescribed by classical game theory. To meet this objective, I start by thoroughly studying a real scientific coauthorship network and showing how it differs from biological or technological networks using divers statistical measurements. Furthermore, I extract and describe its community structure taking into account the intensity of a collaboration. Finally, I investigate the temporal evolution of the network, from its inception to its state at the time of the study in 2006, suggesting also an effective view of it as opposed to a historical one. Thereafter, I combine evolutionary game theory with several network models along with the studied coauthorship network in order to highlight which specific network properties foster cooperation and shed some light on the various mechanisms responsible for the maintenance of this same cooperation. I point out the fact that, to resist defection, cooperators take advantage, whenever possible, of the degree-heterogeneity of social networks and their underlying community structure. Finally, I show that cooperation level and stability depend not only on the game played, but also on the evolutionary dynamic rules used and the individual payoff calculations. Synopsis Le but de la théorie des jeux réside dans l'analyse de situations dans lesquelles différents acteurs sociaux, avec des objectifs souvent conflictuels, doivent individuellement prendre des décisions qui influenceront toutes le résultat global. Dans ce cadre, plusieurs jeux ont été inventés afin de saisir l'essence de divers dilemmes rencontrés dans d'importantes situations socio-économiques. Bien que ces jeux nous permettent souvent de comprendre le comportement d'êtres humains ou d'animaux en interactions, des expériences ont montré que les individus ont parfois tendance à coopérer dans des situations pour lesquelles la théorie classique des jeux prescrit de faire le contraire. Plusieurs mécanismes ont été invoqués pour tenter d'expliquer l'émergence de ce comportement coopératif inattendu. Parmi ceux-ci, la répétition des interactions, la réputation ou encore l'appartenance à des groupes reconnaissables ont souvent été mentionnés. Toutefois, les travaux de Nowak et May (1992) ont montré que le simple fait de disposer les joueurs selon une structure spatiale en leur permettant d'interagir uniquement avec leurs voisins directs est suffisant pour maintenir un certain niveau de coopération même si le jeu est joué de manière anonyme et sans répétitions. L'étude de Nowak et May, ainsi qu'un nombre substantiel de travaux qui ont suivi, étaient basés sur des structures régulières telles que des grilles à deux dimensions. Axelrod et al. (2002) ont montré qu'en randomisant le choix des voisins, i.e. en abandonnant une localisation géographique stricte, la coopération peut malgré tout émerger, pour autant que les schémas d'interactions restent stables au cours du temps. Ceci est un premier pas en direction d'une structure de réseau social. Toutefois, suite aux travaux précurseurs de sociologues des années soixante, tels que ceux de Milgram (1967), il est devenu clair ces dernières années qu'une grande partie des réseaux d'interactions sociaux et biologiques, et même quelques réseaux technologiques, possèdent des propriétés particulières, et partiellement inattendues, qui les distinguent de graphes réguliers ou aléatoires. Entre autres, ils affichent en général une distribution du degré relativement large ainsi qu'une structure de "petit-monde". Grossièrement parlant, un graphe "petit-monde" est un réseau où tout individu se trouve relativement près de tout autre individu en termes de distance sociale, une propriété également présente dans les graphes aléatoires mais absente des grilles régulières. Par contre, les réseaux "petit-monde" ont, contrairement aux graphes aléatoires, une certaine structure de localité, mesurée par exemple par une quantité appelée le "coefficient de clustering". Dans le même esprit, plusieurs situations réelles de conflit en économie et sociologie ne sont pas bien décrites ni par des positions géographiquement fixes des individus en grilles régulières, ni par des graphes aléatoires. De plus, il est bien connu que la structure même d'un réseau peut passablement influencer des phénomènes dynamiques tels que la manière qu'a une maladie de se répandre à travers une population, ou encore la façon dont des idées ou une information s'y propagent. Ainsi, durant cette dernière décennie, l'attention de la recherche s'est tout naturellement déplacée des graphes aléatoires et réguliers vers de meilleurs modèles de structure d'interactions sociales. L'objectif principal de ce travail est de découvrir si la structure sous-jacente de graphe de vrais réseaux sociaux peut fournir des explications quant aux raisons pour lesquelles on trouve, chez certains groupes d'êtres humains ou d'animaux, des niveaux de coopération supérieurs à ce qui est prescrit par la théorie classique des jeux. Dans l'optique d'atteindre ce but, je commence par étudier un véritable réseau de collaborations scientifiques et, en utilisant diverses mesures statistiques, je mets en évidence la manière dont il diffère de réseaux biologiques ou technologiques. De plus, j'extrais et je décris sa structure de communautés en tenant compte de l'intensité d'une collaboration. Finalement, j'examine l'évolution temporelle du réseau depuis son origine jusqu'à son état en 2006, date à laquelle l'étude a été effectuée, en suggérant également une vue effective du réseau par opposition à une vue historique. Par la suite, je combine la théorie évolutionnaire des jeux avec des réseaux comprenant plusieurs modèles et le réseau de collaboration susmentionné, afin de déterminer les propriétés structurelles utiles à la promotion de la coopération et les mécanismes responsables du maintien de celle-ci. Je mets en évidence le fait que, pour ne pas succomber à la défection, les coopérateurs exploitent dans la mesure du possible l'hétérogénéité des réseaux sociaux en termes de degré ainsi que la structure de communautés sous-jacente de ces mêmes réseaux. Finalement, je montre que le niveau de coopération et sa stabilité dépendent non seulement du jeu joué, mais aussi des règles de la dynamique évolutionnaire utilisées et du calcul du bénéfice d'un individu.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Three-dimensional free-breathing coronary magnetic resonance angiography was performed in eight healthy volunteers with use of real-time navigator technology. Images acquired with the navigator localized at the right hemidiaphragm and at the left ventricle were objectively compared. The diaphragmatic navigator was found to be superior for vessel delineation of middle to distal portions of the coronary arteries.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Abstract Traditionally, the common reserving methods used by the non-life actuaries are based on the assumption that future claims are going to behave in the same way as they did in the past. There are two main sources of variability in the processus of development of the claims: the variability of the speed with which the claims are settled and the variability between the severity of the claims from different accident years. High changes in these processes will generate distortions in the estimation of the claims reserves. The main objective of this thesis is to provide an indicator which firstly identifies and quantifies these two influences and secondly to determine which model is adequate for a specific situation. Two stochastic models were analysed and the predictive distributions of the future claims were obtained. The main advantage of the stochastic models is that they provide measures of variability of the reserves estimates. The first model (PDM) combines one conjugate family Dirichlet - Multinomial with the Poisson distribution. The second model (NBDM) improves the first one by combining two conjugate families Poisson -Gamma (for distribution of the ultimate amounts) and Dirichlet Multinomial (for distribution of the incremental claims payments). It was found that the second model allows to find the speed variability in the reporting process and development of the claims severity as function of two above mentioned distributions' parameters. These are the shape parameter of the Gamma distribution and the Dirichlet parameter. Depending on the relation between them we can decide on the adequacy of the claims reserve estimation method. The parameters have been estimated by the Methods of Moments and Maximum Likelihood. The results were tested using chosen simulation data and then using real data originating from the three lines of business: Property/Casualty, General Liability, and Accident Insurance. These data include different developments and specificities. The outcome of the thesis shows that when the Dirichlet parameter is greater than the shape parameter of the Gamma, resulting in a model with positive correlation between the past and future claims payments, suggests the Chain-Ladder method as appropriate for the claims reserve estimation. In terms of claims reserves, if the cumulated payments are high the positive correlation will imply high expectations for the future payments resulting in high claims reserves estimates. The negative correlation appears when the Dirichlet parameter is lower than the shape parameter of the Gamma, meaning low expected future payments for the same high observed cumulated payments. This corresponds to the situation when claims are reported rapidly and fewer claims remain expected subsequently. The extreme case appears in the situation when all claims are reported at the same time leading to expectations for the future payments of zero or equal to the aggregated amount of the ultimate paid claims. For this latter case, the Chain-Ladder is not recommended.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Multi-phase postmortem CT angiography (MPMCTA) is recognized as a valuable tool to explore the vascular system, with higher sensitivity than conventional autopsy. However, a limitation is the impossibility to diagnose pulmonary embolism (PE) due to post-mortem blood clots situated in pulmonary arteries. The purpose of this study was to explore an eventual possibility to distinguish between real PE and artefacts mimicking PE. Our study included 416 medico-legal cases. All of them underwent MPMCTA, conventional autopsy and histological examination. We selected cases presenting arterial luminal filling defects in the pulmonary arteries. Their radiological interpretation was confronted to the one of autopsy and histological examination. We also investigated an eventual correlation between artefacts in pulmonary arteries and those in other parts of the vascular system. In 123 cases, filling defects of pulmonary arteries were described during MPMCTA. In 57 cases, this was interpreted as artefact and in 4 cases as suspected PE. In 62 cases only a differential diagnosis was made. Autopsy and histology could clearly identify the artefacts as such. Only one case of real PE was radiologically misinterpreted as artefact. In 6 of the 62 cases with no interpretation a PE was diagnosed. In 3 out of 4 suspected cases, PE was confirmed. We found out that filling defects in pulmonary arteries are nearly always associated to other vascular artefacts. Therefore, we suggest following some rules for radiological interpretation in order to allow a reliable diagnosis of pulmonary embolism after MPMCTA.