23 resultados para driving under the informed and presumed “Consent”.
Resumo:
Two retrospective epidemiologic studies have shown that cannabis is the main psychoactive substance detected in the blood of drivers suspected of driving under the influence of psychotropic drugs. An oral administration double-blind crossover study was carried out with eight healthy male subjects, aged 22 to 30 years, all occasional cannabis smokers. Three treatments and one placebo were administered to all participants at a two week interval: 20 mg dronabinol, 16.5 mg D9-tétrahydrocannabinol (THC) and 45.7 mg THC as a cannabis milk decoction. Participants were asked to report the subjective drug effects and their willingness to drive under various circumstances on a visual analog scale. Clinical observations, a psychomotor test and a tracking test on a driving simulator were also carried out. Compared to cannabis smoking, THC, 11-OH-THC and THC-COOH blood concentrations remained low through the whole study (<13.1 ng THC/mL,<24.7 ng 11-OH-THC/mL and<99.9 ng THC-COOH/mL). Two subjects experienced deep anxiety symptoms suggesting that this unwanted side-effect may occur when driving under the influence of cannabis or when driving and smoking a joint. No clear association could be found between these adverse reactions and a susceptibility gene to propensity to anxiety and psychotic symptoms (genetic polymorphism of the catechol-O-methyltransferase). The questionnaires have shown that the willingness to drive was lower when the drivers were assigned an insignificant task and was higher when the mission was of crucial importance. The subjects were aware of the effects of cannabis and their performances on the road sign and tracking test were greatly impaired, especially after ingestion of the strongest dose. The Cannabis Influence Factor (CIF) which relies on the molar ratio of active and inactive cannabinoids in blood provided a good estimate of the fitness to drive.
Resumo:
General Summary Although the chapters of this thesis address a variety of issues, the principal aim is common: test economic ideas in an international economic context. The intention has been to supply empirical findings using the largest suitable data sets and making use of the most appropriate empirical techniques. This thesis can roughly be divided into two parts: the first one, corresponding to the first two chapters, investigates the link between trade and the environment, the second one, the last three chapters, is related to economic geography issues. Environmental problems are omnipresent in the daily press nowadays and one of the arguments put forward is that globalisation causes severe environmental problems through the reallocation of investments and production to countries with less stringent environmental regulations. A measure of the amplitude of this undesirable effect is provided in the first part. The third and the fourth chapters explore the productivity effects of agglomeration. The computed spillover effects between different sectors indicate how cluster-formation might be productivity enhancing. The last chapter is not about how to better understand the world but how to measure it and it was just a great pleasure to work on it. "The Economist" writes every week about the impressive population and economic growth observed in China and India, and everybody agrees that the world's center of gravity has shifted. But by how much and how fast did it shift? An answer is given in the last part, which proposes a global measure for the location of world production and allows to visualize our results in Google Earth. A short summary of each of the five chapters is provided below. The first chapter, entitled "Unraveling the World-Wide Pollution-Haven Effect" investigates the relative strength of the pollution haven effect (PH, comparative advantage in dirty products due to differences in environmental regulation) and the factor endowment effect (FE, comparative advantage in dirty, capital intensive products due to differences in endowments). We compute the pollution content of imports using the IPPS coefficients (for three pollutants, namely biological oxygen demand, sulphur dioxide and toxic pollution intensity for all manufacturing sectors) provided by the World Bank and use a gravity-type framework to isolate the two above mentioned effects. Our study covers 48 countries that can be classified into 29 Southern and 19 Northern countries and uses the lead content of gasoline as proxy for environmental stringency. For North-South trade we find significant PH and FE effects going in the expected, opposite directions and being of similar magnitude. However, when looking at world trade, the effects become very small because of the high North-North trade share, where we have no a priori expectations about the signs of these effects. Therefore popular fears about the trade effects of differences in environmental regulations might by exaggerated. The second chapter is entitled "Is trade bad for the Environment? Decomposing worldwide SO2 emissions, 1990-2000". First we construct a novel and large database containing reasonable estimates of SO2 emission intensities per unit labor that vary across countries, periods and manufacturing sectors. Then we use these original data (covering 31 developed and 31 developing countries) to decompose the worldwide SO2 emissions into the three well known dynamic effects (scale, technique and composition effect). We find that the positive scale (+9,5%) and the negative technique (-12.5%) effect are the main driving forces of emission changes. Composition effects between countries and sectors are smaller, both negative and of similar magnitude (-3.5% each). Given that trade matters via the composition effects this means that trade reduces total emissions. We next construct, in a first experiment, a hypothetical world where no trade happens, i.e. each country produces its imports at home and does no longer produce its exports. The difference between the actual and this no-trade world allows us (under the omission of price effects) to compute a static first-order trade effect. The latter now increases total world emissions because it allows, on average, dirty countries to specialize in dirty products. However, this effect is smaller (3.5%) in 2000 than in 1990 (10%), in line with the negative dynamic composition effect identified in the previous exercise. We then propose a second experiment, comparing effective emissions with the maximum or minimum possible level of SO2 emissions. These hypothetical levels of emissions are obtained by reallocating labour accordingly across sectors within each country (under the country-employment and the world industry-production constraints). Using linear programming techniques, we show that emissions are reduced by 90% with respect to the worst case, but that they could still be reduced further by another 80% if emissions were to be minimized. The findings from this chapter go together with those from chapter one in the sense that trade-induced composition effect do not seem to be the main source of pollution, at least in the recent past. Going now to the economic geography part of this thesis, the third chapter, entitled "A Dynamic Model with Sectoral Agglomeration Effects" consists of a short note that derives the theoretical model estimated in the fourth chapter. The derivation is directly based on the multi-regional framework by Ciccone (2002) but extends it in order to include sectoral disaggregation and a temporal dimension. This allows us formally to write present productivity as a function of past productivity and other contemporaneous and past control variables. The fourth chapter entitled "Sectoral Agglomeration Effects in a Panel of European Regions" takes the final equation derived in chapter three to the data. We investigate the empirical link between density and labour productivity based on regional data (245 NUTS-2 regions over the period 1980-2003). Using dynamic panel techniques allows us to control for the possible endogeneity of density and for region specific effects. We find a positive long run elasticity of density with respect to labour productivity of about 13%. When using data at the sectoral level it seems that positive cross-sector and negative own-sector externalities are present in manufacturing while financial services display strong positive own-sector effects. The fifth and last chapter entitled "Is the World's Economic Center of Gravity Already in Asia?" computes the world economic, demographic and geographic center of gravity for 1975-2004 and compares them. Based on data for the largest cities in the world and using the physical concept of center of mass, we find that the world's economic center of gravity is still located in Europe, even though there is a clear shift towards Asia. To sum up, this thesis makes three main contributions. First, it provides new estimates of orders of magnitudes for the role of trade in the globalisation and environment debate. Second, it computes reliable and disaggregated elasticities for the effect of density on labour productivity in European regions. Third, it allows us, in a geometrically rigorous way, to track the path of the world's economic center of gravity.
Resumo:
The aim of radiotherapy is to deliver enough radiation to the tumor in order to achieve maximum tumour control in the irradiated volume with as few serious complications as possible with an irradiation dose as low as possible to normal tissue. The quality of radiotherapy is essential for optimal treatment and quality control is to reduce the bias in clinical trials avoiding possible major deviations. The assurance and quality control programs have been developed in large european (EORTC, GORTEC) and american cooperative groups (RTOG) of radiation oncology since the 1980s. We insist here on the importance of quality assurance in radiotherapy and the current status in this domain and the criteria for quality control especially for current clinical trials within GORTEC are discussed here.
Resumo:
Graph theory has provided a key mathematical framework to analyse the architecture of human brain networks. This architecture embodies an inherently complex relationship between connection topology, the spatial arrangement of network elements, and the resulting network cost and functional performance. An exploration of these interacting factors and driving forces may reveal salient network features that are critically important for shaping and constraining the brain's topological organization and its evolvability. Several studies have pointed to an economic balance between network cost and network efficiency with networks organized in an 'economical' small-world favouring high communication efficiency at a low wiring cost. In this study, we define and explore a network morphospace in order to characterize different aspects of communication efficiency in human brain networks. Using a multi-objective evolutionary approach that approximates a Pareto-optimal set within the morphospace, we investigate the capacity of anatomical brain networks to evolve towards topologies that exhibit optimal information processing features while preserving network cost. This approach allows us to investigate network topologies that emerge under specific selection pressures, thus providing some insight into the selectional forces that may have shaped the network architecture of existing human brains.
Resumo:
This paper analyses how banking regulation was introduced in Switzerland - one of the world's most prominent financial centres - which remained in place until the beginning of the twenty-first century. It shows that the law adopted on 8 November 1934 is a perfect example of capture of the regulator by the regulated. Essentially a political response in the context of the economic crisis of the 1930s, it largely reflected the interests of banking circles by limiting the intervention of the State as much as possible. The introduction of the new legislation was facilitated by the temporary weakness of Swiss banking circles, as they depended on the State to delay or prevent the collapse of many major credit institutions. They did not manage to derail the law as they had two decades earlier when they scuppered the federal bill on banks drawn up between 1914 and 1916. But this time they were better organized and more united, and intervened all the more effectively in the legislative process itself. The 1934 law is thus distinctive in that it made no structural changes to the architecture of the financial centre but merely codified its practices through flexible legislation meant to reassure the public. The law was aimed less at controlling banking activity than at keeping - thanks to skilfully calibrated political concessions - the State from having to intervene more directly in the internal management of banks or in the fixing of interest rates and the export of capital.
Resumo:
L'exposition professionnelle aux nanomatériaux manufacturés dans l'air présente des risques potentiels pour la santé des travailleurs dans les secteurs de la nanotechnologie. Il est important de comprendre les scénarios de libération des aérosols de nanoparticules dans les processus et les activités associées à l'exposition humaine. Les mécanismes de libération, y compris les taux de libération et les propriétés physico-chimiques des nanoparticules, déterminent leurs comportements de transport ainsi que les effets biologiques néfastes. La distribution de taille des particules d'aérosols est l'un des paramètres les plus importants dans ces processus. La stabilité mécanique d'agglomérats de nanoparticules affecte leurs distributions de tailles. Les potentiels de désagglomération de ces agglomérats déterminent les possibilités de leur déformation sous énergies externes. Cela rend les changements possibles dans leur distribution de taille et de la concentration en nombre qui vont finalement modifier leurs risques d'exposition. Les conditions environnementales, telles que l'humidité relative, peuvent influencer les processus de désagglomération par l'adhérence de condensation capillaire de l'humidité. L'objectif général de cette thèse était d'évaluer les scénarios de libération des nanomatériaux manufacturés des processus et activités sur le lieu de travail. Les sous-objectifs étaient les suivants: 1. Etudier les potentiels de désagglomération des nanoparticules dans des conditions environnementales variées. 2. Etudier la libération des nano-objets à partir de nanocomposites polymères; 3. Evaluer la libération de nanoparticules sur le lieu de travail dans des situations concrètes. Nous avons comparé différents systèmes de laboratoire qui présentaient différents niveau d'énergie dans l'aérosolisation des poudres. Des nanopoudres de TiO2 avec des hydrophilicités de surface distinctes ont été testées. Un spectromètre à mobilité électrique (SMPS), un spectromètre à mobilité aérodynamique (APS) et un spectromètre optique (OPC) ont été utilisés pour mesurer la concentration de particules et la distribution de taille des particules. La microscopie électronique à transmission (TEM) a été utilisée pour l'analyse morphologique d'échantillons de particules dans l'air. Les propriétés des aérosols (distribution de taille et concentration en nombre) étaient différentes suivant la méthode employée. Les vitesses des flux d'air d'aérosolisation ont été utilisées pour estimer le niveau d'énergie dans ces systèmes, et il a été montré que les tailles modales des particules étaient inversement proportionnelles à la vitesse appliquée. En général, les particules hydrophiles ont des diamètres plus grands et des nombres inférieurs à ceux des particules hydrophobes. Toutefois, cela dépend aussi des méthodes utilisées. La vitesse de l'air peut donc être un paramètre efficace pour le classement de l'énergie des procédés pour des systèmes d'aérosolisation similaires. Nous avons développé un système laboratoire pour tester les potentiels de désagglomération des nanoparticules dans l'air en utilisant des orifices critiques et un humidificateur. Sa performance a été comparée à un système similaire dans un institut partenaire. Une variété de nanopoudres différentes a été testée. Le niveau d'énergie appliquée et l'humidité ont été modifiés. Le SMPS et l'OPC ont été utilisés pour mesurer la concentration de particules et la distribution de la taille. Un TEM a été utilisé pour l'analyse morphologique d'échantillons de particules dans l'air. Le diamètre moyen des particules a diminué et la concentration en nombre s'est accrue lorsque des énergies externes ont été appliquées. Le nombre de particules inférieures à 100 nm a été augmenté, et celui au-dessus de 350 nm réduits. Les conditions humides ont faits exactement le contraire, en particulier pour les petites particules. En outre, ils ont réduits les effets de la différence de pression due à l'orifice. Les résultats suggèrent que la désagglomération d'agglomérats de nanoparticules dans l'air est possible dans la gamme d'énergie appliquée. Cependant, l'atmosphère humide peut favoriser leur agglomération et améliorer leurs stabilités en réduisant la libération de nanoparticules dans l'environnement. Nous proposons d'utiliser notre système pour le test de routine des potentiels de désagglomération des nanomatériaux manufacturés et de les classer. Un tel classement faciliterait la priorisation de l'exposition et du risque encouru en fonction du niveau d'ENM. Un système de perçage automatique et un système de sciage manuel ont été développés pour étudier la libération de nanoparticules à partir de différents types de nanocomposites. La vitesse de perçage et taille de la mèche ont été modifiées dans les expériences. La distribution de taille des particules et leur concentration en nombre ont été mesurées par un SMPS et un miniature diffusion size classifier (DISCmini). Les distributions de nanoparticules dans les composites et les particules libérées ont été analysés par un TEM et un microscope électronique à balayage (SEM). Les tests de perçage ont libérés un plus grand nombre de particules que le sciage. Des vitesses de perçage plus rapide et les mèches plus grandes ont augmentés la génération de particules. Les charges de nanoparticules manufacturées dans les composites ne modifient pas leurs comportements de libération dans les expériences de perçage. Toutefois, le sciage différencie les niveaux de libération entre les composites et les échantillons blancs. De plus, les vapeurs de polymères ont été générées par la chaleur de sciage. La plupart des particules libérées sont des polymères contenant des nanoparticules ou sur leurs surface. Les résultats ont souligné l'importance du type de processus et paramètres pour déterminer la libération de nanoparticules de composites. Les émissions secondaires telles que les fumées polymères appellent à la nécessité d'évaluations de l'exposition et de risque pour de tels scénarios. Une revue systématique de la littérature sur le sujet de libérations de nanoparticules dans l'air dans les secteurs industriels et laboratoires de recherche a été effectuée. Des stratégies de recherche des informations pertinentes et de stockage ont été développées. Les mécanismes de libération, tels que la taille de particules d'aérosol et de leur concentration en nombre, ont été comparés pour différentes activités. La disponibilité de l'information contextuelle qui est pertinente pour l'estimation de l'exposition humaine a été évaluée. Il a été constaté que les données relatives à l'exposition ne sont pas toujours disponibles dans la littérature actuelle. Les propriétés des aérosols libérés semblent dépendre de la nature des activités. Des procédés à haute énergie ont tendance à générer des plus hauts niveaux de concentrations de particules dans les gammes de plus petite taille. Les résultats peuvent être utiles pour déterminer la priorité des procédés industriels pour l'évaluation les risques associés dans une approche à plusieurs niveaux. Pour l'évaluation de l'exposition, la disponibilité de l'information peut être améliorée par le développement d'une meilleure méthode de communication des données.