29 resultados para Nearest Neighbour


Relevância:

10.00% 10.00%

Publicador:

Resumo:

The paper deals with the development and application of the generic methodology for automatic processing (mapping and classification) of environmental data. General Regression Neural Network (GRNN) is considered in detail and is proposed as an efficient tool to solve the problem of spatial data mapping (regression). The Probabilistic Neural Network (PNN) is considered as an automatic tool for spatial classifications. The automatic tuning of isotropic and anisotropic GRNN/PNN models using cross-validation procedure is presented. Results are compared with the k-Nearest-Neighbours (k-NN) interpolation algorithm using independent validation data set. Real case studies are based on decision-oriented mapping and classification of radioactively contaminated territories.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

We report a new set of nine primer pairs specifically developed for amplification of Brassica plastid SSR markers. The wide utility of these markers is demonstrated for haplotype identification and detection of polymorphism in B. napus, B. nigra, B. oleracea, B. rapa and in related genera Arabidopsis, Camelina, Raphanus and Sinapis. Eleven gene regions (ndhB-rps7 spacer, rbcL-accD spacer, rpl16 intron, rps16 intron, atpB-rbcL spacer, trnE-trnT spacer, trnL intron, trnL-trnF spacer, trnM-atpE spacer, trnR-rpoC2 spacer, ycf3-psaA spacer) were sequenced from a range of Brassica and related genera for SSR detection and primer design. Other sequences were obtained from GenBank/EMBL. Eight out of nine selected SSR loci showed polymorphism when amplified using the new primers and a combined analysis detected variation within and between Brassica species, with the number of alleles detected per locus ranging from 5 (loci MF-6, MF-1) to 11 (locus MF-7). The combined SSR data were used in a neighbour-joining analysis (SMM, D (DM) distances) to group the samples based on the presence and absence of alleles. The analysis was generally able to separate plastid types into taxon-specific groups. Multi-allelic haplotypes were plotted onto the neighbour joining tree. A total number of 28 haplotypes were detected and these differentiated 22 of the 41 accessions screened from all other accessions. None of these haplotypes was shared by more than one species and some were not characteristic of their predicted type. We interpret our results with respect to taxon differentiation, hybridisation and introgression patterns relating to the 'Triangle of U'.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The present research deals with an important public health threat, which is the pollution created by radon gas accumulation inside dwellings. The spatial modeling of indoor radon in Switzerland is particularly complex and challenging because of many influencing factors that should be taken into account. Indoor radon data analysis must be addressed from both a statistical and a spatial point of view. As a multivariate process, it was important at first to define the influence of each factor. In particular, it was important to define the influence of geology as being closely associated to indoor radon. This association was indeed observed for the Swiss data but not probed to be the sole determinant for the spatial modeling. The statistical analysis of data, both at univariate and multivariate level, was followed by an exploratory spatial analysis. Many tools proposed in the literature were tested and adapted, including fractality, declustering and moving windows methods. The use of Quan-tité Morisita Index (QMI) as a procedure to evaluate data clustering in function of the radon level was proposed. The existing methods of declustering were revised and applied in an attempt to approach the global histogram parameters. The exploratory phase comes along with the definition of multiple scales of interest for indoor radon mapping in Switzerland. The analysis was done with a top-to-down resolution approach, from regional to local lev¬els in order to find the appropriate scales for modeling. In this sense, data partition was optimized in order to cope with stationary conditions of geostatistical models. Common methods of spatial modeling such as Κ Nearest Neighbors (KNN), variography and General Regression Neural Networks (GRNN) were proposed as exploratory tools. In the following section, different spatial interpolation methods were applied for a par-ticular dataset. A bottom to top method complexity approach was adopted and the results were analyzed together in order to find common definitions of continuity and neighborhood parameters. Additionally, a data filter based on cross-validation was tested with the purpose of reducing noise at local scale (the CVMF). At the end of the chapter, a series of test for data consistency and methods robustness were performed. This lead to conclude about the importance of data splitting and the limitation of generalization methods for reproducing statistical distributions. The last section was dedicated to modeling methods with probabilistic interpretations. Data transformation and simulations thus allowed the use of multigaussian models and helped take the indoor radon pollution data uncertainty into consideration. The catego-rization transform was presented as a solution for extreme values modeling through clas-sification. Simulation scenarios were proposed, including an alternative proposal for the reproduction of the global histogram based on the sampling domain. The sequential Gaussian simulation (SGS) was presented as the method giving the most complete information, while classification performed in a more robust way. An error measure was defined in relation to the decision function for data classification hardening. Within the classification methods, probabilistic neural networks (PNN) show to be better adapted for modeling of high threshold categorization and for automation. Support vector machines (SVM) on the contrary performed well under balanced category conditions. In general, it was concluded that a particular prediction or estimation method is not better under all conditions of scale and neighborhood definitions. Simulations should be the basis, while other methods can provide complementary information to accomplish an efficient indoor radon decision making.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Plants forming a rosette during their juvenile growth phase, such as Arabidopsis thaliana (L.) Heynh., are able to adjust the size, position and orientation of their leaves. These growth responses are under the control of the plants circadian clock and follow a characteristic diurnal rhythm. For instance, increased leaf elongation and hyponasty - defined here as the increase in leaf elevation angle - can be observed when plants are shaded. Shading can either be caused by a decrease in the fluence rate of photosynthetically active radiation (direct shade) or a decrease in the fluence rate of red compared with far-red radiation (neighbour detection). In this paper we report on a phenotyping approach based on laser scanning to measure the diurnal pattern of leaf hyponasty and increase in rosette size. In short days, leaves showed constitutively increased leaf elevation angles compared with long days, but the overall diurnal pattern and the magnitude of up and downward leaf movement was independent of daylength. Shade treatment led to elevated leaf angles during the first day of application, but did not affect the magnitude of up and downward leaf movement in the following day. Using our phenotyping device, individual plants can be non-invasively monitored during several days under different light conditions. Hence, it represents a proper tool to phenotype light- and circadian clock-mediated growth responses in order to better understand the underlying regulatory genetic network.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Although various foot models were proposed for kinematics assessment using skin makers, no objective justification exists for the foot segmentations. This study proposed objective kinematic criteria to define which foot joints are relevant (dominant) in skin markers assessments. Among the studied joints, shank-hindfoot, hindfoot-midfoot and medial-lateral forefoot joints were found to have larger mobility than flexibility of their neighbour bonesets. The amplitude and pattern consistency of these joint angles confirmed their dominancy. Nevertheless, the consistency of the medial-lateral forefoot joint amplitude was lower. These three joints also showed acceptable sensibility to experimental errors which supported their dominancy. This study concluded that to be reliable for assessments using skin markers, the foot and ankle complex could be divided into shank, hindfoot, medial forefoot, lateral forefoot and toes. Kinematics of foot models with more segments must be more cautiously used.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Dans cette thèse, nous étudions les aspects comportementaux d'agents qui interagissent dans des systèmes de files d'attente à l'aide de modèles de simulation et de méthodologies expérimentales. Chaque période les clients doivent choisir un prestataire de servivce. L'objectif est d'analyser l'impact des décisions des clients et des prestataires sur la formation des files d'attente. Dans un premier cas nous considérons des clients ayant un certain degré d'aversion au risque. Sur la base de leur perception de l'attente moyenne et de la variabilité de cette attente, ils forment une estimation de la limite supérieure de l'attente chez chacun des prestataires. Chaque période, ils choisissent le prestataire pour lequel cette estimation est la plus basse. Nos résultats indiquent qu'il n'y a pas de relation monotone entre le degré d'aversion au risque et la performance globale. En effet, une population de clients ayant un degré d'aversion au risque intermédiaire encoure généralement une attente moyenne plus élevée qu'une population d'agents indifférents au risque ou très averses au risque. Ensuite, nous incorporons les décisions des prestataires en leur permettant d'ajuster leur capacité de service sur la base de leur perception de la fréquence moyenne d'arrivées. Les résultats montrent que le comportement des clients et les décisions des prestataires présentent une forte "dépendance au sentier". En outre, nous montrons que les décisions des prestataires font converger l'attente moyenne pondérée vers l'attente de référence du marché. Finalement, une expérience de laboratoire dans laquelle des sujets jouent le rôle de prestataire de service nous a permis de conclure que les délais d'installation et de démantèlement de capacité affectent de manière significative la performance et les décisions des sujets. En particulier, les décisions du prestataire, sont influencées par ses commandes en carnet, sa capacité de service actuellement disponible et les décisions d'ajustement de capacité qu'il a prises, mais pas encore implémentées. - Queuing is a fact of life that we witness daily. We all have had the experience of waiting in line for some reason and we also know that it is an annoying situation. As the adage says "time is money"; this is perhaps the best way of stating what queuing problems mean for customers. Human beings are not very tolerant, but they are even less so when having to wait in line for service. Banks, roads, post offices and restaurants are just some examples where people must wait for service. Studies of queuing phenomena have typically addressed the optimisation of performance measures (e.g. average waiting time, queue length and server utilisation rates) and the analysis of equilibrium solutions. The individual behaviour of the agents involved in queueing systems and their decision making process have received little attention. Although this work has been useful to improve the efficiency of many queueing systems, or to design new processes in social and physical systems, it has only provided us with a limited ability to explain the behaviour observed in many real queues. In this dissertation we differ from this traditional research by analysing how the agents involved in the system make decisions instead of focusing on optimising performance measures or analysing an equilibrium solution. This dissertation builds on and extends the framework proposed by van Ackere and Larsen (2004) and van Ackere et al. (2010). We focus on studying behavioural aspects in queueing systems and incorporate this still underdeveloped framework into the operations management field. In the first chapter of this thesis we provide a general introduction to the area, as well as an overview of the results. In Chapters 2 and 3, we use Cellular Automata (CA) to model service systems where captive interacting customers must decide each period which facility to join for service. They base this decision on their expectations of sojourn times. Each period, customers use new information (their most recent experience and that of their best performing neighbour) to form expectations of sojourn time at the different facilities. Customers update their expectations using an adaptive expectations process to combine their memory and their new information. We label "conservative" those customers who give more weight to their memory than to the xiv Summary new information. In contrast, when they give more weight to new information, we call them "reactive". In Chapter 2, we consider customers with different degree of risk-aversion who take into account uncertainty. They choose which facility to join based on an estimated upper-bound of the sojourn time which they compute using their perceptions of the average sojourn time and the level of uncertainty. We assume the same exogenous service capacity for all facilities, which remains constant throughout. We first analyse the collective behaviour generated by the customers' decisions. We show that the system achieves low weighted average sojourn times when the collective behaviour results in neighbourhoods of customers loyal to a facility and the customers are approximately equally split among all facilities. The lowest weighted average sojourn time is achieved when exactly the same number of customers patronises each facility, implying that they do not wish to switch facility. In this case, the system has achieved the Nash equilibrium. We show that there is a non-monotonic relationship between the degree of risk-aversion and system performance. Customers with an intermediate degree of riskaversion typically achieve higher sojourn times; in particular they rarely achieve the Nash equilibrium. Risk-neutral customers have the highest probability of achieving the Nash Equilibrium. Chapter 3 considers a service system similar to the previous one but with risk-neutral customers, and relaxes the assumption of exogenous service rates. In this sense, we model a queueing system with endogenous service rates by enabling managers to adjust the service capacity of the facilities. We assume that managers do so based on their perceptions of the arrival rates and use the same principle of adaptive expectations to model these perceptions. We consider service systems in which the managers' decisions take time to be implemented. Managers are characterised by a profile which is determined by the speed at which they update their perceptions, the speed at which they take decisions, and how coherent they are when accounting for their previous decisions still to be implemented when taking their next decision. We find that the managers' decisions exhibit a strong path-dependence: owing to the initial conditions of the model, the facilities of managers with identical profiles can evolve completely differently. In some cases the system becomes "locked-in" into a monopoly or duopoly situation. The competition between managers causes the weighted average sojourn time of the system to converge to the exogenous benchmark value which they use to estimate their desired capacity. Concerning the managers' profile, we found that the more conservative Summary xv a manager is regarding new information, the larger the market share his facility achieves. Additionally, the faster he takes decisions, the higher the probability that he achieves a monopoly position. In Chapter 4 we consider a one-server queueing system with non-captive customers. We carry out an experiment aimed at analysing the way human subjects, taking on the role of the manager, take decisions in a laboratory regarding the capacity of a service facility. We adapt the model proposed by van Ackere et al (2010). This model relaxes the assumption of a captive market and allows current customers to decide whether or not to use the facility. Additionally the facility also has potential customers who currently do not patronise it, but might consider doing so in the future. We identify three groups of subjects whose decisions cause similar behavioural patterns. These groups are labelled: gradual investors, lumpy investors, and random investor. Using an autocorrelation analysis of the subjects' decisions, we illustrate that these decisions are positively correlated to the decisions taken one period early. Subsequently we formulate a heuristic to model the decision rule considered by subjects in the laboratory. We found that this decision rule fits very well for those subjects who gradually adjust capacity, but it does not capture the behaviour of the subjects of the other two groups. In Chapter 5 we summarise the results and provide suggestions for further work. Our main contribution is the use of simulation and experimental methodologies to explain the collective behaviour generated by customers' and managers' decisions in queueing systems as well as the analysis of the individual behaviour of these agents. In this way, we differ from the typical literature related to queueing systems which focuses on optimising performance measures and the analysis of equilibrium solutions. Our work can be seen as a first step towards understanding the interaction between customer behaviour and the capacity adjustment process in queueing systems. This framework is still in its early stages and accordingly there is a large potential for further work that spans several research topics. Interesting extensions to this work include incorporating other characteristics of queueing systems which affect the customers' experience (e.g. balking, reneging and jockeying); providing customers and managers with additional information to take their decisions (e.g. service price, quality, customers' profile); analysing different decision rules and studying other characteristics which determine the profile of customers and managers.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Résumé Cette thèse est consacrée à l'analyse, la modélisation et la visualisation de données environnementales à référence spatiale à l'aide d'algorithmes d'apprentissage automatique (Machine Learning). L'apprentissage automatique peut être considéré au sens large comme une sous-catégorie de l'intelligence artificielle qui concerne particulièrement le développement de techniques et d'algorithmes permettant à une machine d'apprendre à partir de données. Dans cette thèse, les algorithmes d'apprentissage automatique sont adaptés pour être appliqués à des données environnementales et à la prédiction spatiale. Pourquoi l'apprentissage automatique ? Parce que la majorité des algorithmes d'apprentissage automatiques sont universels, adaptatifs, non-linéaires, robustes et efficaces pour la modélisation. Ils peuvent résoudre des problèmes de classification, de régression et de modélisation de densité de probabilités dans des espaces à haute dimension, composés de variables informatives spatialisées (« géo-features ») en plus des coordonnées géographiques. De plus, ils sont idéaux pour être implémentés en tant qu'outils d'aide à la décision pour des questions environnementales allant de la reconnaissance de pattern à la modélisation et la prédiction en passant par la cartographie automatique. Leur efficacité est comparable au modèles géostatistiques dans l'espace des coordonnées géographiques, mais ils sont indispensables pour des données à hautes dimensions incluant des géo-features. Les algorithmes d'apprentissage automatique les plus importants et les plus populaires sont présentés théoriquement et implémentés sous forme de logiciels pour les sciences environnementales. Les principaux algorithmes décrits sont le Perceptron multicouches (MultiLayer Perceptron, MLP) - l'algorithme le plus connu dans l'intelligence artificielle, le réseau de neurones de régression généralisée (General Regression Neural Networks, GRNN), le réseau de neurones probabiliste (Probabilistic Neural Networks, PNN), les cartes auto-organisées (SelfOrganized Maps, SOM), les modèles à mixture Gaussiennes (Gaussian Mixture Models, GMM), les réseaux à fonctions de base radiales (Radial Basis Functions Networks, RBF) et les réseaux à mixture de densité (Mixture Density Networks, MDN). Cette gamme d'algorithmes permet de couvrir des tâches variées telle que la classification, la régression ou l'estimation de densité de probabilité. L'analyse exploratoire des données (Exploratory Data Analysis, EDA) est le premier pas de toute analyse de données. Dans cette thèse les concepts d'analyse exploratoire de données spatiales (Exploratory Spatial Data Analysis, ESDA) sont traités selon l'approche traditionnelle de la géostatistique avec la variographie expérimentale et selon les principes de l'apprentissage automatique. La variographie expérimentale, qui étudie les relations entre pairs de points, est un outil de base pour l'analyse géostatistique de corrélations spatiales anisotropiques qui permet de détecter la présence de patterns spatiaux descriptible par une statistique. L'approche de l'apprentissage automatique pour l'ESDA est présentée à travers l'application de la méthode des k plus proches voisins qui est très simple et possède d'excellentes qualités d'interprétation et de visualisation. Une part importante de la thèse traite de sujets d'actualité comme la cartographie automatique de données spatiales. Le réseau de neurones de régression généralisée est proposé pour résoudre cette tâche efficacement. Les performances du GRNN sont démontrées par des données de Comparaison d'Interpolation Spatiale (SIC) de 2004 pour lesquelles le GRNN bat significativement toutes les autres méthodes, particulièrement lors de situations d'urgence. La thèse est composée de quatre chapitres : théorie, applications, outils logiciels et des exemples guidés. Une partie importante du travail consiste en une collection de logiciels : Machine Learning Office. Cette collection de logiciels a été développée durant les 15 dernières années et a été utilisée pour l'enseignement de nombreux cours, dont des workshops internationaux en Chine, France, Italie, Irlande et Suisse ainsi que dans des projets de recherche fondamentaux et appliqués. Les cas d'études considérés couvrent un vaste spectre de problèmes géoenvironnementaux réels à basse et haute dimensionnalité, tels que la pollution de l'air, du sol et de l'eau par des produits radioactifs et des métaux lourds, la classification de types de sols et d'unités hydrogéologiques, la cartographie des incertitudes pour l'aide à la décision et l'estimation de risques naturels (glissements de terrain, avalanches). Des outils complémentaires pour l'analyse exploratoire des données et la visualisation ont également été développés en prenant soin de créer une interface conviviale et facile à l'utilisation. Machine Learning for geospatial data: algorithms, software tools and case studies Abstract The thesis is devoted to the analysis, modeling and visualisation of spatial environmental data using machine learning algorithms. In a broad sense machine learning can be considered as a subfield of artificial intelligence. It mainly concerns with the development of techniques and algorithms that allow computers to learn from data. In this thesis machine learning algorithms are adapted to learn from spatial environmental data and to make spatial predictions. Why machine learning? In few words most of machine learning algorithms are universal, adaptive, nonlinear, robust and efficient modeling tools. They can find solutions for the classification, regression, and probability density modeling problems in high-dimensional geo-feature spaces, composed of geographical space and additional relevant spatially referenced features. They are well-suited to be implemented as predictive engines in decision support systems, for the purposes of environmental data mining including pattern recognition, modeling and predictions as well as automatic data mapping. They have competitive efficiency to the geostatistical models in low dimensional geographical spaces but are indispensable in high-dimensional geo-feature spaces. The most important and popular machine learning algorithms and models interesting for geo- and environmental sciences are presented in details: from theoretical description of the concepts to the software implementation. The main algorithms and models considered are the following: multi-layer perceptron (a workhorse of machine learning), general regression neural networks, probabilistic neural networks, self-organising (Kohonen) maps, Gaussian mixture models, radial basis functions networks, mixture density networks. This set of models covers machine learning tasks such as classification, regression, and density estimation. Exploratory data analysis (EDA) is initial and very important part of data analysis. In this thesis the concepts of exploratory spatial data analysis (ESDA) is considered using both traditional geostatistical approach such as_experimental variography and machine learning. Experimental variography is a basic tool for geostatistical analysis of anisotropic spatial correlations which helps to understand the presence of spatial patterns, at least described by two-point statistics. A machine learning approach for ESDA is presented by applying the k-nearest neighbors (k-NN) method which is simple and has very good interpretation and visualization properties. Important part of the thesis deals with a hot topic of nowadays, namely, an automatic mapping of geospatial data. General regression neural networks (GRNN) is proposed as efficient model to solve this task. Performance of the GRNN model is demonstrated on Spatial Interpolation Comparison (SIC) 2004 data where GRNN model significantly outperformed all other approaches, especially in case of emergency conditions. The thesis consists of four chapters and has the following structure: theory, applications, software tools, and how-to-do-it examples. An important part of the work is a collection of software tools - Machine Learning Office. Machine Learning Office tools were developed during last 15 years and was used both for many teaching courses, including international workshops in China, France, Italy, Ireland, Switzerland and for realizing fundamental and applied research projects. Case studies considered cover wide spectrum of the real-life low and high-dimensional geo- and environmental problems, such as air, soil and water pollution by radionuclides and heavy metals, soil types and hydro-geological units classification, decision-oriented mapping with uncertainties, natural hazards (landslides, avalanches) assessments and susceptibility mapping. Complementary tools useful for the exploratory data analysis and visualisation were developed as well. The software is user friendly and easy to use.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

OBJECTIVES: Inequalities and inequities in health are an important public health concern. In Switzerland, mortality in the general population varies according to the socio-economic position (SEP) of neighbourhoods. We examined the influence of neighbourhood SEP on presentation and outcomes in HIV-positive individuals in the era of combination antiretroviral therapy (cART). METHODS: The neighbourhood SEP of patients followed in the Swiss HIV Cohort Study (SHCS) 2000-2013 was obtained on the basis of 2000 census data on the 50 nearest households (education and occupation of household head, rent, mean number of persons per room). We used Cox and logistic regression models to examine the probability of late presentation, virologic response to cART, loss to follow-up and death across quintiles of neighbourhood SEP. RESULTS: A total of 4489 SHCS participants were included. Presentation with advanced disease [CD4 cell count <200 cells/μl or AIDS] and with AIDS was less common in neighbourhoods of higher SEP: the age and sex-adjusted odds ratio (OR) comparing the highest with the lowest quintile of SEP was 0.71 [95% confidence interval (95% CI) 0.58-0.87] and 0.59 (95% CI 0.45-0.77), respectively. An undetectable viral load at 6 months of cART was more common in the highest than in the lowest quintile (OR 1.52; 95% CI 1.14-2.04). Loss to follow-up, mortality and causes of death were not associated with neighbourhood SEP. CONCLUSION: Late presentation was more common and virologic response to cART less common in HIV-positive individuals living in neighbourhoods of lower SEP, but in contrast to the general population, there was no clear trend for mortality.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The paper deals with the development and application of the methodology for automatic mapping of pollution/contamination data. General Regression Neural Network (GRNN) is considered in detail and is proposed as an efficient tool to solve this problem. The automatic tuning of isotropic and an anisotropic GRNN model using cross-validation procedure is presented. Results are compared with k-nearest-neighbours interpolation algorithm using independent validation data set. Quality of mapping is controlled by the analysis of raw data and the residuals using variography. Maps of probabilities of exceeding a given decision level and ?thick? isoline visualization of the uncertainties are presented as examples of decision-oriented mapping. Real case study is based on mapping of radioactively contaminated territories.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

BACKGROUND: Previous studies on childhood cancer and nuclear power plants (NPPs) produced conflicting results. We used a cohort approach to examine whether residence near NPPs was associated with leukaemia or any childhood cancer in Switzerland. METHODS: We computed person-years at risk for children aged 0-15 years born in Switzerland from 1985 to 2009, based on the Swiss censuses 1990 and 2000 and identified cancer cases from the Swiss Childhood Cancer Registry. We geo-coded place of residence at birth and calculated incidence rate ratios (IRRs) with 95% confidence intervals (CIs) comparing the risk of cancer in children born <5 km, 5-10 km and 10-15 km from the nearest NPP with children born >15 km away, using Poisson regression models. RESULTS: We included 2925 children diagnosed with cancer during 21 117 524 person-years of follow-up; 953 (32.6%) had leukaemia. Eight and 12 children diagnosed with leukaemia at ages 0-4 and 0-15 years, and 18 and 31 children diagnosed with any cancer were born <5 km from a NPP. Compared with children born >15 km away, the IRRs (95% CI) for leukaemia in 0-4 and 0-15 year olds were 1.20 (0.60-2.41) and 1.05 (0.60-1.86), respectively. For any cancer, corresponding IRRs were 0.97 (0.61-1.54) and 0.89 (0.63-1.27). There was no evidence of a dose-response relationship with distance (P > 0.30). Results were similar for residence at diagnosis and at birth, and when adjusted for potential confounders. Results from sensitivity analyses were consistent with main results. CONCLUSIONS: This nationwide cohort study found little evidence of an association between residence near NPPs and the risk of leukaemia or any childhood cancer.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Natural selection favours the genes which are able to introduce replicates of themselves in the next generation with higher certainty than do rival genes (Hamilton 1963). The fitness of an individual, it?s ability to produce future parents, depends on it?s own behaviour as well as on the behaviour of other individuals in the population. For instance, the intensity of competition an individual experience depends on the exploitation of resources by neighbours. The fitness is thus frequency dependent on what neighbours do. Behaviours can be classified according to the costs and benefits they have on the fitness of the behaver and it?s neighbours (Hamilton 1964, Hamilton 1975). According to this classification there exist four distinct social behaviours. (1) A gene confering the ability to use a new ressource is called selfish because it has a positive e_ect on the bearer of the gene but a negative e_ect on neighbours by the concomitant increase in competition. (2) An altruistic behaviour is defined as an action where an individual increases the fitness of a neighbour at the expense of it?s own. The e_ect is deleterious for the actor but positive for the receptor. (3) More surprinsingly, an individual might sacrifice a fraction of it?s ressources to harm another at no direct benefits. This spitefull behaviour incurs a cost for the actor but is also deleterious for the receptor. (4) Finally a cooperative behaviour breeds benefits for both actors and neighbours. In this thesis I will continue on the path traced by numerous evolutionnary biologist which attempt to fine tune our understanding of the evolution of social behaviours since Hamilton?s foundation (1963, 1964). A critical development over the last 40 years has been the realisation that competition between kin can partly or completely cancel out the role of relatedness as an agent favouring altruism (Wilson et al., 1992; Taylor, 1992a,b). Of importance is thus to determine the scale at which competition and altruism occur. One mechanism avoiding the complete dilution of relatedness by competition is the conditionnal expression of the social behaviors. Focus will be given in this thesis at the role played by di_erent recognition mechanism in paving the way to altruism (Komdeur and Hatchwell, 1999) when the population has a spatial structure. Further, the evolution of spite will also be considered in these settings. The thesis is fractionated into two parts. First, di_erent models promoting altruism cooperation and spite will be compared under the same theoretical umbrella. This is a rather informal and more personnal part of my thesis. It also serve as a justification and basis to "Altruism among kin and non-kin individuals" which is an article attempting to clas- sify the mechanisms leading to altruism and cooperation. Second, in the annexe, there are three research papers about kin selection, altruism and dispersal: "Is sociality driven by the costs of dispersal or the benefits of philopatry?: A role for kin-discrimination mechanism", "Altruism, dispersal and phenotype kin recognition" and "Inbreeding avoidance through kin recognition: choosy female boost male dispersal" this last paper incorporates kin recognition as an agent favoring sex-biased dispersal.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Neurons and astrocytes, the two major cell populations in the adult brain, are characterized by their own mode of intercellular communication--the synapses and the gap junctions (GJ), respectively. In addition, there is increasing evidence for dynamic and metabolic neuroglial interactions resulting in the modulation of synaptic transmission at the so-called "tripartite synapse". Based on this, we have investigated at the ultrastructural level how excitatory synapses (ES) and astroglial GJ are spatially distributed in layer IV of the barrel cortex of the adult mouse. We used specific antibodies for connexin (Cx) 30 and 43 to identify astroglial GJ, these two proteins are known to be present in the majority of astroglial GJ in the cerebral cortex. In electron-microscopic images, we measured the distance between two ES, between two GJ and between a GJ and its nearest ES. We found a ratio of two GJ per three ES in the hollow and septal areas. Taking into account the size of an astrocyte domain, the high density of GJ suggests the occurrence of reflexive type, i.e. GJ between processes of the same astrocyte. Interestingly, the distance between an ES and an astroglial GJ was found to be significantly lower than that between either two synapses or between two GJ. These observations indicate that the two modes of cell-to-cell communication are not randomly distributed in layer IV of the barrel cortex. Consequently, this feature may provide the morphological support for the recently reported functional interactions between neuronal circuits and astroglial networks.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

We propose a novel formulation to solve the problem of intra-voxel reconstruction of the fibre orientation distribution function (FOD) in each voxel of the white matter of the brain from diffusion MRI data. The majority of the state-of-the-art methods in the field perform the reconstruction on a voxel-by-voxel level, promoting sparsity of the orientation distribution. Recent methods have proposed a global denoising of the diffusion data using spatial information prior to reconstruction, while others promote spatial regularisation through an additional empirical prior on the diffusion image at each q-space point. Our approach reconciles voxelwise sparsity and spatial regularisation and defines a spatially structured FOD sparsity prior, where the structure originates from the spatial coherence of the fibre orientation between neighbour voxels. The method is shown, through both simulated and real data, to enable accurate FOD reconstruction from a much lower number of q-space samples than the state of the art, typically 15 samples, even for quite adverse noise conditions.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

We consider the problem of multiple correlated sparse signals reconstruction and propose a new implementation of structured sparsity through a reweighting scheme. We present a particular application for diffusion Magnetic Resonance Imaging data and show how this procedure can be used for fibre orientation reconstruction in the white matter of the brain. In that framework, our structured sparsity prior can be used to exploit the fundamental coherence between fibre directions in neighbour voxels. Our method approaches the ℓ0 minimisation through a reweighted ℓ1-minimisation scheme. The weights are here defined in such a way to promote correlated sparsity between neighbour signals.