953 resultados para Models and Principles
Resumo:
In groundwater applications, Monte Carlo methods are employed to model the uncertainty on geological parameters. However, their brute-force application becomes computationally prohibitive for highly detailed geological descriptions, complex physical processes, and a large number of realizations. The Distance Kernel Method (DKM) overcomes this issue by clustering the realizations in a multidimensional space based on the flow responses obtained by means of an approximate (computationally cheaper) model; then, the uncertainty is estimated from the exact responses that are computed only for one representative realization per cluster (the medoid). Usually, DKM is employed to decrease the size of the sample of realizations that are considered to estimate the uncertainty. We propose to use the information from the approximate responses for uncertainty quantification. The subset of exact solutions provided by DKM is then employed to construct an error model and correct the potential bias of the approximate model. Two error models are devised that both employ the difference between approximate and exact medoid solutions, but differ in the way medoid errors are interpolated to correct the whole set of realizations. The Local Error Model rests upon the clustering defined by DKM and can be seen as a natural way to account for intra-cluster variability; the Global Error Model employs a linear interpolation of all medoid errors regardless of the cluster to which the single realization belongs. These error models are evaluated for an idealized pollution problem in which the uncertainty of the breakthrough curve needs to be estimated. For this numerical test case, we demonstrate that the error models improve the uncertainty quantification provided by the DKM algorithm and are effective in correcting the bias of the estimate computed solely from the MsFV results. The framework presented here is not specific to the methods considered and can be applied to other combinations of approximate models and techniques to select a subset of realizations
Resumo:
A general scheme for devising efficient cluster dynamics proposed in a previous paper [Phys. Rev. Lett. 72, 1541 (1994)] is extensively discussed. In particular, the strong connection among equilibrium properties of clusters and dynamic properties as the correlation time for magnetization is emphasized. The general scheme is applied to a number of frustrated spin models and the results discussed.
Resumo:
Bacterial reporters are live, genetically engineered cells with promising application in bioanalytics. They contain genetic circuitry to produce a cellular sensing element, which detects the target compound and relays the detection to specific synthesis of so-called reporter proteins (the presence or activity of which is easy to quantify). Bioassays with bacterial reporters are a useful complement to chemical analytics because they measure biological responses rather than total chemical concentrations. Simple bacterial reporter assays may also replace more costly chemical methods as a first line sample analysis technique. Recent promising developments integrate bacterial reporter cells with microsystems to produce bacterial biosensors. This lecture presents an in-depth treatment of the synthetic biological design principles of bacterial reporters, the engineering of which started as simple recombinant DNA puzzles, but has now become a more rational approach of choosing and combining sensing, controlling and reporting DNA 'parts'. Several examples of existing bacterial reporter designs and their genetic circuitry will be illustrated. Besides the design principles, the lecture also focuses on the application principles of bacterial reporter assays. A variety of assay formats will be illustrated, and principles of quantification will be dealt with. In addition to this discussion, substantial reference material is supplied in various Annexes.
Integrating species distribution models (SDMs) and phylogeography for two species of Alpine Primula.
Resumo:
The major intention of the present study was to investigate whether an approach combining the use of niche-based palaeodistribution modeling and phylo-geography would support or modify hypotheses about the Quaternary distributional history derived from phylogeographic methods alone. Our study system comprised two closely related species of Alpine Primula. We used species distribution models based on the extant distribution of the species and last glacial maximum (LGM) climate models to predict the distribution of the two species during the LGM. Phylogeographic data were generated using amplified fragment length polymorphisms (AFLPs). In Primula hirsuta, models of past distribution and phylogeographic data are partly congruent and support the hypothesis of widespread nunatak survival in the Central Alps. Species distribution models (SDMs) allowed us to differentiate between alpine regions that harbor potential nunatak areas and regions that have been colonized from other areas. SDMs revealed that diversity is a good indicator for nunataks, while rarity is a good indicator for peripheral relict populations that were not source for the recolonization of the inner Alps. In P. daonensis, palaeo-distribution models and phylogeographic data are incongruent. Besides the uncertainty inherent to this type of modeling approach (e.g., relatively coarse 1-km grain size), disagreement of models and data may partly be caused by shifts of ecological niche in both species. Nevertheless, we demonstrate that the combination of palaeo-distribution modeling with phylogeographical approaches provides a more differentiated picture of the distributional history of species and partly supports (P. hirsuta) and partly modifies (P. daonensis and P. hirsuta) hypotheses of Quaternary distributional history. Some of the refugial area indicated by palaeodistribution models could not have been identified with phylogeographic data.
Resumo:
Indirect topographic variables have been used successfully as surrogates for disturbance processes in plant species distribution models (SDM) in mountain environments. However, no SDM studies have directly tested the performance of disturbance variables. In this study, we developed two disturbance variables: a geomorphic index (GEO) and an index of snow redistribution by wind (SNOW). These were developed in order to assess how they improved both the fit and predictive power of presenceabsence SDM based on commonly used topoclimatic (TC) variables for 91 plants in the Western Swiss Alps. The individual contribution of the disturbance variables was compared to TC variables. Maps of models were prepared to spatially test the effect of disturbance variables. On average, disturbance variables significantly improved the fit but not the predictive power of the TC models and their individual contribution was weak (5.6% for GEO and 3.3% for SNOW). However their maximum individual contribution was important (24.7% and 20.7%). Finally, maps including disturbance variables (i) were significantly divergent from TC models in terms of predicted suitable surfaces and connectivity between potential habitats, and (ii) were interpreted as more ecologically relevant. Disturbance variables did not improve the transferability of models at the local scale in a complex mountain system, and the performance and contribution of these variables were highly species-specific. However, improved spatial projections and change in connectivity are important issues when preparing projections under climate change because the future range size of the species will determine the sensitivity to changing conditions.
Resumo:
At the beginning of the 1990s, the concept of "European integration" could still be said to be fairly unambiguous. Nowadays, it has become plural and complex almost to the point of unintelligibility. This is due, of course, to the internal differentiation of EU membership, with several Member States pulling out of key integrative projects such as establishing an area without frontiers, the "Schengen" area, and a common currency. But this is also due to the differentiated extension of key integrative projects to European non-EU countries - Schengen is again a case in point. Such processes of "integration without membership", the focus of the present publication, are acquiring an ever-growing topicality both in the political arena and in academia. International relations between the EU and its neighbouring countries are crucial for both, and their development through new agreements features prominently on the continent's political agenda. Over and above this aspect, the dissemination of EU values and standards beyond the Union's borders raises a whole host of theoretical and methodological questions, unsettling in some cases traditional conceptions of the autonomy and separation of national legal orders. This publication brings together the papers presented at the Integration without EU Membership workshop held in May 2008 at the EUI (Max Weber Programme and Department of Law). It aims to compare different models and experiences of integration between the EU, on the one hand, and those European countries that do not currently have an accession perspective on the other hand. In delimiting the geographical scope of the inquiry, so as to scale it down to manageable proportions, the guiding principles have been to include both the "Eastern" and "Western" neighbours of the EU, and to examine both structured frameworks of cooperation, such as the European Neighbourhood Policy and the European Economic Area, and bilateral relations developing on a more ad hoc basis. These principles are reflected in the arrangement of the papers, which consider in turn the positions of Ukraine, Russia, Norway, and Switzerland in European integration - current standing, perspectives for evolution, consequences in terms of the EU-ization of their respective legal orders1. These subjects are examined from several perspectives. We had the privilege of receiving contributions from leading practitioners and scholars from the countries concerned, from EU highranking officials, from prominent specialists in EU external relations law, and from young and talented researchers. We wish to thank them all here for their invaluable insights. We are moreover deeply indebted to Marise Cremona (EUI, Law Department, EUI) for her inspiring advice and encouragement, as well as to Ramon Marimon, Karin Tilmans, Lotte Holm, Alyson Price and Susan Garvin (Max Weber Programme, EUI) for their unflinching support throughout this project. A word is perhaps needed on the propriety and usefulness of the research concept embodied in this publication. Does it make sense to compare the integration models and experiences of countries as different as Norway, Russia, Switzerland, and Ukraine? Needless to say, this list of four evokes a staggering diversity of political, social, cultural, and economic conditions, and at least as great a diversity of approaches to European integration. Still, we would argue that such diversity only makes comparisons more meaningful. Indeed, while the particularities and idiosyncratic elements of each "model" of integration are fully displayed in the present volume, common themes and preoccupations run through the pages of every contribution: the difficulty in conceptualizing the finalité and essence of integration, which is evident in the EU today but which is greatly amplified for non-EU countries; the asymmetries and tradeoffs between integration and autonomy that are inherent in any attempt to participate in European integration from outside; the alteration of deeply seated legal concepts, and concepts about the law, that are already observable in the most integrated of the non-EU countries concerned. These issues are not transient or coincidental: they are inextricably bound up with the integration of non-EU countries in the EU project. By publishing this collection, we make no claim to have dealt with them in an exhaustive, still less in a definitive manner. Our ambition is more modest: to highlight the relevance of these themes, to place them more firmly on the scientific agenda, and to provide a stimulating basis for future research and reflection.
Resumo:
Characterizing the risks posed by nanomaterials is extraordinarily complex because these materials can have a wide range of sizes, shapes, chemical compositions and surface modifications, all of which may affect toxicity. There is an urgent need for a testing strategy that can rapidly and efficiently provide a screening approach for evaluating the potential hazard of nanomaterials and inform the prioritization of additional toxicological testing where necessary. Predictive toxicity models could form an integral component of such an approach by predicting which nanomaterials, as a result of their physico-chemical characteristics, have potentially hazardous properties. Strategies for directing research towards predictive models and the ancillary benefits of such research are presented here.
Resumo:
Les reconstructions palinspastiques fournissent le cadre idéal à de nombreuses études géologiques, géographiques, océanographique ou climatiques. En tant qu?historiens de la terre, les "reconstructeurs" essayent d?en déchiffrer le passé. Depuis qu?ils savent que les continents bougent, les géologues essayent de retracer leur évolution à travers les âges. Si l?idée originale de Wegener était révolutionnaire au début du siècle passé, nous savons depuis le début des années « soixante » que les continents ne "dérivent" pas sans but au milieu des océans mais sont inclus dans un sur-ensemble associant croûte « continentale » et « océanique »: les plaques tectoniques. Malheureusement, pour des raisons historiques aussi bien que techniques, cette idée ne reçoit toujours pas l'écho suffisant parmi la communauté des reconstructeurs. Néanmoins, nous sommes intimement convaincus qu?en appliquant certaines méthodes et certains principes il est possible d?échapper à l?approche "Wégenerienne" traditionnelle pour enfin tendre vers la tectonique des plaques. Le but principal du présent travail est d?exposer, avec tous les détails nécessaires, nos outils et méthodes. Partant des données paléomagnétiques et paléogéographiques classiquement utilisées pour les reconstructions, nous avons développé une nouvelle méthodologie replaçant les plaques tectoniques et leur cinématique au coeur du problème. En utilisant des assemblages continentaux (aussi appelés "assemblées clés") comme des points d?ancrage répartis sur toute la durée de notre étude (allant de l?Eocène jusqu?au Cambrien), nous développons des scénarios géodynamiques permettant de passer de l?une à l?autre en allant du passé vers le présent. Entre deux étapes, les plaques lithosphériques sont peu à peu reconstruites en additionnant/ supprimant les matériels océaniques (symbolisés par des isochrones synthétiques) aux continents. Excepté lors des collisions, les plaques sont bougées comme des entités propres et rigides. A travers les âges, les seuls éléments évoluant sont les limites de plaques. Elles sont préservées aux cours du temps et suivent une évolution géodynamique consistante tout en formant toujours un réseau interconnecté à travers l?espace. Cette approche appelée "limites de plaques dynamiques" intègre de multiples facteurs parmi lesquels la flottabilité des plaques, les taux d'accrétions aux rides, les courbes de subsidence, les données stratigraphiques et paléobiogéographiques aussi bien que les évènements tectoniques et magmatiques majeurs. Cette méthode offre ainsi un bon contrôle sur la cinématique des plaques et fournit de sévères contraintes au modèle. Cette approche "multi-source" nécessite une organisation et une gestion des données efficaces. Avant le début de cette étude, les masses de données nécessaires était devenues un obstacle difficilement surmontable. Les SIG (Systèmes d?Information Géographiques) et les géo-databases sont des outils informatiques spécialement dédiés à la gestion, au stockage et à l?analyse des données spatialement référencées et de leurs attributs. Grâce au développement dans ArcGIS de la base de données PaleoDyn nous avons pu convertir cette masse de données discontinues en informations géodynamiques précieuses et facilement accessibles pour la création des reconstructions. Dans le même temps, grâce à des outils spécialement développés, nous avons, tout à la fois, facilité le travail de reconstruction (tâches automatisées) et amélioré le modèle en développant fortement le contrôle cinématique par la création de modèles de vitesses des plaques. Sur la base des 340 terranes nouvellement définis, nous avons ainsi développé un set de 35 reconstructions auxquelles est toujours associé un modèle de vitesse. Grâce à cet ensemble de données unique, nous pouvons maintenant aborder des problématiques majeurs de la géologie moderne telles que l?étude des variations du niveau marin et des changements climatiques. Nous avons commencé par aborder un autre problème majeur (et non définitivement élucidé!) de la tectonique moderne: les mécanismes contrôlant les mouvements des plaques. Nous avons pu observer que, tout au long de l?histoire de la terre, les pôles de rotation des plaques (décrivant les mouvements des plaques à la surface de la terre) tendent à se répartir le long d'une bande allant du Pacifique Nord au Nord de l'Amérique du Sud, l'Atlantique Central, l'Afrique du Nord, l'Asie Centrale jusqu'au Japon. Fondamentalement, cette répartition signifie que les plaques ont tendance à fuir ce plan médian. En l'absence d'un biais méthodologique que nous n'aurions pas identifié, nous avons interprété ce phénomène comme reflétant l'influence séculaire de la Lune sur le mouvement des plaques. La Lune sur le mouvement des plaques. Le domaine océanique est la clé de voute de notre modèle. Nous avons attaché un intérêt tout particulier à le reconstruire avec beaucoup de détails. Dans ce modèle, la croûte océanique est préservée d?une reconstruction à l?autre. Le matériel crustal y est symbolisé sous la forme d?isochrones synthétiques dont nous connaissons les âges. Nous avons également reconstruit les marges (actives ou passives), les rides médio-océaniques et les subductions intra-océaniques. En utilisant ce set de données très détaillé, nous avons pu développer des modèles bathymétriques 3-D unique offrant une précision bien supérieure aux précédents.<br/><br/>Palinspastic reconstructions offer an ideal framework for geological, geographical, oceanographic and climatology studies. As historians of the Earth, "reconstructers" try to decipher the past. Since they know that continents are moving, geologists a trying to retrieve the continents distributions through ages. If Wegener?s view of continent motions was revolutionary at the beginning of the 20th century, we know, since the Early 1960?s that continents are not drifting without goal in the oceanic realm but are included in a larger set including, all at once, the oceanic and the continental crust: the tectonic plates. Unfortunately, mainly due to technical and historical issues, this idea seems not to receive a sufficient echo among our particularly concerned community. However, we are intimately convinced that, by applying specific methods and principles we can escape the traditional "Wegenerian" point of view to, at last, reach real plate tectonics. This is the main aim of this study to defend this point of view by exposing, with all necessary details, our methods and tools. Starting with the paleomagnetic and paleogeographic data classically used in reconstruction studies, we developed a modern methodology placing the plates and their kinematics at the centre of the issue. Using assemblies of continents (referred as "key assemblies") as anchors distributed all along the scope of our study (ranging from Eocene time to Cambrian time) we develop geodynamic scenarios leading from one to the next, from the past to the present. In between, lithospheric plates are progressively reconstructed by adding/removing oceanic material (symbolized by synthetic isochrones) to major continents. Except during collisions, plates are moved as single rigid entities. The only evolving elements are the plate boundaries which are preserved and follow a consistent geodynamical evolution through time and form an interconnected network through space. This "dynamic plate boundaries" approach integrates plate buoyancy factors, oceans spreading rates, subsidence patterns, stratigraphic and paleobiogeographic data, as well as major tectonic and magmatic events. It offers a good control on plate kinematics and provides severe constraints for the model. This multi-sources approach requires an efficient data management. Prior to this study, the critical mass of necessary data became a sorely surmountable obstacle. GIS and geodatabases are modern informatics tools of specifically devoted to store, analyze and manage data and associated attributes spatially referenced on the Earth. By developing the PaleoDyn database in ArcGIS software we converted the mass of scattered data offered by the geological records into valuable geodynamical information easily accessible for reconstructions creation. In the same time, by programming specific tools we, all at once, facilitated the reconstruction work (tasks automation) and enhanced the model (by highly increasing the kinematic control of plate motions thanks to plate velocity models). Based on the 340 terranes properly defined, we developed a revised set of 35 reconstructions associated to their own velocity models. Using this unique dataset we are now able to tackle major issues of the geology (such as the global sea-level variations and climate changes). We started by studying one of the major unsolved issues of the modern plate tectonics: the driving mechanism of plate motions. We observed that, all along the Earth?s history, plates rotation poles (describing plate motions across the Earth?s surface) tend to follow a slight linear distribution along a band going from the Northern Pacific through Northern South-America, Central Atlantic, Northern Africa, Central Asia up to Japan. Basically, it sighifies that plates tend to escape this median plan. In the absence of a non-identified methodological bias, we interpreted it as the potential secular influence ot the Moon on plate motions. The oceanic realms are the cornerstone of our model and we attached a particular interest to reconstruct them with many details. In this model, the oceanic crust is preserved from one reconstruction to the next. The crustal material is symbolised by the synthetic isochrons from which we know the ages. We also reconstruct the margins (active or passive), ridges and intra-oceanic subductions. Using this detailed oceanic dataset, we developed unique 3-D bathymetric models offering a better precision than all the previously existing ones.
Resumo:
Background: Understanding the relationship between gene expression changes, enzyme activity shifts, and the corresponding physiological adaptive response of organisms to environmental cues is crucial in explaining how cells cope with stress. For example, adaptation of yeast to heat shock involves a characteristic profile of changes to the expression levels of genes coding for enzymes of the glycolytic pathway and some of its branches. The experimental determination of changes in gene expression profiles provides a descriptive picture of the adaptive response to stress. However, it does not explain why a particular profile is selected for any given response. Results: We used mathematical models and analysis of in silico gene expression profiles (GEPs) to understand how changes in gene expression correlate to an efficient response of yeast cells to heat shock. An exhaustive set of GEPs, matched with the corresponding set of enzyme activities, was simulated and analyzed. The effectiveness of each profile in the response to heat shock was evaluated according to relevant physiological and functional criteria. The small subset of GEPs that lead to effective physiological responses after heat shock was identified as the result of the tuning of several evolutionary criteria. The experimentally observed transcriptional changes in response to heat shock belong to this set and can be explained by quantitative design principles at the physiological level that ultimately constrain changes in gene expression. Conclusion: Our theoretical approach suggests a method for understanding the combined effect of changes in the expression of multiple genes on the activity of metabolic pathways, and consequently on the adaptation of cellular metabolism to heat shock. This method identifies quantitative design principles that facilitate understating the response of the cell to stress.
Resumo:
Tämä työ luo katsauksen ajallisiin ja stokastisiin ohjelmien luotettavuus malleihin sekä tutkii muutamia malleja käytännössä. Työn teoriaosuus sisältää ohjelmien luotettavuuden kuvauksessa ja arvioinnissa käytetyt keskeiset määritelmät ja metriikan sekä varsinaiset mallien kuvaukset. Työssä esitellään kaksi ohjelmien luotettavuusryhmää. Ensimmäinen ryhmä ovat riskiin perustuvat mallit. Toinen ryhmä käsittää virheiden ”kylvöön” ja merkitsevyyteen perustuvat mallit. Työn empiirinen osa sisältää kokeiden kuvaukset ja tulokset. Kokeet suoritettiin käyttämällä kolmea ensimmäiseen ryhmään kuuluvaa mallia: Jelinski-Moranda mallia, ensimmäistä geometrista mallia sekä yksinkertaista eksponenttimallia. Kokeiden tarkoituksena oli tutkia, kuinka syötetyn datan distribuutio vaikuttaa mallien toimivuuteen sekä kuinka herkkiä mallit ovat syötetyn datan määrän muutoksille. Jelinski-Moranda malli osoittautui herkimmäksi distribuutiolle konvergaatio-ongelmien vuoksi, ensimmäinen geometrinen malli herkimmäksi datan määrän muutoksille.
Resumo:
Alpine tree-line ecotones are characterized by marked changes at small spatial scales that may result in a variety of physiognomies. A set of alternative individual-based models was tested with data from four contrasting Pinus uncinata ecotones in the central Spanish Pyrenees to reveal the minimal subset of processes required for tree-line formation. A Bayesian approach combined with Markov chain Monte Carlo methods was employed to obtain the posterior distribution of model parameters, allowing the use of model selection procedures. The main features of real tree lines emerged only in models considering nonlinear responses in individual rates of growth or mortality with respect to the altitudinal gradient. Variation in tree-line physiognomy reflected mainly changes in the relative importance of these nonlinear responses, while other processes, such as dispersal limitation and facilitation, played a secondary role. Different nonlinear responses also determined the presence or absence of krummholz, in agreement with recent findings highlighting a different response of diffuse and abrupt or krummholz tree lines to climate change. The method presented here can be widely applied in individual-based simulation models and will turn model selection and evaluation in this type of models into a more transparent, effective, and efficient exercise.
Resumo:
Three-dimensional reconstruction of reservoir analogues can be improved combining data from different geophysical methods. Ground Penetrating Radar (GPR) and Electrical Resistivity Tomography (ERT) data are valuable tools, since they provide subsurface information from internal architecture and facies distribution of sedimentary rock bodies, enabling the upgrading of depositional models and heterogeneity reconstruction. The Lower Eocene Roda Sandstone is a well-known deltaic complex widely studied as a reservoir analogue that displays a series of sandstone wedges with a general NE to SW progradational trend. To provide a better understanding of internal heterogeneity of a 10m-thick progradational delta-front sandstone unit, 3D GPR data were acquired. In addition, common midpoints (CMP) to measure the sandstone subsoil velocity, test profiles with different frequency antennas (25, 50 and 100MHz) and topographic data for subsequent correction in the geophysical data were also obtained. Three ERT profiles were also acquired to further constrain GPR analysis. These geophysical results illustrate the geometry of reservoir analogue heterogeneities both depositional and diagenetic in nature, improving and complementing previous outcrop-derived data. GPR interpretation using radar stratigraphy principles and attributes analysis provided: 1)tridimensional geometry of major stratigraphic surfaces that define four units in the GPR Prism, 2) image the internal architecture of the units and their statistical study of azimuth and dips, useful for a quick determination of paleocurrent directions. These results were used to define the depositional architecture of the progradational sandbody that shows an arrangement in very-high-frequency sequences characterized by clockwise paleocurrent variations and decrease of the sedimentary flow, similar to those observed at a greater scale in the same system. This high-frequency sequential arrangement has been attributed to the autocyclic dynamics of a supply-dominated delta- front where fluvial and tidal currents are in competition. The resistivity models enhanced the viewing of reservoir quality associated with cement distribution caused by depositional and early diagenetic processes related to the development of transgressive and regressive systems tracts in igh-frequency sequences.
Resumo:
The use of private funding and management is enjoying an increasing trend in airports. The literature has not paid enough attention to the mixed management models in this industry, although many European airports take the form of mixed public-private companies, where ownership is shared between public and private sectors. We examine the determinants of the degree of private participation in the European airport sector. Drawing on a sample of the 100 largest European airports, we estimate a multivariate equation in order to determine the role of airport characteristics, fiscal variables, and political factors on the extent of private involvement. Our results confirm the alignment between public and private interests in partially privatized airports. Fiscal constraints and market attractiveness promote private participation. Integrated governance models and the share of network carriers prevent the presence of private ownership, while the degree of private participation appears to be pragmatic rather than ideological.
Resumo:
The use of private funding and management enjoys an increasing trend in airports. The literature has not paid enough attention to the mixed management models in this industry, although many European airports take the form of mixed firms or Institutional PPP, where ownership is shared between public and private sectors. We examine the determinants of the degree of private participation in the European airport sector. Drawing on a sample of the 100 largest European airports we estimate a multivariate equation in order to determine the role of airport characteristics, fiscal variables and political factors on the extent of private involvement. Our results confirm the alignment between public and private interests in PPPs. Fiscal constraints and market attractiveness promote private participation. Integrated governance models and the share of network carriers prevent the presence of private ownership, while the degree of private participation appears to be pragmatic rather than ideological.
Resumo:
The present study was performed in an attempt to develop an in vitro integrated testing strategy (ITS) to evaluate drug-induced neurotoxicity. A number of endpoints were analyzed using two complementary brain cell culture models and an in vitro blood-brain barrier (BBB) model after single and repeated exposure treatments with selected drugs that covered the major biological, pharmacological and neuro-toxicological responses. Furthermore, four drugs (diazepam, cyclosporine A, chlorpromazine and amiodarone) were tested more in depth as representatives of different classes of neurotoxicants, inducing toxicity through different pathways of toxicity. The developed in vitro BBB model allowed detection of toxic effects at the level of BBB and evaluation of drug transport through the barrier for predicting free brain concentrations of the studied drugs. The measurement of neuronal electrical activity was found to be a sensitive tool to predict the neuroactivity and neurotoxicity of drugs after acute exposure. The histotypic 3D re-aggregating brain cell cultures, containing all brain cell types, were found to be well suited for OMICs analyses after both acute and long term treatment. The obtained data suggest that an in vitro ITS based on the information obtained from BBB studies and combined with metabolomics, proteomics and neuronal electrical activity measurements performed in stable in vitro neuronal cell culture systems, has high potential to improve current in vitro drug-induced neurotoxicity evaluation.