907 resultados para Integration of Programming Techniques
Resumo:
La compréhension de processus biologiques complexes requiert des approches expérimentales et informatiques sophistiquées. Les récents progrès dans le domaine des stratégies génomiques fonctionnelles mettent dorénavant à notre disposition de puissants outils de collecte de données sur l’interconnectivité des gènes, des protéines et des petites molécules, dans le but d’étudier les principes organisationnels de leurs réseaux cellulaires. L’intégration de ces connaissances au sein d’un cadre de référence en biologie systémique permettrait la prédiction de nouvelles fonctions de gènes qui demeurent non caractérisées à ce jour. Afin de réaliser de telles prédictions à l’échelle génomique chez la levure Saccharomyces cerevisiae, nous avons développé une stratégie innovatrice qui combine le criblage interactomique à haut débit des interactions protéines-protéines, la prédiction de la fonction des gènes in silico ainsi que la validation de ces prédictions avec la lipidomique à haut débit. D’abord, nous avons exécuté un dépistage à grande échelle des interactions protéines-protéines à l’aide de la complémentation de fragments protéiques. Cette méthode a permis de déceler des interactions in vivo entre les protéines exprimées par leurs promoteurs naturels. De plus, aucun biais lié aux interactions des membranes n’a pu être mis en évidence avec cette méthode, comparativement aux autres techniques existantes qui décèlent les interactions protéines-protéines. Conséquemment, nous avons découvert plusieurs nouvelles interactions et nous avons augmenté la couverture d’un interactome d’homéostasie lipidique dont la compréhension demeure encore incomplète à ce jour. Par la suite, nous avons appliqué un algorithme d’apprentissage afin d’identifier huit gènes non caractérisés ayant un rôle potentiel dans le métabolisme des lipides. Finalement, nous avons étudié si ces gènes et un groupe de régulateurs transcriptionnels distincts, non préalablement impliqués avec les lipides, avaient un rôle dans l’homéostasie des lipides. Dans ce but, nous avons analysé les lipidomes des délétions mutantes de gènes sélectionnés. Afin d’examiner une grande quantité de souches, nous avons développé une plateforme à haut débit pour le criblage lipidomique à contenu élevé des bibliothèques de levures mutantes. Cette plateforme consiste en la spectrométrie de masse à haute resolution Orbitrap et en un cadre de traitement des données dédié et supportant le phénotypage des lipides de centaines de mutations de Saccharomyces cerevisiae. Les méthodes expérimentales en lipidomiques ont confirmé les prédictions fonctionnelles en démontrant certaines différences au sein des phénotypes métaboliques lipidiques des délétions mutantes ayant une absence des gènes YBR141C et YJR015W, connus pour leur implication dans le métabolisme des lipides. Une altération du phénotype lipidique a également été observé pour une délétion mutante du facteur de transcription KAR4 qui n’avait pas été auparavant lié au métabolisme lipidique. Tous ces résultats démontrent qu’un processus qui intègre l’acquisition de nouvelles interactions moléculaires, la prédiction informatique des fonctions des gènes et une plateforme lipidomique innovatrice à haut débit , constitue un ajout important aux méthodologies existantes en biologie systémique. Les développements en méthodologies génomiques fonctionnelles et en technologies lipidomiques fournissent donc de nouveaux moyens pour étudier les réseaux biologiques des eucaryotes supérieurs, incluant les mammifères. Par conséquent, le stratégie présenté ici détient un potentiel d’application au sein d’organismes plus complexes.
Resumo:
La scoliose idiopathique de l’adolescent (SIA) est une déformation tri-dimensionelle du rachis. Son traitement comprend l’observation, l’utilisation de corsets pour limiter sa progression ou la chirurgie pour corriger la déformation squelettique et cesser sa progression. Le traitement chirurgical reste controversé au niveau des indications, mais aussi de la chirurgie à entreprendre. Malgré la présence de classifications pour guider le traitement de la SIA, une variabilité dans la stratégie opératoire intra et inter-observateur a été décrite dans la littérature. Cette variabilité s’accentue d’autant plus avec l’évolution des techniques chirurgicales et de l’instrumentation disponible. L’avancement de la technologie et son intégration dans le milieu médical a mené à l’utilisation d’algorithmes d’intelligence artificielle informatiques pour aider la classification et l’évaluation tridimensionnelle de la scoliose. Certains algorithmes ont démontré être efficace pour diminuer la variabilité dans la classification de la scoliose et pour guider le traitement. L’objectif général de cette thèse est de développer une application utilisant des outils d’intelligence artificielle pour intégrer les données d’un nouveau patient et les évidences disponibles dans la littérature pour guider le traitement chirurgical de la SIA. Pour cela une revue de la littérature sur les applications existantes dans l’évaluation de la SIA fut entreprise pour rassembler les éléments qui permettraient la mise en place d’une application efficace et acceptée dans le milieu clinique. Cette revue de la littérature nous a permis de réaliser que l’existence de “black box” dans les applications développées est une limitation pour l’intégration clinique ou la justification basée sur les évidence est essentielle. Dans une première étude nous avons développé un arbre décisionnel de classification de la scoliose idiopathique basé sur la classification de Lenke qui est la plus communément utilisée de nos jours mais a été critiquée pour sa complexité et la variabilité inter et intra-observateur. Cet arbre décisionnel a démontré qu’il permet d’augmenter la précision de classification proportionnellement au temps passé à classifier et ce indépendamment du niveau de connaissance sur la SIA. Dans une deuxième étude, un algorithme de stratégies chirurgicales basé sur des règles extraites de la littérature a été développé pour guider les chirurgiens dans la sélection de l’approche et les niveaux de fusion pour la SIA. Lorsque cet algorithme est appliqué à une large base de donnée de 1556 cas de SIA, il est capable de proposer une stratégie opératoire similaire à celle d’un chirurgien expert dans prêt de 70% des cas. Cette étude a confirmé la possibilité d’extraire des stratégies opératoires valides à l’aide d’un arbre décisionnel utilisant des règles extraites de la littérature. Dans une troisième étude, la classification de 1776 patients avec la SIA à l’aide d’une carte de Kohonen, un type de réseaux de neurone a permis de démontrer qu’il existe des scoliose typiques (scoliose à courbes uniques ou double thoracique) pour lesquelles la variabilité dans le traitement chirurgical varie peu des recommandations par la classification de Lenke tandis que les scolioses a courbes multiples ou tangentielles à deux groupes de courbes typiques étaient celles avec le plus de variation dans la stratégie opératoire. Finalement, une plateforme logicielle a été développée intégrant chacune des études ci-dessus. Cette interface logicielle permet l’entrée de données radiologiques pour un patient scoliotique, classifie la SIA à l’aide de l’arbre décisionnel de classification et suggère une approche chirurgicale basée sur l’arbre décisionnel de stratégies opératoires. Une analyse de la correction post-opératoire obtenue démontre une tendance, bien que non-statistiquement significative, à une meilleure balance chez les patients opérés suivant la stratégie recommandée par la plateforme logicielle que ceux aillant un traitement différent. Les études exposées dans cette thèse soulignent que l’utilisation d’algorithmes d’intelligence artificielle dans la classification et l’élaboration de stratégies opératoires de la SIA peuvent être intégrées dans une plateforme logicielle et pourraient assister les chirurgiens dans leur planification préopératoire.
Resumo:
This Thesis deals with the fabrication and characterization of novel all-fiber components for access networks. All fiber components offer distinctive advantages due to low forward and backward losses, epoxy free optical path and high power handling. A novel fabrication method for monolithic 1x4 couplers, which are vital components in distributed passive optical networks, is realized. The fabrication method differs from conventional structures with a symmetric coupling profile and hence offers ultra wideband performance and easy process control. New structure for 1x4 couplers, by fusing five fibers is proposed to achieve high uniformity, which gives equivalent uniformity performance to 1x4 planar lightwave splitters, isolation in fused fiber WDM is improved with integration of long period gratings. Packaging techniques of fused couplers are analyzed for long term stability.
Resumo:
Medical fields requires fast, simple and noninvasive methods of diagnostic techniques. Several methods are available and possible because of the growth of technology that provides the necessary means of collecting and processing signals. The present thesis details the work done in the field of voice signals. New methods of analysis have been developed to understand the complexity of voice signals, such as nonlinear dynamics aiming at the exploration of voice signals dynamic nature. The purpose of this thesis is to characterize complexities of pathological voice from healthy signals and to differentiate stuttering signals from healthy signals. Efficiency of various acoustic as well as non linear time series methods are analysed. Three groups of samples are used, one from healthy individuals, subjects with vocal pathologies and stuttering subjects. Individual vowels/ and a continuous speech data for the utterance of the sentence "iruvarum changatimaranu" the meaning in English is "Both are good friends" from Malayalam language are recorded using a microphone . The recorded audio are converted to digital signals and are subjected to analysis.Acoustic perturbation methods like fundamental frequency (FO), jitter, shimmer, Zero Crossing Rate(ZCR) were carried out and non linear measures like maximum lyapunov exponent(Lamda max), correlation dimension (D2), Kolmogorov exponent(K2), and a new measure of entropy viz., Permutation entropy (PE) are evaluated for all three groups of the subjects. Permutation Entropy is a nonlinear complexity measure which can efficiently distinguish regular and complex nature of any signal and extract information about the change in dynamics of the process by indicating sudden change in its value. The results shows that nonlinear dynamical methods seem to be a suitable technique for voice signal analysis, due to the chaotic component of the human voice. Permutation entropy is well suited due to its sensitivity to uncertainties, since the pathologies are characterized by an increase in the signal complexity and unpredictability. Pathological groups have higher entropy values compared to the normal group. The stuttering signals have lower entropy values compared to the normal signals.PE is effective in charaterising the level of improvement after two weeks of speech therapy in the case of stuttering subjects. PE is also effective in characterizing the dynamical difference between healthy and pathological subjects. This suggests that PE can improve and complement the recent voice analysis methods available for clinicians. The work establishes the application of the simple, inexpensive and fast algorithm of PE for diagnosis in vocal disorders and stuttering subjects.
Resumo:
Spike disease in sandal is generally diagnosed by the manifestation of external symptoms. Attempts have been made to detect the diseased plants by determining the length/breadth ratio of leaves (lyengar, 1961) and histochemical tests using Mann's stain (Parthasarathi et al., 1966), Dienes' stain (Ananthapadmanabha et a/., 1973) aniline blue and Hoechst 33258 (Ghosh et a/., 1985, Rangaswamy, 1995). But most of these techniques are insensitive, indirect detection methods leading to misinterpretation of results. Moreover, to identify disease resistant sandal trees, highly sensitive techniques are needed to detect the presence of the pathogen. In sandal forests, several host plants of sandal like Zizyphus oenop/ea (Fig. 1.3) also exhibit the yellows type disease symptoms. Immunological and molecular assays have to be developed to confirm the presence of sandal spike phytoplasma in such hosts. The major objectives of the present work includes:In situ detection of sandal spike phytoplasma by epifluorescence microscopy and scanning electron microscopy.,Purification of sandal spike phytoplasma and production of polyclonal antibodies.,Amino acid and total protein estimation of sandal spike phytoplasma.,Immunological detection of sandal spike phytoplasma., Molecular detection of sandal spike phytoplasma.,Screening for phytoplasma in host plants of spike disease affected sandal using immunological and molecular techniques.
Resumo:
In recent years, there is a visible trend for products/services which demand seamless integration of cellular networks, WLANs and WPANs. This is a strong indication for the inclusion of high speed short range wireless technology in future applications. In this context UWB radio has a significant role to play as an extension/complement to existing cellular/access technology. In the present work, three major types of ultra wide band planar antennas are investigated: Monopole and Slot. Three novel compact UWB antennas, suitable for poratble applications, are designed and characterized, namely 1) Ground modified monopole 2) Serrated monopole 3) Triangular slot The performance of these designs have been studied using standard simulation tools used in industry/academia and they have been experimentally verified. Antenna design guidelines are also deduced by accounting the resonances in each structure. In addition to having compact sized, high efficiency and broad bandwidth antennas, one of the major criterion in the design of impulse-UWB systems have been the transmission of narrow band pulses with minimum distortion. The key challenge is not only to design a broad band antenna with constant and stable gain but to maintain a flat group delay or linear phase response in the frequency domain or excellent transient response in time domain. One of the major contributions of the thesis lies in the analysis of the frequency and timedomain response of the designed UWB antennas to confirm their suitability for portable pulsed-UWB systems. Techniques to avoid narrowband interference by engraving narrow slot resonators on the antenna is also proposed and their effect on a nano-second pulse have been investigated
Resumo:
Speech is a natural mode of communication for people and speech recognition is an intensive area of research due to its versatile applications. This paper presents a comparative study of various feature extraction methods based on wavelets for recognizing isolated spoken words. Isolated words from Malayalam, one of the four major Dravidian languages of southern India are chosen for recognition. This work includes two speech recognition methods. First one is a hybrid approach with Discrete Wavelet Transforms and Artificial Neural Networks and the second method uses a combination of Wavelet Packet Decomposition and Artificial Neural Networks. Features are extracted by using Discrete Wavelet Transforms (DWT) and Wavelet Packet Decomposition (WPD). Training, testing and pattern recognition are performed using Artificial Neural Networks (ANN). The proposed method is implemented for 50 speakers uttering 20 isolated words each. The experimental results obtained show the efficiency of these techniques in recognizing speech
Resumo:
Knowledge discovery in databases is the non-trivial process of identifying valid, novel potentially useful and ultimately understandable patterns from data. The term Data mining refers to the process which does the exploratory analysis on the data and builds some model on the data. To infer patterns from data, data mining involves different approaches like association rule mining, classification techniques or clustering techniques. Among the many data mining techniques, clustering plays a major role, since it helps to group the related data for assessing properties and drawing conclusions. Most of the clustering algorithms act on a dataset with uniform format, since the similarity or dissimilarity between the data points is a significant factor in finding out the clusters. If a dataset consists of mixed attributes, i.e. a combination of numerical and categorical variables, a preferred approach is to convert different formats into a uniform format. The research study explores the various techniques to convert the mixed data sets to a numerical equivalent, so as to make it equipped for applying the statistical and similar algorithms. The results of clustering mixed category data after conversion to numeric data type have been demonstrated using a crime data set. The thesis also proposes an extension to the well known algorithm for handling mixed data types, to deal with data sets having only categorical data. The proposed conversion has been validated on a data set corresponding to breast cancer. Moreover, another issue with the clustering process is the visualization of output. Different geometric techniques like scatter plot, or projection plots are available, but none of the techniques display the result projecting the whole database but rather demonstrate attribute-pair wise analysis
Resumo:
The thesis explores the area of still image compression. The image compression techniques can be broadly classified into lossless and lossy compression. The most common lossy compression techniques are based on Transform coding, Vector Quantization and Fractals. Transform coding is the simplest of the above and generally employs reversible transforms like, DCT, DWT, etc. Mapped Real Transform (MRT) is an evolving integer transform, based on real additions alone. The present research work aims at developing new image compression techniques based on MRT. Most of the transform coding techniques employ fixed block size image segmentation, usually 8×8. Hence, a fixed block size transform coding is implemented using MRT and the merits and demerits are analyzed for both 8×8 and 4×4 blocks. The N2 unique MRT coefficients, for each block, are computed using templates. Considering the merits and demerits of fixed block size transform coding techniques, a hybrid form of these techniques is implemented to improve the performance of compression. The performance of the hybrid coder is found to be better compared to the fixed block size coders. Thus, if the block size is made adaptive, the performance can be further improved. In adaptive block size coding, the block size may vary from the size of the image to 2×2. Hence, the computation of MRT using templates is impractical due to memory requirements. So, an adaptive transform coder based on Unique MRT (UMRT), a compact form of MRT, is implemented to get better performance in terms of PSNR and HVS The suitability of MRT in vector quantization of images is then experimented. The UMRT based Classified Vector Quantization (CVQ) is implemented subsequently. The edges in the images are identified and classified by employing a UMRT based criteria. Based on the above experiments, a new technique named “MRT based Adaptive Transform Coder with Classified Vector Quantization (MATC-CVQ)”is developed. Its performance is evaluated and compared against existing techniques. A comparison with standard JPEG & the well-known Shapiro’s Embedded Zero-tree Wavelet (EZW) is done and found that the proposed technique gives better performance for majority of images
Resumo:
The evolution of coast through geological time scale is dependent on the transgression-regression event subsequent to the rise or fall of sea level. This event is accounted by investigation of the vertical sediment deposition patterns and their interrelationship for paleo-enviornmental reconstruction. Different methods like sedimentological (grain size and micro-morphological) and geochemical (elemental relationship) analyses as well as radiocarbon dating are generally used to decipher the sea level changes and paleoclimatic conditions of the Quaternary sediment sequence. For the Indian coast with a coastline length of about 7500 km, studies on geological and geomorphological signatures of sea level changes during the Quaternary were reported in general by researchers during the last two decades. However, for the southwest coast of India particularily Kerala which is famous for its coastal landforms comprising of estuaries, lagoons, backwaters, coastal plains, cliffs and barrier beaches, studies pertaining to the marine transgression-regression events in the southern region are limited. The Neendakara-Kayamkulam coastal stretch in central Kerala where the coast is manifested with shore parallel Kayamkulam Lagoon on one side and shore perpendicular Ashtamudi Estuary on the other side indicating existence of an uplifted prograded coastal margin followed by barrier beaches, backwater channels, ridge and runnel topography is an ideal site for studying such events. Hence the present study has been taken up in this context to address the gap area. The location for collection of core samples representing coastal plain, estuarylagoon and offshore regions have been identified based on published literature and available sedimentary records. The objectives of the research work are: To study the lithological variations and depositional environments of sediment cores along the coastal plain, estuary-lagoon and offshore regions between Kollam and Kayamkulam in the central Kerala coast To study the transportation and diagenetic history of sediments in the area To investigate the geochemical characterization of sediments and to elucidate the source-sink relationship To understand the marine transgression-regression events and to propose a conceptual model for the region The thesis comprises of 8 chapters. The first chapter embodies the preamble for the selection and significance of this research work. The study area is introduced with details on its physiographical, geological, geomorphological, rainfall and climate information. A review of literature, compiling the research on different aspects such as physico-chemical, geomorphological, tectonics, transgression-regression events are presented in the second chapter and they are broadly classified into three viz:- International, National and Kerala. The field data collection and laboratory analyses adopted in the research work are discussed in the third chapter. For collection of sediment core samples from the coastal plains, rotary drilling method was employed whereas for the estuary-lagoon and offshore locations the gravity/piston corer method was adopted. The collected subsurficial samples were analysed for texture, surface micro-texture, elemental analysis, XRD and radiocarbon dating techniques for age determination. The fourth chapter deals with the textural analysis of the core samples collected from various predefined locations of the study area. The result reveals that the Ashtamudi Estuary is composed of silty clay to clayey type of sediments whereas offshore cores are carpeted with silty clay to relict sand. Investigation of the source of sediments deposited in the coastal plain located on either side of the estuary indicates the dominance of terrigenous to marine origin in the southern region whereas it is predominantly of marine origin towards the north. Further the hydrodynamic conditions as well as the depositional enviornment of the sediment cores are elucidated based on statistical parameters that decipher the deposition pattern at various locations viz., coastal plain (open to closed basin), Ashtamudi Estuary (partially open to restricted estuary to closed basin) and offshore (open channel). The intensity of clay minerals is also discussed. From the results of radiocarbon dating the sediment depositional environments were deciphered.The results of the microtextural study of sediment samples (quartz grains) using Scanning Electron Microscope (SEM) are presented in the fifth chapter. These results throw light on the processes of transport and diagenetic history of the detrital sediments. Based on the lithological variations, selected quartz grains of different environments were also analysed. The study indicates that the southern coastal plain sediments were transported and deposited mechanically under fluvial environment followed by diagenesis under prolonged marine incursion. But in the case of the northern coastal plain, the sediments were transported and deposited under littoral environment indicating the dominance of marine incursion through mechanical as well as chemical processes. The quartz grains of the Ashtamudi Estuary indicate fluvial origin. The surface texture features of the offshore sediments suggest that the quartz grains are of littoral origin and represent the relict beach deposits. The geochemical characterisation of sediment cores based on geochemical classification, sediment maturity, palaeo-weathering and provenance in different environments are discussed in the sixth chapter. In the seventh chapter the integration of multiproxies data along with radiocarbon dates are presented and finally evolution and depositional history based on transgression–regression events is deciphered. The eighth chapter summarizes the major findings and conclusions of the study with recommendation for future work.
Resumo:
Each player in the financial industry, each bank, stock exchange, government agency, or insurance company operates its own financial information system or systems. By its very nature, financial information, like the money that it represents, changes hands. Therefore the interoperation of financial information systems is the cornerstone of the financial services they support. E-services frameworks such as web services are an unprecedented opportunity for the flexible interoperation of financial systems. Naturally the critical economic role and the complexity of financial information led to the development of various standards. Yet standards alone are not the panacea: different groups of players use different standards or different interpretations of the same standard. We believe that the solution lies in the convergence of flexible E-services such as web-services and semantically rich meta-data as promised by the semantic Web; then a mediation architecture can be used for the documentation, identification, and resolution of semantic conflicts arising from the interoperation of heterogeneous financial services. In this paper we illustrate the nature of the problem in the Electronic Bill Presentment and Payment (EBPP) industry and the viability of the solution we propose. We describe and analyze the integration of services using four different formats: the IFX, OFX and SWIFT standards, and an example proprietary format. To accomplish this integration we use the COntext INterchange (COIN) framework. The COIN architecture leverages a model of sources and receivers’ contexts in reference to a rich domain model or ontology for the description and resolution of semantic heterogeneity.
Resumo:
This paper presents the use of a mobile robot platform as an innovative educational tool in order to promote and integrate different curriculum knowledge. Hence, it is presented the acquired experience within a summer course named ldquoapplied mobile roboticsrdquo. The main aim of the course is to integrate different subjects as electronics, programming, architecture, perception systems, communications, control and trajectory planning by using the educational open mobile robot platform PRIM. The summer course is addressed to a wide range of student profiles. However, it is of special interests to the students of electrical and computer engineering around their final academic year. The summer course consists of the theoretical and laboratory sessions, related to the following topics: design & programming of electronic devices, modelling and control systems, trajectory planning and control, and computer vision systems. Therefore, the clues for achieving a renewed path of progress in robotics are the integration of several knowledgeable fields, such as computing, communications, and control sciences, in order to perform a higher level reasoning and use decision tools with strong theoretical base
Resumo:
Expert supervision systems are software applications specially designed to automate process monitoring. The goal is to reduce the dependency on human operators to assure the correct operation of a process including faulty situations. Construction of this kind of application involves an important task of design and development in order to represent and to manipulate process data and behaviour at different degrees of abstraction for interfacing with data acquisition systems connected to the process. This is an open problem that becomes more complex with the number of variables, parameters and relations to account for the complexity of the process. Multiple specialised modules tuned to solve simpler tasks that operate under a co-ordination provide a solution. A modular architecture based on concepts of software agents, taking advantage of the integration of diverse knowledge-based techniques, is proposed for this purpose. The components (software agents, communication mechanisms and perception/action mechanisms) are based on ICa (Intelligent Control architecture), software middleware supporting the build-up of applications with software agent features
Resumo:
La present tesi pretén recollir l'experiència viscuda en desenvolupar un sistema supervisor intel·ligent per a la millora de la gestió de plantes depuradores d'aigües residuals., implementar-lo en planta real (EDAR Granollers) i avaluar-ne el funcionament dia a dia amb situacions típiques de la planta. Aquest sistema supervisor combina i integra eines de control clàssic de les plantes depuradores (controlador automàtic del nivell d'oxigen dissolt al reactor biològic, ús de models descriptius del procés...) amb l'aplicació d'eines del camp de la intel·ligència artificial (sistemes basats en el coneixement, concretament sistemes experts i sistemes basats en casos, i xarxes neuronals). Aquest document s'estructura en 9 capítols diferents. Hi ha una primera part introductòria on es fa una revisió de l'estat actual del control de les EDARs i s'explica el perquè de la complexitat de la gestió d'aquests processos (capítol 1). Aquest capítol introductori juntament amb el capítol 2, on es pretén explicar els antecedents d'aquesta tesi, serveixen per establir els objectius d'aquest treball (capítol 3). A continuació, el capítol 4 descriu les peculiaritats i especificitats de la planta que s'ha escollit per implementar el sistema supervisor. Els capítols 5 i 6 del present document exposen el treball fet per a desenvolupar el sistema basat en regles o sistema expert (capítol 6) i el sistema basat en casos (capítol 7). El capítol 8 descriu la integració d'aquestes dues eines de raonament en una arquitectura multi nivell distribuïda. Finalment, hi ha una darrer capítol que correspon a la avaluació (verificació i validació), en primer lloc, de cadascuna de les eines per separat i, posteriorment, del sistema global en front de situacions reals que es donin a la depuradora
Resumo:
In the lead-up to the creation of a Eurasian Economic Union in 2015, the Customs Union and the Common Economic Space between Russia, Belarus and Kazakhstan represent two elements of the most ambitious regional integration project launched in the post-Soviet era since 1991. This CEPS Special Report examines both the potential and the limits of Eurasian economic integration. For the purpose of assessing the Eurasian integration process, CEPS applied a modified version of a framework first developed by Ernest B. Haas and Philippe C. Schmitter in 1964 to project whether economic integration of a group of countries automatically engenders political unity. Taking the data available for the early stages of the European integration process as a benchmark, the results for the Customs Union and the Common Economic Space point to a rather unfavourable outlook for Eurasian economic integration.