839 resultados para Time-sharing computer systems
Resumo:
The European Space Agency's Gaia mission will create the largest and most precise three dimensional chart of our galaxy (the Milky Way), by providing unprecedented position, parallax, proper motion, and radial velocity measurements for about one billion stars. The resulting catalogue will be made available to the scientific community and will be analyzed in many different ways, including the production of a variety of statistics. The latter will often entail the generation of multidimensional histograms and hypercubes as part of the precomputed statistics for each data release, or for scientific analysis involving either the final data products or the raw data coming from the satellite instruments. In this paper we present and analyze a generic framework that allows the hypercube generation to be easily done within a MapReduce infrastructure, providing all the advantages of the new Big Data analysis paradigmbut without dealing with any specific interface to the lower level distributed system implementation (Hadoop). Furthermore, we show how executing the framework for different data storage model configurations (i.e. row or column oriented) and compression techniques can considerably improve the response time of this type of workload for the currently available simulated data of the mission. In addition, we put forward the advantages and shortcomings of the deployment of the framework on a public cloud provider, benchmark against other popular solutions available (that are not always the best for such ad-hoc applications), and describe some user experiences with the framework, which was employed for a number of dedicated astronomical data analysis techniques workshops.
Resumo:
The European Space Agency's Gaia mission will create the largest and most precise three dimensional chart of our galaxy (the Milky Way), by providing unprecedented position, parallax, proper motion, and radial velocity measurements for about one billion stars. The resulting catalogue will be made available to the scientific community and will be analyzed in many different ways, including the production of a variety of statistics. The latter will often entail the generation of multidimensional histograms and hypercubes as part of the precomputed statistics for each data release, or for scientific analysis involving either the final data products or the raw data coming from the satellite instruments. In this paper we present and analyze a generic framework that allows the hypercube generation to be easily done within a MapReduce infrastructure, providing all the advantages of the new Big Data analysis paradigmbut without dealing with any specific interface to the lower level distributed system implementation (Hadoop). Furthermore, we show how executing the framework for different data storage model configurations (i.e. row or column oriented) and compression techniques can considerably improve the response time of this type of workload for the currently available simulated data of the mission. In addition, we put forward the advantages and shortcomings of the deployment of the framework on a public cloud provider, benchmark against other popular solutions available (that are not always the best for such ad-hoc applications), and describe some user experiences with the framework, which was employed for a number of dedicated astronomical data analysis techniques workshops.
Resumo:
In this work, we present an integral scheduling system for non-dedicated clusters, termed CISNE-P, which ensures the performance required by the local applications, while simultaneously allocating cluster resources to parallel jobs. Our approach solves the problem efficiently by using a social contract technique. This kind of technique is based on reserving computational resources, preserving a predetermined response time to local users. CISNE-P is a middleware which includes both a previously developed space-sharing job scheduler and a dynamic coscheduling system, a time sharing scheduling component. The experimentation performed in a Linux cluster shows that these two scheduler components are complementary and a good coordination improves global performance significantly. We also compare two different CISNE-P implementations: one developed inside the kernel, and the other entirely implemented in the user space.
Resumo:
The advent of the Internet had a great impact on distance education and rapidly e-learning has become a killer application. Education institutions worldwide are taking advantage of the available technology in order to facilitate education to a growing audience. Everyday, more and more people use e-learning systems, environments and contents for both training and learning. E-learning promotes educationamong people that due to different reasons could not have access to education: people who could nottravel, people with very little free time, or withdisabilities, etc. As e-learning systems grow and more people are accessing them, it is necessary to consider when designing virtual environments the diverse needs and characteristics that different users have. This allows building systems that people can use easily, efficiently and effectively, where the learning process leads to a good user experience and becomes a good learning experience.
Resumo:
Network virtualisation is considerably gaining attentionas a solution to ossification of the Internet. However, thesuccess of network virtualisation will depend in part on how efficientlythe virtual networks utilise substrate network resources.In this paper, we propose a machine learning-based approachto virtual network resource management. We propose to modelthe substrate network as a decentralised system and introducea learning algorithm in each substrate node and substrate link,providing self-organization capabilities. We propose a multiagentlearning algorithm that carries out the substrate network resourcemanagement in a coordinated and decentralised way. The taskof these agents is to use evaluative feedback to learn an optimalpolicy so as to dynamically allocate network resources to virtualnodes and links. The agents ensure that while the virtual networkshave the resources they need at any given time, only the requiredresources are reserved for this purpose. Simulations show thatour dynamic approach significantly improves the virtual networkacceptance ratio and the maximum number of accepted virtualnetwork requests at any time while ensuring that virtual networkquality of service requirements such as packet drop rate andvirtual link delay are not affected.
Resumo:
The objective of this project was to develop firmware for both the Arduino-compatible boards of the Smart Citizen initiative, and for the RTX4100 low-power WiFi.
Resumo:
Distributed storage systems are studied. The interest in such system has become relatively wide due to the increasing amount of information needed to be stored in data centers or different kinds of cloud systems. There are many kinds of solutions for storing the information into distributed devices regarding the needs of the system designer. This thesis studies the questions of designing such storage systems and also fundamental limits of such systems. Namely, the subjects of interest of this thesis include heterogeneous distributed storage systems, distributed storage systems with the exact repair property, and locally repairable codes. For distributed storage systems with either functional or exact repair, capacity results are proved. In the case of locally repairable codes, the minimum distance is studied. Constructions for exact-repairing codes between minimum bandwidth regeneration (MBR) and minimum storage regeneration (MSR) points are given. These codes exceed the time-sharing line of the extremal points in many cases. Other properties of exact-regenerating codes are also studied. For the heterogeneous setup, the main result is that the capacity of such systems is always smaller than or equal to the capacity of a homogeneous system with symmetric repair with average node size and average repair bandwidth. A randomized construction for a locally repairable code with good minimum distance is given. It is shown that a random linear code of certain natural type has a good minimum distance with high probability. Other properties of locally repairable codes are also studied.
Resumo:
Please consult the paper edition of this thesis to read. It is available on the 5th Floor of the Library at Call Number: Z 9999 E38 D56 1992
Resumo:
Les émotions jouent un rôle primordial dans les processus cognitifs et plus particulièrement dans les tâches d’apprentissage. D’ailleurs, plusieurs recherches neurologiques ont montré l’interrelation qui existe entre la cognition et les émotions. Elles ont aussi déterminé plusieurs formes d’intelligence humaine autre que l’intelligence rationnelle parmi lesquelles nous distinguons la forme ayant comme dimension émotionnelle, à savoir l’intelligence émotionnelle, vu son impact sur les processus d’apprentissage. L’intelligence émotionnelle est alors un facteur significatif de réussite scolaire et professionnelle. Sous la lumière de ces constatations présentées, les informaticiens à leur tour, vont alors tenter de consentir de plus en plus de place au facteur émotionnel dans les systèmes informatiques et plus particulièrement dans ceux dédiés à l’apprentissage. L’intégration de l’intelligence émotionnelle dans ces systèmes et plus précisément, dans les Systèmes Tutoriels Intelligents (STI), va leur permettre de gérer les émotions de l’apprenant et par la suite améliorer ses performances. Dans ce mémoire, notre objectif principal est d’élaborer une stratégie d’apprentissage visant à favoriser et accentuer la mémorisation chez les enfants. Pour atteindre cet objectif, nous avons développé un cours d’anglais en ligne ainsi qu’un tuteur virtuel utilisant des ressources multimédia tels que le ton de la voix, la musique, les images et les gestes afin de susciter des émotions chez l’apprenant. Nous avons conduit une expérience pour tester l’efficacité de quelques stratégies émotionnelles ainsi qu’évaluer l’impact des émotions suscitées sur la capacité de mémorisation des connaissances à acquérir par l’apprenant. Les résultats de cette étude expérimentale ont prouvé que l’induction implicite des émotions chez ce dernier a une influence significative sur ses performances. Ils ont également montré qu’il n’existe pas une stratégie efficace pour tous les apprenants à la fois, cependant l’efficacité d’une telle stratégie par rapport à une autre dépend essentiellement du profil comportemental de l’apprenant déterminé à partir de son tempérament.
Resumo:
L’avancée des infrastructures informatiques a permis l’émergence de la modélisation moléculaire. À cet effet, une multitude de modèles mathématiques sont aujourd’hui disponibles pour simuler différents systèmes chimiques. À l’aide de la modélisation moléculaire, différents types d’interactions chimiques ont été observés. À partir des systèmes les plus simples permettant l’utilisation de modèles quantiques rigoureux, une série d’approximations a été considérée pour rendre envisageable la simulation de systèmes moléculaires de plus en plus complexes. En premier lieu, la théorie de la fonctionnelle de densité dépendante du temps a été utilisée pour simuler les énergies d’excitation de molécules photoactives. De manière similaire, la DFT indépendante du temps a permis la simulation du pont hydrogène intramoléculaire de structures analogues au 1,3,5-triazapentadiène et la rationalisation de la stabilité des états de transition. Par la suite, la dynamique moléculaire et la mécanique moléculaire ont permis de simuler les interactions d’un trimère d’acide cholique et d’un pyrène dans différents solvants. Cette même méthodologie a été utilisée pour simuler les interactions d’un rotaxane-parapluie à l’interface d’un système biphasique. Finalement, l’arrimage moléculaire et les fonctions de score ont été utilisés pour simuler les interactions intermoléculaires entre une protéine et des milliers de candidats moléculaires. Les résultats ont permis de mettre en place une stratégie de développement d’un nouvel inhibiteur enzymatique.
Resumo:
Modern computer systems are plagued with stability and security problems: applications lose data, web servers are hacked, and systems crash under heavy load. Many of these problems or anomalies arise from rare program behavior caused by attacks or errors. A substantial percentage of the web-based attacks are due to buffer overflows. Many methods have been devised to detect and prevent anomalous situations that arise from buffer overflows. The current state-of-art of anomaly detection systems is relatively primitive and mainly depend on static code checking to take care of buffer overflow attacks. For protection, Stack Guards and I-leap Guards are also used in wide varieties.This dissertation proposes an anomaly detection system, based on frequencies of system calls in the system call trace. System call traces represented as frequency sequences are profiled using sequence sets. A sequence set is identified by the starting sequence and frequencies of specific system calls. The deviations of the current input sequence from the corresponding normal profile in the frequency pattern of system calls is computed and expressed as an anomaly score. A simple Bayesian model is used for an accurate detection.Experimental results are reported which show that frequency of system calls represented using sequence sets, captures the normal behavior of programs under normal conditions of usage. This captured behavior allows the system to detect anomalies with a low rate of false positives. Data are presented which show that Bayesian Network on frequency variations responds effectively to induced buffer overflows. It can also help administrators to detect deviations in program flow introduced due to errors.
Resumo:
A program that simulates a Digital Equipment Corporation PDP-11 computer and many of its peripherals on the AI Laboratory Time Sharing System (ITS) is described from a user's reference point of view. This simulator has a built in DDT-like command level which provides the user with the normal range of DDT facilities but also with several special debugging features built into the simulator. The DDT command language was implemented by Richard M. Stallman while the simulator was written by the author of this memo.
Resumo:
The memory hierarchy is the main bottleneck in modern computer systems as the gap between the speed of the processor and the memory continues to grow larger. The situation in embedded systems is even worse. The memory hierarchy consumes a large amount of chip area and energy, which are precious resources in embedded systems. Moreover, embedded systems have multiple design objectives such as performance, energy consumption, and area, etc. Customizing the memory hierarchy for specific applications is a very important way to take full advantage of limited resources to maximize the performance. However, the traditional custom memory hierarchy design methodologies are phase-ordered. They separate the application optimization from the memory hierarchy architecture design, which tend to result in local-optimal solutions. In traditional Hardware-Software co-design methodologies, much of the work has focused on utilizing reconfigurable logic to partition the computation. However, utilizing reconfigurable logic to perform the memory hierarchy design is seldom addressed. In this paper, we propose a new framework for designing memory hierarchy for embedded systems. The framework will take advantage of the flexible reconfigurable logic to customize the memory hierarchy for specific applications. It combines the application optimization and memory hierarchy design together to obtain a global-optimal solution. Using the framework, we performed a case study to design a new software-controlled instruction memory that showed promising potential.
Resumo:
This article describes an application of computers to a consumer-based production engineering environment. Particular consideration is given to the utilisation of low-cost computer systems for the visual inspection of components on a production line in real time. The process of installation is discussed, from identifying the need for artificial vision and justifying the cost, through to choosing a particular system and designing the physical and program structure.
Resumo:
Terahertz pulse imaging (TPI) is a novel noncontact, nondestructive technique for the examination of cultural heritage artifacts. It has the advantage of broadband spectral range, time-of-flight depth resolution, and penetration through optically opaque materials. Fiber-coupled, portable, time-domain terahertz systems have enabled this technique to move out of the laboratory and into the field. Much like the rings of a tree, stratified architectural materials give the chronology of their environmental and aesthetic history. This work concentrates on laboratory models of stratified mosaics and fresco paintings, specimens extracted from a neolithic excavation site in Catalhoyuk, Turkey, and specimens measured at the medieval Eglise de Saint Jean-Baptiste in Vif, France. Preparatory spectroscopic studies of various composite materials, including lime, gypsum and clay plasters are presented to enhance the interpretation of results and with the intent to aid future computer simulations of the TPI of stratified architectural material. The breadth of the sample range is a demonstration of the cultural demand and public interest in the life history of buildings. The results are an illustration of the potential role of TPI in providing both a chronological history of buildings and in the visualization of obscured wall paintings and mosaics.