937 resultados para parallel computer systems


Relevância:

40.00% 40.00%

Publicador:

Resumo:

Parallel kinematic structures are considered very adequate architectures for positioning and orienti ng the tools of robotic mechanisms. However, developing dynamic models for this kind of systems is sometimes a difficult task. In fact, the direct application of traditional methods of robotics, for modelling and analysing such systems, usually does not lead to efficient and systematic algorithms. This work addre sses this issue: to present a modular approach to generate the dynamic model and through some convenient modifications, how we can make these methods more applicable to parallel structures as well. Kane’s formulati on to obtain the dynamic equations is shown to be one of the easiest ways to deal with redundant coordinates and kinematic constraints, so that a suitable c hoice of a set of coordinates allows the remaining of the modelling procedure to be computer aided. The advantages of this approach are discussed in the modelling of a 3-dof parallel asymmetric mechanisms.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

The increasing precision of current and future experiments in high-energy physics requires a likewise increase in the accuracy of the calculation of theoretical predictions, in order to find evidence for possible deviations of the generally accepted Standard Model of elementary particles and interactions. Calculating the experimentally measurable cross sections of scattering and decay processes to a higher accuracy directly translates into including higher order radiative corrections in the calculation. The large number of particles and interactions in the full Standard Model results in an exponentially growing number of Feynman diagrams contributing to any given process in higher orders. Additionally, the appearance of multiple independent mass scales makes even the calculation of single diagrams non-trivial. For over two decades now, the only way to cope with these issues has been to rely on the assistance of computers. The aim of the xloops project is to provide the necessary tools to automate the calculation procedures as far as possible, including the generation of the contributing diagrams and the evaluation of the resulting Feynman integrals. The latter is based on the techniques developed in Mainz for solving one- and two-loop diagrams in a general and systematic way using parallel/orthogonal space methods. These techniques involve a considerable amount of symbolic computations. During the development of xloops it was found that conventional computer algebra systems were not a suitable implementation environment. For this reason, a new system called GiNaC has been created, which allows the development of large-scale symbolic applications in an object-oriented fashion within the C++ programming language. This system, which is now also in use for other projects besides xloops, is the main focus of this thesis. The implementation of GiNaC as a C++ library sets it apart from other algebraic systems. Our results prove that a highly efficient symbolic manipulator can be designed in an object-oriented way, and that having a very fine granularity of objects is also feasible. The xloops-related parts of this work consist of a new implementation, based on GiNaC, of functions for calculating one-loop Feynman integrals that already existed in the original xloops program, as well as the addition of supplementary modules belonging to the interface between the library of integral functions and the diagram generator.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

This work presents algorithms for the calculation of the electrostatic interaction in partially periodic systems. The framework for these algorithms is provided by the simulation package ESPResSo, of which the author was one of the main developers. The prominent features of the program are listed and the internal structure is described. In the following, algorithms for the calculation of the Coulomb sum in three dimensionally periodic systems are described. These methods are the foundations for the algorithms for partially periodic systems presented in this work. Starting from the MMM2D method for systems with one non-periodic coordinate, the ELC method for these systems is developed. This method consists of a correction term which allows to use methods for three dimensional periodicity also for the case of two periodic coordinates. The computation time of this correction term is neglible for large numbers of particles. The performance of MMM2D and ELC are demonstrated by results from the implementations contained in ESPResSo. It is also discussed, how different dielectric constants inside and outside of the simulation box can be realized. For systems with one periodic coordinate, the MMM1D method is derived from the MMM2D method. This method is applied to the problem of the attraction of like-charged rods in the presence of counterions, and results of the strong coupling theory for the equilibrium distance of the rods at infinite counterion-coupling are checked against results from computer simulations. The degree of agreement between the simulations at finite coupling and the theory can be characterized by a single parameter gamma_RB. In the special case of T=0, one finds under certain circumstances flat configurations, in which all charges are located in the rod-rod plane. The energetically optimal configuration and its stability are determined analytically, which depends on only one parameter gamma_z, similar to gamma_RB. These findings are in good agreement with results from computer simulations.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

The term "Brain Imaging" identi�es a set of techniques to analyze the structure and/or functional behavior of the brain in normal and/or pathological situations. These techniques are largely used in the study of brain activity. In addition to clinical usage, analysis of brain activity is gaining popularity in others recent �fields, i.e. Brain Computer Interfaces (BCI) and the study of cognitive processes. In this context, usage of classical solutions (e.g. f MRI, PET-CT) could be unfeasible, due to their low temporal resolution, high cost and limited portability. For these reasons alternative low cost techniques are object of research, typically based on simple recording hardware and on intensive data elaboration process. Typical examples are ElectroEncephaloGraphy (EEG) and Electrical Impedance Tomography (EIT), where electric potential at the patient's scalp is recorded by high impedance electrodes. In EEG potentials are directly generated from neuronal activity, while in EIT by the injection of small currents at the scalp. To retrieve meaningful insights on brain activity from measurements, EIT and EEG relies on detailed knowledge of the underlying electrical properties of the body. This is obtained from numerical models of the electric �field distribution therein. The inhomogeneous and anisotropic electric properties of human tissues make accurate modeling and simulation very challenging, leading to a tradeo�ff between physical accuracy and technical feasibility, which currently severely limits the capabilities of these techniques. Moreover elaboration of data recorded requires usage of regularization techniques computationally intensive, which influences the application with heavy temporal constraints (such as BCI). This work focuses on the parallel implementation of a work-flow for EEG and EIT data processing. The resulting software is accelerated using multi-core GPUs, in order to provide solution in reasonable times and address requirements of real-time BCI systems, without over-simplifying the complexity and accuracy of the head models.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Monte Carlo simulations are used to study the effect of confinement on a crystal of point particles interacting with an inverse power law potential in d=2 dimensions. This system can describe colloidal particles at the air-water interface, a model system for experimental study of two-dimensional melting. It is shown that the state of the system (a strip of width D) depends very sensitively on the precise boundary conditions at the two ``walls'' providing the confinement. If one uses a corrugated boundary commensurate with the order of the bulk triangular crystalline structure, both orientational order and positional order is enhanced, and such surface-induced order persists near the boundaries also at temperatures where the system in the bulk is in its fluid state. However, using smooth repulsive boundaries as walls providing the confinement, only the orientational order is enhanced, but positional (quasi-) long range order is destroyed: The mean-square displacement of two particles n lattice parameters apart in the y-direction along the walls then crosses over from the logarithmic increase (characteristic for $d=2$) to a linear increase (characteristic for d=1). The strip then exhibits a vanishing shear modulus. These results are interpreted in terms of a phenomenological harmonic theory. Also the effect of incommensurability of the strip width D with the triangular lattice structure is discussed, and a comparison with surface effects on phase transitions in simple Ising- and XY-models is made

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Mainstream hardware is becoming parallel, heterogeneous, and distributed on every desk, every home and in every pocket. As a consequence, in the last years software is having an epochal turn toward concurrency, distribution, interaction which is pushed by the evolution of hardware architectures and the growing of network availability. This calls for introducing further abstraction layers on top of those provided by classical mainstream programming paradigms, to tackle more effectively the new complexities that developers have to face in everyday programming. A convergence it is recognizable in the mainstream toward the adoption of the actor paradigm as a mean to unite object-oriented programming and concurrency. Nevertheless, we argue that the actor paradigm can only be considered a good starting point to provide a more comprehensive response to such a fundamental and radical change in software development. Accordingly, the main objective of this thesis is to propose Agent-Oriented Programming (AOP) as a high-level general purpose programming paradigm, natural evolution of actors and objects, introducing a further level of human-inspired concepts for programming software systems, meant to simplify the design and programming of concurrent, distributed, reactive/interactive programs. To this end, in the dissertation first we construct the required background by studying the state-of-the-art of both actor-oriented and agent-oriented programming, and then we focus on the engineering of integrated programming technologies for developing agent-based systems in their classical application domains: artificial intelligence and distributed artificial intelligence. Then, we shift the perspective moving from the development of intelligent software systems, toward general purpose software development. Using the expertise maturated during the phase of background construction, we introduce a general-purpose programming language named simpAL, which founds its roots on general principles and practices of software development, and at the same time provides an agent-oriented level of abstraction for the engineering of general purpose software systems.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Despite the several issues faced in the past, the evolutionary trend of silicon has kept its constant pace. Today an ever increasing number of cores is integrated onto the same die. Unfortunately, the extraordinary performance achievable by the many-core paradigm is limited by several factors. Memory bandwidth limitation, combined with inefficient synchronization mechanisms, can severely overcome the potential computation capabilities. Moreover, the huge HW/SW design space requires accurate and flexible tools to perform architectural explorations and validation of design choices. In this thesis we focus on the aforementioned aspects: a flexible and accurate Virtual Platform has been developed, targeting a reference many-core architecture. Such tool has been used to perform architectural explorations, focusing on instruction caching architecture and hybrid HW/SW synchronization mechanism. Beside architectural implications, another issue of embedded systems is considered: energy efficiency. Near Threshold Computing is a key research area in the Ultra-Low-Power domain, as it promises a tenfold improvement in energy efficiency compared to super-threshold operation and it mitigates thermal bottlenecks. The physical implications of modern deep sub-micron technology are severely limiting performance and reliability of modern designs. Reliability becomes a major obstacle when operating in NTC, especially memory operation becomes unreliable and can compromise system correctness. In the present work a novel hybrid memory architecture is devised to overcome reliability issues and at the same time improve energy efficiency by means of aggressive voltage scaling when allowed by workload requirements. Variability is another great drawback of near-threshold operation. The greatly increased sensitivity to threshold voltage variations in today a major concern for electronic devices. We introduce a variation-tolerant extension of the baseline many-core architecture. By means of micro-architectural knobs and a lightweight runtime control unit, the baseline architecture becomes dynamically tolerant to variations.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

The monitoring of cognitive functions aims at gaining information about the current cognitive state of the user by decoding brain signals. In recent years, this approach allowed to acquire valuable information about the cognitive aspects regarding the interaction of humans with external world. From this consideration, researchers started to consider passive application of brain–computer interface (BCI) in order to provide a novel input modality for technical systems solely based on brain activity. The objective of this thesis is to demonstrate how the passive Brain Computer Interfaces (BCIs) applications can be used to assess the mental states of the users, in order to improve the human machine interaction. Two main studies has been proposed. The first one allows to investigate whatever the Event Related Potentials (ERPs) morphological variations can be used to predict the users’ mental states (e.g. attentional resources, mental workload) during different reactive BCI tasks (e.g. P300-based BCIs), and if these information can predict the subjects’ performance in performing the tasks. In the second study, a passive BCI system able to online estimate the mental workload of the user by relying on the combination of the EEG and the ECG biosignals has been proposed. The latter study has been performed by simulating an operative scenario, in which the occurrence of errors or lack of performance could have significant consequences. The results showed that the proposed system is able to estimate online the mental workload of the subjects discriminating three different difficulty level of the tasks ensuring a high reliability.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Computer-Simulationen von Kolloidalen Fluiden in Beschränkten Geometrien Kolloidale Suspensionen, die einen Phasenübergang aufweisen, zeigen eine Vielfalt an interessanten Effekten, sobald sie auf eine bestimmte Geometrie beschränkt werden, wie zum Beispiel auf zylindrische Poren, sphärische Hohlräume oder auf einen Spalt mit ebenen Wänden. Der Einfluss dieser verschiedenen Geometrietypen sowohl auf das Phasenverhalten als auch auf die Dynamik von Kolloid-Polymer-Mischungen wird mit Hilfe von Computer-Simulationen unter Verwendung des Asakura-Oosawa- Modells, für welches auf Grund der “Depletion”-Kräfte ein Phasenübergang existiert, untersucht. Im Fall von zylindrischen Poren sieht man ein interessantes Phasenverhalten, welches vom eindimensionalen Charakter des Systems hervorgerufen wird. In einer kurzen Pore findet man im Bereich des Phasendiagramms, in dem das System typischerweise entmischt, entweder eine polymerreiche oder eine kolloidreiche Phase vor. Sobald aber die Länge der zylindrischen Pore die typische Korrelationslänge entlang der Zylinderachse überschreitet, bilden sich mehrere quasi-eindimensionale Bereiche der polymerreichen und der kolloidreichen Phase, welche von nun an koexistieren. Diese Untersuchungen helfen das Verhalten von Adsorptionshysteresekurven in entsprechenden Experimenten zu erklären. Wenn das Kolloid-Polymer-Modellsystem auf einen sphärischen Hohlraum eingeschränkt wird, verschiebt sich der Punkt des Phasenübergangs von der polymerreichen zur kolloidreichen Phase. Es wird gezeigt, dass diese Verschiebung direkt von den Benetzungseigenschaften des Systems abhängt, was die Beobachtung von zwei verschiedenen Morphologien bei Phasenkoexistenz ermöglicht – Schalenstrukturen und Strukturen des Janustyps. Im Rahmen der Untersuchung von heterogener Keimbildung von Kristallen innerhalb einer Flüssigkeit wird eine neue Simulationsmethode zur Berechnung von Freien Energien der Grenzfläche zwischen Kristall- bzw. Flüssigkeitsphase undWand präsentiert. Die Resultate für ein System von harten Kugeln und ein System einer Kolloid- Polymer-Mischung werden anschließend zur Bestimmung von Kontaktwinkeln von Kristallkeimen an Wänden verwendet. Die Dynamik der Phasenseparation eines quasi-zweidimensionalen Systems, welche sich nach einem Quench des Systems aus dem homogenen Zustand in den entmischten Zustand ausbildet, wird mit Hilfe von einer mesoskaligen Simulationsmethode (“Multi Particle Collision Dynamics”) untersucht, die sich für eine detaillierte Untersuchung des Einflusses der hydrodynamischen Wechselwirkung eignet. Die Exponenten universeller Potenzgesetze, die das Wachstum der mittleren Domänengröße beschreiben, welche für rein zwei- bzw. dreidimensionale Systeme bekannt sind, können für bestimmte Parameterbereiche nachgewiesen werden. Die unterschiedliche Dynamik senkrecht bzw. parallel zu den Wänden sowie der Einfluss der Randbedingungen für das Lösungsmittel werden untersucht. Es wird gezeigt, dass die daraus resultierende Abschirmung der hydrodynamischen Wechselwirkungsreichweite starke Auswirkungen auf das Wachstum der mittleren Domänengröße hat.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Eine der offenen Fragen der aktuellen Physik ist das Verständnis von Systemen im Nichtgleichgewicht. Im Gegensatz zu der Gleichgewichtsphysik ist in diesem Bereich aktuell kein Formalismus bekannt der ein systematisches Beschreiben der unterschiedlichen Systeme ermöglicht. Um das Verständnis über diese Systeme zu vergrößern werden in dieser Arbeit zwei unterschiedliche Systeme studiert, die unter einem externen Feld ein starkes nichtlineares Verhalten zeigen. Hierbei handelt es sich zum einen um das Verhalten von Teilchen unter dem Einfluss einer extern angelegten Kraft und zum anderen um das Verhalten eines Systems in der Nähe des kritischen Punktes unter Scherung. Das Modellsystem in dem ersten Teil der Arbeit ist eine binäre Yukawa Mischung, die bei tiefen Temperaturen einen Glassübergang zeigt. Dies führt zu einer stark ansteigenden Relaxationszeit des Systems, so dass man auch bei kleinen Kräften relativ schnell ein nichtlineares Verhalten beobachtet. In Abhängigkeit der angelegten konstanten Kraft können in dieser Arbeit drei Regime, mit stark unterschiedlichem Teilchenverhalten, identifiziert werden. In dem zweiten Teil der Arbeit wird das Ising-Modell unter Scherung betrachtet. In der Nähe des kritischen Punkts kommt es in diesem Modell zu einer Beeinflussung der Fluktuationen in dem System durch das angelegte Scherfeld. Dies hat zur Folge, dass das System stark anisotrop wird und man zwei unterschiedliche Korrelationslängen vorfindet, die mit unterschiedlichen Exponenten divergieren. Infolgedessen lässt sich der normale isotrope Formalismus des "finite-size scaling" nicht mehr auf dieses System anwenden. In dieser Arbeit wird gezeigt, wie dieser auf den anisotropen Fall zu verallgemeinern ist und wie damit die kritischen Punkte, sowie die dazu gehörenden kritischen Exponenten berechnet werden können.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

In this work, electrophoretic preconcentration of protein and peptide samples in microchannels was studied theoretically using the 1D dynamic simulator GENTRANS, and experimentally combined with MS. In all configurations studied, the sample was uniformly distributed throughout the channel before power application, and driving electrodes were used as microchannel ends. In the first part, previously obtained experimental results from carrier-free systems are compared to simulation results, and the effects of atmospheric carbon dioxide and impurities in the sample solution are examined. Simulation provided insight into the dynamics of the transport of all components under the applied electric field and revealed the formation of a pure water zone in the channel center. In the second part, the use of an IEF procedure with simple well defined amphoteric carrier components, i.e. amino acids, for concentration and fractionation of peptides was investigated. By performing simulations a qualitative description of the analyte behavior in this system was obtained. Neurotensin and [Glu1]-Fibrinopeptide B were separated by IEF in microchannels featuring a liquid lid for simple sample handling and placement of the driving electrodes. Component distributions in the channel were detected using MALDI- and nano-ESI-MS and data were in agreement with those obtained by simulation. Dynamic simulations are demonstrated to represent an effective tool to investigate the electrophoretic behavior of all components in the microchannel.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Within the membrane computing research field, there are many papers about software simulations and a few about hardware implementations. In both cases, algorithms for implementing membrane systems in software and hardware that try to take advantages of massive parallelism are implemented. P-systems are parallel and non deterministic systems which simulate membranes behavior when processing information. This paper presents software techniques based on the proper utilization of virtual memory of a computer. There is a study of how much virtual memory is necessary to host a membrane model. This method improves performance in terms of time.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Incorporating the possibility of attaching attributes to variables in a logic programming system has been shown to allow the addition of general constraint solving capabilities to it. This approach is very attractive in that by adding a few primitives any logic programming system can be turned into a generic constraint logic programming system in which constraint solving can be user deñned, and at source level - an extreme example of the "glass box" approach. In this paper we propose a different and novel use for the concept of attributed variables: developing a generic parallel/concurrent (constraint) logic programming system, using the same "glass box" flavor. We argüe that a system which implements attributed variables and a few additional primitives can be easily customized at source level to implement many of the languages and execution models of parallelism and concurrency currently proposed, in both shared memory and distributed systems. We illustrate this through examples and report on an implementation of our ideas.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Incorporating the possibility of attaching attributes to variables in a logic programming system has been shown to allow the addition of general constraint solving capabilities to it. This approach is very attractive in that by adding a few primitives any logic programming system can be turned into a generic constraint logic programming system in which constraint solving can be user defined, and at source level - an extreme example of the "glass box" approach. In this paper we propose a different and novel use for the concept of attributed variables: developing a generic parallel/concurrent (constraint) logic programming system, using the same "glass box" flavor. We argüe that a system which implements attributed variables and a few additional primitives can be easily customized at source level to implement many of the languages and execution models of parallelism and concurrency currently proposed, in both shared memory and distributed systems. We illustrate this through examples.