9 resultados para Parallel computing, Virtual machine, Composition, Determinism, Abstraction
em Universidade Federal do Rio Grande do Norte(UFRN)
Resumo:
The lava Platform is increasing1y being adopted in the development of distributed sys¬tems with higb user demando This kind of application is more complex because it needs beyond attending the functional requirements, to fulfil1 the pre-established performance parameters. This work makes a study on the Java Vutual Machine (JVM), approaching its intemal aspects and exploring the garbage collection strategies existing in the literature and used by the NM. It also presents a set of tools that helps in the job of optimizing applications and others that help in the monitoring of applications in the production envi¬ronment. Doe to the great amount of technologies that aim to solve problems which are common to the application layer, it becomes difficult to choose the one with best time response and less memory usage. This work presents a brief introduction to each one of tbe possible technologies and realize comparative tests through a statistical analysis of the response time and garbage collection activity random variables. The obtained results supply engineers and managers with a subside to decide which technologies to use in large applications through the knowledge of how they behave in their environments and the amount of resources that they consume. The relation between the productivity of the technology and its performance is also considered ao important factor in this choice
Resumo:
Artificial neural networks are usually applied to solve complex problems. In problems with more complexity, by increasing the number of layers and neurons, it is possible to achieve greater functional efficiency. Nevertheless, this leads to a greater computational effort. The response time is an important factor in the decision to use neural networks in some systems. Many argue that the computational cost is higher in the training period. However, this phase is held only once. Once the network trained, it is necessary to use the existing computational resources efficiently. In the multicore era, the problem boils down to efficient use of all available processing cores. However, it is necessary to consider the overhead of parallel computing. In this sense, this paper proposes a modular structure that proved to be more suitable for parallel implementations. It is proposed to parallelize the feedforward process of an RNA-type MLP, implemented with OpenMP on a shared memory computer architecture. The research consistes on testing and analizing execution times. Speedup, efficiency and parallel scalability are analyzed. In the proposed approach, by reducing the number of connections between remote neurons, the response time of the network decreases and, consequently, so does the total execution time. The time required for communication and synchronization is directly linked to the number of remote neurons in the network, and so it is necessary to investigate which one is the best distribution of remote connections
Resumo:
The increasing demand for high performance wireless communication systems has shown the inefficiency of the current model of fixed allocation of the radio spectrum. In this context, cognitive radio appears as a more efficient alternative, by providing opportunistic spectrum access, with the maximum bandwidth possible. To ensure these requirements, it is necessary that the transmitter identify opportunities for transmission and the receiver recognizes the parameters defined for the communication signal. The techniques that use cyclostationary analysis can be applied to problems in either spectrum sensing and modulation classification, even in low signal-to-noise ratio (SNR) environments. However, despite the robustness, one of the main disadvantages of cyclostationarity is the high computational cost for calculating its functions. This work proposes efficient architectures for obtaining cyclostationary features to be employed in either spectrum sensing and automatic modulation classification (AMC). In the context of spectrum sensing, a parallelized algorithm for extracting cyclostationary features of communication signals is presented. The performance of this features extractor parallelization is evaluated by speedup and parallel eficiency metrics. The architecture for spectrum sensing is analyzed for several configuration of false alarm probability, SNR levels and observation time for BPSK and QPSK modulations. In the context of AMC, the reduced alpha-profile is proposed as as a cyclostationary signature calculated for a reduced cyclic frequencies set. This signature is validated by a modulation classification architecture based on pattern matching. The architecture for AMC is investigated for correct classification rates of AM, BPSK, QPSK, MSK and FSK modulations, considering several scenarios of observation length and SNR levels. The numerical results of performance obtained in this work show the eficiency of the proposed architectures
Resumo:
With the advance of the Cloud Computing paradigm, a single service offered by a cloud platform may not be enough to meet all the application requirements. To fulfill such requirements, it may be necessary, instead of a single service, a composition of services that aggregates services provided by different cloud platforms. In order to generate aggregated value for the user, this composition of services provided by several Cloud Computing platforms requires a solution in terms of platforms integration, which encompasses the manipulation of a wide number of noninteroperable APIs and protocols from different platform vendors. In this scenario, this work presents Cloud Integrator, a middleware platform for composing services provided by different Cloud Computing platforms. Besides providing an environment that facilitates the development and execution of applications that use such services, Cloud Integrator works as a mediator by providing mechanisms for building applications through composition and selection of semantic Web services that take into account metadata about the services, such as QoS (Quality of Service), prices, etc. Moreover, the proposed middleware platform provides an adaptation mechanism that can be triggered in case of failure or quality degradation of one or more services used by the running application in order to ensure its quality and availability. In this work, through a case study that consists of an application that use services provided by different cloud platforms, Cloud Integrator is evaluated in terms of the efficiency of the performed service composition, selection and adaptation processes, as well as the potential of using this middleware in heterogeneous computational clouds scenarios
Resumo:
The reverse time migration algorithm (RTM) has been widely used in the seismic industry to generate images of the underground and thus reduce the risk of oil and gas exploration. Its widespread use is due to its high quality in underground imaging. The RTM is also known for its high computational cost. Therefore, parallel computing techniques have been used in their implementations. In general, parallel approaches for RTM use a coarse granularity by distributing the processing of a subset of seismic shots among nodes of distributed systems. Parallel approaches with coarse granularity for RTM have been shown to be very efficient since the processing of each seismic shot can be performed independently. For this reason, RTM algorithm performance can be considerably improved by using a parallel approach with finer granularity for the processing assigned to each node. This work presents an efficient parallel algorithm for 3D reverse time migration with fine granularity using OpenMP. The propagation algorithm of 3D acoustic wave makes up much of the RTM. Different load balancing were analyzed in order to minimize possible losses parallel performance at this stage. The results served as a basis for the implementation of other phases RTM: backpropagation and imaging condition. The proposed algorithm was tested with synthetic data representing some of the possible underground structures. Metrics such as speedup and efficiency were used to analyze its parallel performance. The migrated sections show that the algorithm obtained satisfactory performance in identifying subsurface structures. As for the parallel performance, the analysis clearly demonstrate the scalability of the algorithm achieving a speedup of 22.46 for the propagation of the wave and 16.95 for the RTM, both with 24 threads.
Resumo:
Based on the presupposition that the arts in the West always counted on resources, supports, and devices pertaining to its time context, an reflection is intended regarding the scenic compositions mediated by digital technologies do. Such technologies are inserted in the daily routine, also composing artistic experiments, thus playing a dialogical role with the art/technology intersection. Therefore, the proposal is to investigate what relationships are established in the contemporary theatrical scene from the contagion by digital technologies, aiming at establishing this parallel through a dialogue with the authors discussing the subject, and also based on the group practices having technological resources as a determinant factor in their plays. Furthermore, a reflection should be made on the scene that incorporates or is carried out in intermediatic events, analyzing how digital technologies (re)configure compositional processes of the plays by GAG Phila7, in the city of São Paulo/SP. For such, the dissertation is organized in three sections comprising four moments, to wit: brief overview of the field, contextualization, poetic analysis and synthesis. Qualitative methods are used as the methodological proposal: semi-structure interview, note and document taking (program, website, playing book, disclosure material for advertising text, photographs, and videos). Within the universe of qualitative research, it works with the epistemological perspective of the Gadamer philosophical hermeneutics. The possibilities allowed by the double virtual (Internet/web) generated a type of theater with another material basis and new forms of organization and structure, being possible to perceive that such technological advances and the arts are mutually contaminated, generating a dislocation in the logics of theatrical composition, movement beginning with the artistic vanguards, gradually intensified, thus offering new possibilities of constructions and hybridization of the of the most different possible types. Experiment ―Profanações_superfície de eventos de construção coletiva‖, idealized by Phila7 is inserted in this perspective. Object of the discussion of such research, the experiment works with possible poetics arising from the intersection with the digital technologies, aiming at identifying and problematizing the challenges from the technological evolution and expansion in a scenic context
Resumo:
Currently there is still a high demand for quality control in manufacturing processes of mechanical parts. This keeps alive the need for the inspection activity of final products ranging from dimensional analysis to chemical composition of products. Usually this task may be done through various nondestructive and destructive methods that ensure the integrity of the parts. The result generated by these modern inspection tools ends up not being able to geometrically define the real damage and, therefore, cannot be properly displayed on a computing environment screen. Virtual 3D visualization may help identify damage that would hardly be detected by any other methods. One may find some commercial softwares that seek to address the stages of a design and simulation of mechanical parts in order to predict possible damages trying to diminish potential undesirable events. However, the challenge of developing softwares capable of integrating the various design activities, product inspection, results of non-destructive testing as well as the simulation of damage still needs the attention of researchers. This was the motivation to conduct a methodological study for implementation of a versatile CAD/CAE computer kernel capable of helping programmers in developing softwares applied to the activities of design and simulation of mechanics parts under stress. In this research it is presented interesting results obtained from the use of the developed kernel showing that it was successfully applied to case studies of design including parts presenting specific geometries, namely: mechanical prostheses, heat exchangers and piping of oil and gas. Finally, the conclusions regarding the experience of merging CAD and CAE theories to develop the kernel, so as to result in a tool adaptable to various applications of the metalworking industry are presented
Resumo:
In this paper we present the methodological procedures involved in the digital imaging in mesoscale of a block of travertines rock of quaternary age, originating from the city of Acquasanta, located in the Apennines, Italy. This rocky block, called T-Block, was stored in the courtyard of the Laboratório Experimental Petróleo "Kelsen Valente" (LabPetro), of Universidade Estadual de Campinas (UNICAMP), so that from it were performed Scientific studies, mainly for research groups universities and research centers working in brazilian areas of reservoir characterization and 3D digital imaging. The purpose of this work is the development of a Model Solid Digital, from the use of non-invasive techniques of digital 3D imaging of internal and external surfaces of the T-Block. For the imaging of the external surfaces technology has been used LIDAR (Light Detection and Range) and the imaging surface Interior was done using Ground Penetrating Radar (GPR), moreover, profiles were obtained with a Gamma Ray Gamae-spectômetro laptop. The goal of 3D digital imaging involved the identification and parameterization of surface geological and sedimentary facies that could represent heterogeneities depositional mesoscale, based on study of a block rocky with dimensions of approximately 1.60 m x 1.60 m x 2.70 m. The data acquired by means of terrestrial laser scanner made available georeferenced spatial information of the surface of the block (X, Y, Z), and varying the intensity values of the return laser beam and high resolution RGB data (3 mm x 3 mm), total points acquired 28,505,106. This information was used as an aid in the interpretation of radargrams and are ready to be displayed in rooms virtual reality. With the GPR was obtained 15 profiles of 2.3 m and 2 3D grids, each with 24 sections horizontal of 1.3 and 14 m vertical sections of 2.3 m, both the Antenna 900 MHz to about 2600 MHz antenna. Finally, the use of GPR associated with Laser Scanner enabled the identification and 3D mapping of 3 different radarfácies which were correlated with three sedimentary facies as had been defined at the outset. The 6 profiles showed gamma a low amplitude variation in the values of radioactivity. This is likely due to the fact of the sedimentary layers profiled have the same mineralogical composition, being composed by carbonate sediments, with no clay in siliciclastic pellitic layers or other mineral carrier elements radioactive
Resumo:
Based on the presupposition that the arts in the West always counted on resources, supports, and devices pertaining to its time context, an reflection is intended regarding the scenic compositions mediated by digital technologies do. Such technologies are inserted in the daily routine, also composing artistic experiments, thus playing a dialogical role with the art/technology intersection. Therefore, the proposal is to investigate what relationships are established in the contemporary theatrical scene from the contagion by digital technologies, aiming at establishing this parallel through a dialogue with the authors discussing the subject, and also based on the group practices having technological resources as a determinant factor in their plays. Furthermore, a reflection should be made on the scene that incorporates or is carried out in intermediatic events, analyzing how digital technologies (re)configure compositional processes of the plays by GAG Phila7, in the city of São Paulo/SP. For such, the dissertation is organized in three sections comprising four moments, to wit: brief overview of the field, contextualization, poetic analysis and synthesis. Qualitative methods are used as the methodological proposal: semi-structure interview, note and document taking (program, website, playing book, disclosure material for advertising text, photographs, and videos). Within the universe of qualitative research, it works with the epistemological perspective of the Gadamer philosophical hermeneutics. The possibilities allowed by the double virtual (Internet/web) generated a type of theater with another material basis and new forms of organization and structure, being possible to perceive that such technological advances and the arts are mutually contaminated, generating a dislocation in the logics of theatrical composition, movement beginning with the artistic vanguards, gradually intensified, thus offering new possibilities of constructions and hybridization of the of the most different possible types. Experiment ―Profanações_superfície de eventos de construção coletiva‖, idealized by Phila7 is inserted in this perspective. Object of the discussion of such research, the experiment works with possible poetics arising from the intersection with the digital technologies, aiming at identifying and problematizing the challenges from the technological evolution and expansion in a scenic context