935 resultados para Programação orientada a objetos (computação)
Resumo:
Hospital Automation is an area that is constantly growing. The emergency of new technologies and hardware is transforming the processes more efficient. Nevertheless, some of the hospital processes are still being performed manually, such as monitoring of patients that is considered critical because it involves human lives. One of the factors that should be taken into account during a monitoring is the agility to detect any abnormality in vital signs of patients, as well as warning of this anomaly to the medical team involved. So, this master's thesis aims to develop an architecture to automate this process of monitoring and reporting of possible alert to a professional, so that emergency care can be done effectively. The computing mobile was used to improve the communication by distributing messages between a central located into the hospital and the mobile carried by the duty
Resumo:
This study shows the implementation and the embedding of an Artificial Neural Network (ANN) in hardware, or in a programmable device, as a field programmable gate array (FPGA). This work allowed the exploration of different implementations, described in VHDL, of multilayer perceptrons ANN. Due to the parallelism inherent to ANNs, there are disadvantages in software implementations due to the sequential nature of the Von Neumann architectures. As an alternative to this problem, there is a hardware implementation that allows to exploit all the parallelism implicit in this model. Currently, there is an increase in use of FPGAs as a platform to implement neural networks in hardware, exploiting the high processing power, low cost, ease of programming and ability to reconfigure the circuit, allowing the network to adapt to different applications. Given this context, the aim is to develop arrays of neural networks in hardware, a flexible architecture, in which it is possible to add or remove neurons, and mainly, modify the network topology, in order to enable a modular network of fixed-point arithmetic in a FPGA. Five synthesis of VHDL descriptions were produced: two for the neuron with one or two entrances, and three different architectures of ANN. The descriptions of the used architectures became very modular, easily allowing the increase or decrease of the number of neurons. As a result, some complete neural networks were implemented in FPGA, in fixed-point arithmetic, with a high-capacity parallel processing
Resumo:
Foundation Fieldbus Industrial networks are the high standard technology which allows users to create complex control logic and totally decentralized. Although being so advanced, they still have some limitations imposed by their own technology. Attempting to solve one of these limitations, this paper describes how to design a Fuzzy controller in a Foundation Fieldbus network using their basic elements of programming, the functional blocks, so that the network remains fully independent of other devices other than the same instruments that constitute it. Moreover, in this work was developed a tool that aids this process of building the Fuzzy controller, setting the internal parameters of functional blocks and informing how many and which blocks should be used for a given structure. The biggest challenge in creating this controller is exactly the choice of blocks and how to arrange them in order to effectuate the same functions of a Fuzzy controller implemented in other kind of environment. The methodology adopted was to divide each one of the phases of a traditional Fuzzy controller and then create simple structures with the functional blocks to implement them. At the end of the work, the developed controller is compared with a Fuzzy controller implemented in a mathematical program that it has a proper tool for the development and implementation of Fuzzy controllers, obtaining comparatives graphics of performance between both
Resumo:
A seleção de pulverizadores agrícolas que se adaptem às necessidades da propriedade, é um processo trabalhoso, sendo uma das etapas mais importantes dentro do processo produtivo. O objetivo do presente trabalho foi o de desenvolver e utilizar um modelo de programação linear para auxiliar na seleção de pulverizadores agrícolas de barras, baseado no menor custo horário do equipamento. Foram utilizadas as informações técnicas referentes a 20 modelos de pulverizadores disponíveis no mercado, sendo quatro autopropelidos, oito de arrasto e oito do tipo montado. A análise de sensibilidade dos componentes dos custos operacionais mostrou que as taxas de reparo e depreciação foram os fatores que mais interferiram na variação do custo horário do conjunto trator-pulverizador. O modelo matemático desenvolvido facilitou a realização da análise de sensibilidade que foi processada em um tempo muito pequeno.
Resumo:
This work develops a methodology for defining the maximum active power being injected into predefined nodes in the studied distribution networks, considering the possibility of multiple accesses of generating units. The definition of these maximum values is obtained from an optimization study, in which further losses should not exceed those of the base case, i.e., without the presence of distributed generation. The restrictions on the loading of the branches and voltages of the system are respected. To face the problem it is proposed an algorithm, which is based on the numerical method called particle swarm optimization, applied to the study of AC conventional load flow and optimal load flow for maximizing the penetration of distributed generation. Alternatively, the Newton-Raphson method was incorporated to resolution of the load flow. The computer program is performed with the SCILAB software. The proposed algorithm is tested with the data from the IEEE network with 14 nodes and from another network, this one from the Rio Grande do Norte State, at a high voltage (69 kV), with 25 nodes. The algorithm defines allowed values of nominal active power of distributed generation, in percentage terms relative to the demand of the network, from reference values
Resumo:
Conselho Nacional de Desenvolvimento Científico e Tecnológico (CNPq)
Resumo:
The use of graphical objects three-dimensional (3D) multimedia applications is gaining more space in the media. Networks with high transmission rates, computers with large processing and graphics boost and popularize such three-dimensional applications. The areas of 3D applications ranging from military applications, entertainment applications geared up for education. Within the applications related to education, we highlight the applications that create virtual copies of cultural spaces such as museums. Through this copy, you can virtually visit a museum, see other users, communicate, exchange information on works, etc. Thereby allowing the visit museums physically distant remote users. A major problem of such virtual environments is its update. By dealing with various media (text, images, sounds, and 3D models), its subsequent handling and update on a virtual environment requires staff with specialized knowledge. Speaking of museums, they hardly have people on your team with this profile. Inside the GT-MV (Grupo de Trabalho de Museus Virtuais), funded by RNP (Rede Nacional de Ensino e Pesquisa) propose a portal for registration, amendment and seen collaborative virtual museums of Brazil. The update, be it related to work or physical space, a system with a national scale like this, would be impossible if done only by the project team. Within this scenario, we propose the modeling and implementation of a tool that allows editing of virtual spaces in an easy and intuitive as compared with available tools. Within the context of GT-MV, we apply the SAMVC (Sistema de Autoria de Museus Virtuais Colaborativos) to museums where curators build the museum from a 3D floor plan (2D). The system, from these twodimensional information, recreates the equivalent in three dimensions. With this, through little or no training, team members from each museum may be responsible for updating the system
Resumo:
Considering the transition from industrial society to information society, we realize that the digital training that is addressed is currently insufficient to navigate within a digitized reality. As proposed to minimize this problem, this paper assesses, validates and develops the software RoboEduc to work with educational robotics with the main differential programming of robotic devices in levels, considering the specifics of reality training . One of the emphases of this work isthe presentation of materials and procedures involving the development, analysis and evolution of this software. For validation of usability tests were performed, based on analysis of these tests was developed version 4.0 of RoboEduc
Resumo:
This work presents the design and construction of an X-Y table of two degrees of freedom, as well as the development of a fuzzy system for its position and trajectory control. The table is composed of two bases that move perpendicularly to each other in the horizontal plane, and are driven by two DC motors. Base position is detected by position sensors attached to the motor axes. A data acquisition board performs the interface between a laptop and the plant. The fuzzy system algorithm was implemented in LabVIEW® programming environment that processes the sensors signals and determines the control variables values that drive the motors. Experimental results using position reference signals (step type signal) and straight and circular paths reference signals are presented to demonstrate the dynamic behavior of fuzzy system
Resumo:
Due to the large amount of television content, which emerged from the Digital TV, viewers are facing a new challenge, how to find interesting content intuitively and efficiently. The Personalized Electronic Programming Guides (pEPG) arise as an answer to this complex challenge. We propose TrendTV a layered architecture that allows the formation of social networks among viewers of Interactive Digital TV based on online microblogging. Associated with a pEPG, this social network allows the viewer to perform content filtering on a particular subject from the indications made by other viewers of his network. Allowing the viewer to create his own indications for a particular content when it is displayed, or to analyze the importance of a particular program online, based on these indications. This allows any user to perform filtering on content and generate or exchange information with other users in a flexible and transparent way, using several different devices (TVs, Smartphones, Tablets or PCs). Moreover, this architecture defines a mechanism to perform the automatic exchange of channels based on the best program that is showing at the moment, suggesting new components to be added to the middleware of the Brazilian Digital TV System (Ginga). The result is a constructed and dynamic database containing the classification of several TV programs as well as an application to automatically switch to the best channel of the moment
Resumo:
This work shows a study about the Generalized Predictive Controllers with Restrictions and their implementation in physical plants. Three types of restrictions will be discussed: restrictions in the variation rate of the signal control, restrictions in the amplitude of the signal control and restrictions in the amplitude of the Out signal (plant response). At the predictive control, the control law is obtained by the minimization of an objective function. To consider the restrictions, this minimization of the objective function is done by the use of a method to solve optimizing problems with restrictions. The chosen method was the Rosen Algorithm (based on the Gradient-projection). The physical plants in this study are two didactical systems of water level control. The first order one (a simple tank) and another of second order, which is formed by two tanks connected in cascade. The codes are implemented in C++ language and the communication with the system to be done through using a data acquisition panel offered by the system producer
Resumo:
The present work shows the development and construction of a robot manipulator with two rotary joints and two degrees of freedom, driven by three-phase induction motors. The positions of the arm and base are made, for comparison, by a fuzzy controller and a PID controller implemented in LabVIEW® programming environment. The robot manipulator moves in an area equivalent to a quarter of a sphere. Experimental results have shown that the fuzzy controller has superior performance to PID controller when tracking single and multiple step trajectories, for the cases of load and no load
Resumo:
The last years have presented an increase in the acceptance and adoption of the parallel processing, as much for scientific computation of high performance as for applications of general intention. This acceptance has been favored mainly for the development of environments with massive parallel processing (MPP - Massively Parallel Processing) and of the distributed computation. A common point between distributed systems and MPPs architectures is the notion of message exchange, that allows the communication between processes. An environment of message exchange consists basically of a communication library that, acting as an extension of the programming languages that allow to the elaboration of applications parallel, such as C, C++ and Fortran. In the development of applications parallel, a basic aspect is on to the analysis of performance of the same ones. Several can be the metric ones used in this analysis: time of execution, efficiency in the use of the processing elements, scalability of the application with respect to the increase in the number of processors or to the increase of the instance of the treat problem. The establishment of models or mechanisms that allow this analysis can be a task sufficiently complicated considering parameters and involved degrees of freedom in the implementation of the parallel application. An joined alternative has been the use of collection tools and visualization of performance data, that allow the user to identify to points of strangulation and sources of inefficiency in an application. For an efficient visualization one becomes necessary to identify and to collect given relative to the execution of the application, stage this called instrumentation. In this work it is presented, initially, a study of the main techniques used in the collection of the performance data, and after that a detailed analysis of the main available tools is made that can be used in architectures parallel of the type to cluster Beowulf with Linux on X86 platform being used libraries of communication based in applications MPI - Message Passing Interface, such as LAM and MPICH. This analysis is validated on applications parallel bars that deal with the problems of the training of neural nets of the type perceptrons using retro-propagation. The gotten conclusions show to the potentiality and easinesses of the analyzed tools.
Resumo:
This work uses computer vision algorithms related to features in the identification of medicine boxes for the visually impaired. The system is for people who have a disease that compromises his vision, hindering the identification of the correct medicine to be ingested. We use the camera, available in several popular devices such as computers, televisions and phones, to identify the box of the correct medicine and audio through the image, showing the poor information about the medication, such: as the dosage, indication and contraindications of the medication. We utilize a model of object detection using algorithms to identify the features in the boxes of drugs and playing the audio at the time of detection of feauteres in those boxes. Experiments carried out with 15 people show that where 93 % think that the system is useful and very helpful in identifying drugs for boxes. So, it is necessary to make use of this technology to help several people with visual impairments to take the right medicine, at the time indicated in advance by the physician
Resumo:
The seismic method is of extreme importance in geophysics. Mainly associated with oil exploration, this line of research focuses most of all investment in this area. The acquisition, processing and interpretation of seismic data are the parts that instantiate a seismic study. Seismic processing in particular is focused on the imaging that represents the geological structures in subsurface. Seismic processing has evolved significantly in recent decades due to the demands of the oil industry, and also due to the technological advances of hardware that achieved higher storage and digital information processing capabilities, which enabled the development of more sophisticated processing algorithms such as the ones that use of parallel architectures. One of the most important steps in seismic processing is imaging. Migration of seismic data is one of the techniques used for imaging, with the goal of obtaining a seismic section image that represents the geological structures the most accurately and faithfully as possible. The result of migration is a 2D or 3D image which it is possible to identify faults and salt domes among other structures of interest, such as potential hydrocarbon reservoirs. However, a migration fulfilled with quality and accuracy may be a long time consuming process, due to the mathematical algorithm heuristics and the extensive amount of data inputs and outputs involved in this process, which may take days, weeks and even months of uninterrupted execution on the supercomputers, representing large computational and financial costs, that could derail the implementation of these methods. Aiming at performance improvement, this work conducted the core parallelization of a Reverse Time Migration (RTM) algorithm, using the parallel programming model Open Multi-Processing (OpenMP), due to the large computational effort required by this migration technique. Furthermore, analyzes such as speedup, efficiency were performed, and ultimately, the identification of the algorithmic scalability degree with respect to the technological advancement expected by future processors