974 resultados para Software-reconfigurable array processing architectures
Resumo:
In a leading service economy like India, services lie at the very center of economic activity. Competitive organizations now look not only at the skills and knowledge, but also at the behavior required by an employee to be successful on the job. Emotionally competent employees can effectively deal with occupational stress and maintain psychological well-being. This study explores the scope of the first two formants and jitter to assess seven common emotional states present in the natural speech in English. The k-means method was used to classify emotional speech as neutral, happy, surprised, angry, disgusted and sad. The accuracy of classification obtained using raw jitter was more than 65 percent for happy and sad but less accurate for the others. The overall classification accuracy was 72% in the case of preprocessed jitter. The experimental study was done on 1664 English utterances of 6 females. This is a simple, interesting and more proactive method for employees from varied backgrounds to become aware of their own communication styles as well as that of their colleagues' and customers and is therefore socially beneficial. It is a cheap method also as it requires only a computer. Since knowledge of sophisticated software or signal processing is not necessary, it is easy to analyze
Resumo:
To coordinate ambulances for emergency medical services, a multiagent system uses an auction mechanism based on trust. Results of tests using real data show that this system can efficiently assign ambulances to patients, thereby reducing transportation time. Emergency transportation on specialized vehicles is needed when a person's health is in risk of irreparable damage. A patient can't benefit from sophisticated medical treatments and technologies if she or he isn't placed in a proper healthcare center with the appropriate medical team. For example, strokes are neurological emergencies involving a limited amount of time in which treatment measures are effective
Resumo:
One of the current major concerns in engineering is the development of aircrafts that have low power consumption and high performance. So, airfoils that have a high value of Lift Coefficient and a low value for the Drag Coefficient, generating a High-Efficiency airfoil are studied and designed. When the value of the Efficiency increases, the aircraft s fuel consumption decreases, thus improving its performance. Therefore, this work aims to develop a tool for designing of airfoils from desired characteristics, as Lift and Drag coefficients and the maximum Efficiency, using an algorithm based on an Artificial Neural Network (ANN). For this, it was initially collected an aerodynamic characteristics database, with a total of 300 airfoils, from the software XFoil. Then, through the software MATLAB, several network architectures were trained, between modular and hierarchical, using the Back-propagation algorithm and the Momentum rule. For data analysis, was used the technique of cross- validation, evaluating the network that has the lowest value of Root Mean Square (RMS). In this case, the best result was obtained for a hierarchical architecture with two modules and one layer of hidden neurons. The airfoils developed for that network, in the regions of lower RMS, were compared with the same airfoils imported into the software XFoil
Resumo:
Uma arquitetura reconfigurável e multiprocessada para a implementação física de Redes de Petri foi desenvolvida em VHDL e mapeada sobre um FPGA. Convencionalmente, as Redes de Petri são transformadas em uma linguagem de descrição de hardware no nível de transferências entre registradores e um processo de síntese de alto nível é utilizado para gerar as funções booleanas e tabelas de transição de estado para que se possa, finalmente, mapeá-las num FPGA (Morris et al., 2000) (Soto and Pereira, 2001). A arquitetura proposta possui blocos lógicos reconfiguráveis desenvolvidos exclusivamente para a implementação dos lugares e das transições da rede, não sendo necessária a descrição da rede em níveis de abstração intermediários e nem a utilização de um processo de síntese para realizar o mapeamento da rede na arquitetura. A arquitetura permite o mapeamento de modelos de Redes de Petri com diferenciação entre as marcas e associação de tempo no disparo das transições, sendo composta por um arranjo de processadores reconfiguráveis, cada um dos quais representando o comportamento de uma transição da Rede de Petri a ser mapeada e por um sistema de comunicação, implementado por um conjunto de roteadores que são capazes de enviar pacotes de dados de um processador reconfigurável a outro. A arquitetura proposta foi validada num FPGA de 10.570 elementos lógicos com uma topologia que permitiu a implementação de Redes de Petri de até 9 transições e 36 lugares, atingindo uma latência de 15,4ns e uma vazão de até 17,12GB/s com uma freqüência de operação de 64,58MHz.
Resumo:
Artificial Neural Networks are widely used in various applications in engineering, as such solutions of nonlinear problems. The implementation of this technique in reconfigurable devices is a great challenge to researchers by several factors, such as floating point precision, nonlinear activation function, performance and area used in FPGA. The contribution of this work is the approximation of a nonlinear function used in ANN, the popular hyperbolic tangent activation function. The system architecture is composed of several scenarios that provide a tradeoff of performance, precision and area used in FPGA. The results are compared in different scenarios and with current literature on error analysis, area and system performance. © 2013 IEEE.
Resumo:
The trade fair industry has great relevance in the national and international economic environment and it is constantly expanding. The general objective of this work is to analyze the main linguistic variants found in the trade fair terminological set. Our research is based on the theories of Cabré (1993, 1999), Barros (2004), Krieger & Finatto (2004), Alves (2007), Barbosa (2009), Dubuc (1985), Berber Sardinha (2004), Babini (2006) and Faulstich (1998, 2001). For this work we constitute two corpora of specialized texts, one for English language and another for Portuguese language. Successively we performed a collection of terms using software for corpora processing. These terms were organized into two notional systems, one in English and another in Portuguese. Then we analyzed the main types of variants found in our terminological set. Among them, those who had higher productivity are the lexical variants, followed by graphical, morphological and syntactic variants.
Resumo:
In a statistical inference scenario, the estimation of target signal or its parameters is done by processing data from informative measurements. The estimation performance can be enhanced if we choose the measurements based on some criteria that help to direct our sensing resources such that the measurements are more informative about the parameter we intend to estimate. While taking multiple measurements, the measurements can be chosen online so that more information could be extracted from the data in each measurement process. This approach fits well in Bayesian inference model often used to produce successive posterior distributions of the associated parameter. We explore the sensor array processing scenario for adaptive sensing of a target parameter. The measurement choice is described by a measurement matrix that multiplies the data vector normally associated with the array signal processing. The adaptive sensing of both static and dynamic system models is done by the online selection of proper measurement matrix over time. For the dynamic system model, the target is assumed to move with some distribution and the prior distribution at each time step is changed. The information gained through adaptive sensing of the moving target is lost due to the relative shift of the target. The adaptive sensing paradigm has many similarities with compressive sensing. We have attempted to reconcile the two approaches by modifying the observation model of adaptive sensing to match the compressive sensing model for the estimation of a sparse vector.
Resumo:
An analytical method for evaluating the uncertainty of the performance of active antenna arrays in the whole spatial spectrum is presented. Since array processing algorithms based on spatial reference are widely used to track moving targets, it is essential to be aware of the impact of the uncertainty sources on the antenna response. Furthermore, the estimation of the direction of arrival (DOA) depends on the array uncertainty. The aim of the uncertainties analysis is to provide an exhaustive characterization of the behavior of the active antenna array associated with its main uncertainty sources. The result of this analysis helps to select the proper calibration technique to be implemented. An illustrative example for a triangular antenna array used for satellite tracking is presented showing the suitability of the proposed method to carry out an efficient characterization of an active antenna array.
Resumo:
Nas últimas décadas, a poluição sonora tornou-se um grande problema para a sociedade. É por esta razão que a indústria tem aumentado seus esforços para reduzir a emissão de ruído. Para fazer isso, é importante localizar quais partes das fontes sonoras são as que emitem maior energia acústica. Conhecer os pontos de emissão é necessário para ter o controle das mesmas e assim poder reduzir o impacto acústico-ambiental. Técnicas como \"beamforming\" e \"Near-Field Acoustic Holography\" (NAH) permitem a obtenção de imagens acústicas. Essas imagens são obtidas usando um arranjo de microfones localizado a uma distância relativa de uma fonte emissora de ruído. Uma vez adquiridos os dados experimentais pode-se obter a localização e magnitude dos principais pontos de emissão de ruído. Do mesmo modo, ajudam a localizar fontes aeroacústicas e vibro acústicas porque são ferramentas de propósito geral. Usualmente, estes tipos de fontes trabalham em diferentes faixas de frequência de emissão. Recentemente, foi desenvolvida a transformada de Kronecker para arranjos de microfones, a qual fornece uma redução significativa do custo computacional quando aplicada a diversos métodos de reconstrução de imagens, desde que os microfones estejam distribuídos em um arranjo separável. Este trabalho de mestrado propõe realizar medições com sinais reais, usando diversos algoritmos desenvolvidos anteriormente em uma tese de doutorado, quanto à qualidade do resultado obtido e à complexidade computacional, e o desenvolvimento de alternativas para tratamento de dados quando alguns microfones do arranjo apresentarem defeito. Para reduzir o impacto de falhas em microfones e manter a condição de que o arranjo seja separável, foi desenvolvida uma alternativa para utilizar os algoritmos rápidos, eliminando-se apenas os microfones com defeito, de maneira que os resultados finais serão obtidos levando-se em conta todos os microfones do arranjo.
Resumo:
Uma grande diversidade de macrofibras poliméricas para reforço de concreto se encontram disponibilizadas hoje em dia. Por natureza estas fibras apresentam grande diversidade de características e propriedades. Estas variações afetam sua atuação como reforço no concreto. No entanto, não há normas brasileiras sobre o assunto e as metodologias de caracterização de normas estrangeiras apresentam divergências. Algumas normas definem que a caracterização do comportamento mecânico deva ser feita nos fios originais e outras que se devam utilizar métodos definidos para caracterização de materiais metálicos. A norma EN14889-2:2006 apresenta maior abrangência, mas deixa dúvidas quanto à adequação dos critérios de caracterização geométrica das fibras e não define um método de ensaio específico para sua caracterização mecânica. Assim, há a necessidade de estabelecimento de uma metodologia que permita a realização de um programa de controle de qualidade da fibra nas condições de emprego. Esta metodologia também proporcionaria uma forma de caracterização do material para estudos experimentais, o que permitiria maior fundamentação científica desses trabalhos que, frequentemente, fundamentam-se apenas em dados dos fabricantes. Assim, foi desenvolvido um estudo experimental focando a caracterização de duas macrofibras poliméricas disponíveis no mercado brasileiro. Focou-se o estudo na determinação dos parâmetros geométricos e na caracterização mecânica através da determinação da resistência à tração e avaliação do módulo de elasticidade. Na caracterização geométrica foi adotada como referência a norma europeia EN14889-2:2006. As medições do comprimento se efetuaram por dois métodos: o método do paquímetro e o método de análise de imagens digitais, empregando um software para processamento das imagens. Para a medição do diâmetro, além das metodologias mencionadas, foi usado o método da densidade. Conclui-se que o método do paquímetro, com o cuidado de esticar previamente as macrofibras, e o método das imagens digitais podem ser igualmente utilizados para medir o comprimento. Já parar determinar o diâmetro, recomenda-se o método da densidade. Quanto à caracterização mecânica, foi desenvolvida uma metodologia própria a partir de informações obtidas de outros ensaios. Assim, efetuaram-se ensaios de tração direta nas macrofibras coladas em molduras de tecido têxtil. Complementarmente, foi avaliado também o efeito do contato abrasivo das macrofibras com os agregados durante a mistura em betoneira no comportamento mecânico do material. Também se avaliou o efeito do método de determinação da área da seção transversal nos resultados medidos no ensaio de tração da fibra. Conclui-se que o método proposto para o ensaio de tração direta da fibra é viável, especialmente para a determinação da resistência à tração. O valor do módulo de elasticidade, por sua vez, acaba sendo subestimado. A determinação da área da seção da fibra através do método da densidade forneceu também os melhores resultados. Além disso, comprovou-se que o atrito das fibras com o agregado durante a mistura compromete o comportamento mecânico, reduzindo tanto a resistência quanto o módulo de elasticidade. Assim, pode-se afirmar que a metodologia proposta para o controle geométrico e mecânico das macrofibras poliméricas é adequada para a caracterização do material.
Resumo:
The paper describes three software packages - the main components of a software system for processing and web-presentation of Bulgarian language resources – parallel corpora and bilingual dictionaries. The author briefly presents current versions of the core components “Dictionary” and “Corpus” as well as the recently developed component “Connection” that links both “Dictionary” and “Corpus”. The components main functionalities are described as well. Some examples of the usage of the system’s web-applications are included.
Resumo:
Human motion monitoring is an important function in numerous applications. In this dissertation, two systems for monitoring motions of multiple human targets in wide-area indoor environments are discussed, both of which use radio frequency (RF) signals to detect, localize, and classify different types of human motion. In the first system, a coherent monostatic multiple-input multiple-output (MIMO) array is used, and a joint spatial-temporal adaptive processing method is developed to resolve micro-Doppler signatures at each location in a wide-area for motion mapping. The downranges are obtained by estimating time-delays from the targets, and the crossranges are obtained by coherently filtering array spatial signals. Motion classification is then applied to each target based on micro-Doppler analysis. In the second system, multiple noncoherent multistatic transmitters (Tx's) and receivers (Rx's) are distributed in a wide-area, and motion mapping is achieved by noncoherently combining bistatic range profiles from multiple Tx-Rx pairs. Also, motion classification is applied to each target by noncoherently combining bistatic micro-Doppler signatures from multiple Tx-Rx pairs. For both systems, simulation and real data results are shown to demonstrate the ability of the proposed methods for monitoring patient repositioning activities for pressure ulcer prevention.
Resumo:
Dissertação (mestrado)—Universidade de Brasília, Faculdade de Tecnologia, Departamento de Engenharia Elétrica, 2015.
Resumo:
General-purpose computing devices allow us to (1) customize computation after fabrication and (2) conserve area by reusing expensive active circuitry for different functions in time. We define RP-space, a restricted domain of the general-purpose architectural space focussed on reconfigurable computing architectures. Two dominant features differentiate reconfigurable from special-purpose architectures and account for most of the area overhead associated with RP devices: (1) instructions which tell the device how to behave, and (2) flexible interconnect which supports task dependent dataflow between operations. We can characterize RP-space by the allocation and structure of these resources and compare the efficiencies of architectural points across broad application characteristics. Conventional FPGAs fall at one extreme end of this space and their efficiency ranges over two orders of magnitude across the space of application characteristics. Understanding RP-space and its consequences allows us to pick the best architecture for a task and to search for more robust design points in the space. Our DPGA, a fine- grained computing device which adds small, on-chip instruction memories to FPGAs is one such design point. For typical logic applications and finite- state machines, a DPGA can implement tasks in one-third the area of a traditional FPGA. TSFPGA, a variant of the DPGA which focuses on heavily time-switched interconnect, achieves circuit densities close to the DPGA, while reducing typical physical mapping times from hours to seconds. Rigid, fabrication-time organization of instruction resources significantly narrows the range of efficiency for conventional architectures. To avoid this performance brittleness, we developed MATRIX, the first architecture to defer the binding of instruction resources until run-time, allowing the application to organize resources according to its needs. Our focus MATRIX design point is based on an array of 8-bit ALU and register-file building blocks interconnected via a byte-wide network. With today's silicon, a single chip MATRIX array can deliver over 10 Gop/s (8-bit ops). On sample image processing tasks, we show that MATRIX yields 10-20x the computational density of conventional processors. Understanding the cost structure of RP-space helps us identify these intermediate architectural points and may provide useful insight more broadly in guiding our continual search for robust and efficient general-purpose computing structures.
Resumo:
This study shows the implementation and the embedding of an Artificial Neural Network (ANN) in hardware, or in a programmable device, as a field programmable gate array (FPGA). This work allowed the exploration of different implementations, described in VHDL, of multilayer perceptrons ANN. Due to the parallelism inherent to ANNs, there are disadvantages in software implementations due to the sequential nature of the Von Neumann architectures. As an alternative to this problem, there is a hardware implementation that allows to exploit all the parallelism implicit in this model. Currently, there is an increase in use of FPGAs as a platform to implement neural networks in hardware, exploiting the high processing power, low cost, ease of programming and ability to reconfigure the circuit, allowing the network to adapt to different applications. Given this context, the aim is to develop arrays of neural networks in hardware, a flexible architecture, in which it is possible to add or remove neurons, and mainly, modify the network topology, in order to enable a modular network of fixed-point arithmetic in a FPGA. Five synthesis of VHDL descriptions were produced: two for the neuron with one or two entrances, and three different architectures of ANN. The descriptions of the used architectures became very modular, easily allowing the increase or decrease of the number of neurons. As a result, some complete neural networks were implemented in FPGA, in fixed-point arithmetic, with a high-capacity parallel processing