906 resultados para Experimental performance metrics
Resumo:
Mestrado em Contabilidade e Gestão das Instituições Financeiras
Resumo:
Trabalho Final de Mestrado para obtenção do grau de Mestre em Engenharia Química e Biológica
Resumo:
This paper focuses on evaluating the usability of an Intelligent Wheelchair (IW) in both real and simulated environments. The wheelchair is controlled at a high-level by a flexible multimodal interface, using voice commands, facial expressions, head movements and joystick as its main inputs. A Quasi-experimental design was applied including a deterministic sample with a questionnaire that enabled to apply the System Usability Scale. The subjects were divided in two independent samples: 46 individuals performing the experiment with an Intelligent Wheelchair in a simulated environment (28 using different commands in a sequential way and 18 with the liberty to choose the command); 12 individuals performing the experiment with a real IW. The main conclusion achieved by this study is that the usability of the Intelligent Wheelchair in a real environment is higher than in the simulated environment. However there were not statistical evidences to affirm that there are differences between the real and simulated wheelchairs in terms of safety and control. Also, most of users considered the multimodal way of driving the wheelchair very practical and satisfactory. Thus, it may be concluded that the multimodal interfaces enables very easy and safe control of the IW both in simulated and real environments.
Resumo:
The IEEE 802.15.4 has been adopted as a communication protocol standard for Low-Rate Wireless Private Area Networks (LRWPANs). While it appears as a promising candidate solution for Wireless Sensor Networks (WSNs), its adequacy must be carefully evaluated. In this paper, we analyze the performance limits of the slotted CSMA/CA medium access control (MAC) mechanism in the beacon-enabled mode for broadcast transmissions in WSNs. The motivation for evaluating the beacon-enabled mode is due to its flexibility and potential for WSN applications as compared to the non-beacon enabled mode. Our analysis is based on an accurate simulation model of the slotted CSMA/CA mechanism on top of a realistic physical layer, with respect to the IEEE 802.15.4 standard specification. The performance of the slotted CSMA/CA is evaluated and analyzed for different network settings to understand the impact of the protocol attributes (superframe order, beacon order and backoff exponent), the number of nodes and the data frame size on the network performance, namely in terms of throughput (S), average delay (D) and probability of success (Ps). We also analytically evaluate the impact of the slotted CSMA/CA overheads on the saturation throughput. We introduce the concept of utility (U) as a combination of two or more metrics, to determine the best offered load range for an optimal behavior of the network. We show that the optimal network performance using slotted CSMA/CA occurs in the range of 35% to 60% with respect to an utility function proportional to the network throughput (S) divided by the average delay (D).
Resumo:
Applications involving biosignals, such as Electrocardiography (ECG), are becoming more pervasive with the extension towards non-intrusive scenarios helping targeting ambulatory healthcare monitoring, emotion assessment, among many others. In this study we introduce a new type of silver/silver chloride (Ag/AgCl) electrodes based on a paper substrate and produced using an inkjet printing technique. This type of electrodes can increase the potential applications of biosignal acquisition technologies for everyday life use, given that there are several advantages, such as cost reduction and easier recycling, resultant from the approach explored in our work. We performed a comparison study to assess the quality of this new electrode type, in which ECG data was collected with three types of Ag/AgCl electrodes: i) gelled; ii) dry iii) paper-based inkjet printed. We also compared the performance of each electrode when acquired using a professional-grade gold standard device, and a low cost platform. Experimental results showed that data acquired using our proposed inkjet printed electrode is highly correlated with data obtained through conventional electrodes. Moreover, the electrodes are robust to high-end and low-end data acquisition devices. Copyright © 2014 SCITEPRESS - Science and Technology Publications. All rights reserved.
Resumo:
A unified architecture for fast and efficient computation of the set of two-dimensional (2-D) transforms adopted by the most recent state-of-the-art digital video standards is presented in this paper. Contrasting to other designs with similar functionality, the presented architecture is supported on a scalable, modular and completely configurable processing structure. This flexible structure not only allows to easily reconfigure the architecture to support different transform kernels, but it also permits its resizing to efficiently support transforms of different orders (e. g. order-4, order-8, order-16 and order-32). Consequently, not only is it highly suitable to realize high-performance multi-standard transform cores, but it also offers highly efficient implementations of specialized processing structures addressing only a reduced subset of transforms that are used by a specific video standard. The experimental results that were obtained by prototyping several configurations of this processing structure in a Xilinx Virtex-7 FPGA show the superior performance and hardware efficiency levels provided by the proposed unified architecture for the implementation of transform cores for the Advanced Video Coding (AVC), Audio Video coding Standard (AVS), VC-1 and High Efficiency Video Coding (HEVC) standards. In addition, such results also demonstrate the ability of this processing structure to realize multi-standard transform cores supporting all the standards mentioned above and that are capable of processing the 8k Ultra High Definition Television (UHDTV) video format (7,680 x 4,320 at 30 fps) in real time.
Resumo:
The reduction of the power loss generated in mechanical transmissions and the use of low friction biodegradable lubricants has been attracting considerable attention in recent times. Therefore, it is necessary to develop methods to test and evaluate the performance of such lubricants and compare them with conventional ones. In this sense, a Four-Ball Machine was modified allowing the test of rolling bearings. A 51107 thrust ball bearing was used to test two different greases and the corresponding base oils. Friction torque and operating temperatures were continuously monitored to quantify the power loss and the heat evacuation for each lubricant tested. Copyright © 2010 John Wiley & Sons, Ltd.
Resumo:
It is considered that using crushed recycled concrete as aggregate for concrete production is a viable alternative to dumping and would help to conserve abiotic resources. This use has fundamentally been based on the coarse fraction because the fine fraction is likely to degrade the performance of the resulting concrete. This paper presents results from a research work undertaken at Institut Superior Tecnico (IST), Lisbon, Portugal, in which the effects of incorporating two types of superplasticizer on the mechanical performance of concrete containing fine recycled aggregate were evaluated. The purpose was to see if the addition of superplasticizer would offset the detrimental effects associated with the use of fine recycled concrete aggregate. The experimental programme is described and the results of tests for splitting tensile strength, modulus of elasticity and abrasion resistance are presented. The relative performance of concrete made with recycled aggregate was found to decrease. However, the same concrete with admixtures in general exhibited a better mechanical performance than the reference mixes without admixtures or with a less active superplasticizer. Therefore, it is argued that the mechanical performance of concrete made with fine recycled concrete aggregates can be as good as that of conventional concrete, if superplasticizers are used to reduce the water-cement ratio of the former concrete.
Resumo:
Current Electrocardiographic (ECG) signal acquisition methods are generally highly intrusive, as they involve the use of pre-gelled electrodes and cabled sensors placed directly on the person, at the chest or limbs level. Moreover, systems that make use of alternative conductive materials to overcome this issue, only provide heart rate information and not the detailed signal itself. We present a comparison and evaluation of two types of dry electrodes as interface with the skin, targeting wearable and low intrusiveness applications, which enable ECG measurement without the need for any apparatus permanently fitted to the individual. In particular, our approach is targeted at ECG biometrics using signals collected at the hand or finger level. A custom differential circuit with virtual ground was also developed for enhanced usability. Our work builds upon the current stateof-the-art in sensoring devices and processing tools, and enables novel data acquisition settings through the use of dry electrodes. Experimental evaluation was performed for Ag/AgCl and Electrolycra materials, and results show that both materials exhibit adequate performance for the intended application.
Resumo:
Buildings account for 40% of total energy consumption in the European Union. The reduction of energy consumption in the buildings sector constitute an important measure needed to reduce the Union's energy dependency and greenhouse gas emissions. The Portuguese legislation incorporate this principles in order to regulate the energy performance of buildings. This energy performance should be accompanied by good conditions for the occupants of the buildings. According to EN 15251 (2007) the four factors that affect the occupant comfort in the buildings are: Indoor Air Quality (IAQ), thermal comfort, acoustics and lighting. Ventilation directly affects all except the lighting, so it is crucial to understand the performance of it. The ventilation efficiency concept therefore earn significance, because it is an attempt to quantify a parameter that can easily distinguish the different options for air diffusion in the spaces. The two indicators most internationally accepted are the Air Change Efficiency (ACE) and the Contaminant Removal Effectiveness (CRE). Nowadays with the developed of the Computational Fluid Dynamics (CFD) the behaviour of ventilation can be more easily predicted. Thirteen strategies of air diffusion were measured in a test chamber through the application of the tracer gas method, with the objective to validate the calculation by the MicroFlo module of the IES-VE software for this two indicators. The main conclusions from this work were: that the values of the numerical simulations are in agreement with experimental measurements; the value of the CRE is more dependent of the position of the contamination source, that the strategy used for the air diffusion; the ACE indicator is more appropriate for quantifying the quality of the air diffusion; the solutions to be adopted, to maximize the ventilation efficiency should be, the schemes that operate with low speeds of supply air and small differences between supply air temperature and the room temperature.
Resumo:
STRIPPING is a software application developed for the automatic design of a randomly packing column where the transfer of volatile organic compounds (VOCs) from water to air can be performed and to simulate it’s behaviour in a steady-state. This software completely purges any need of experimental work for the selection of diameter of the column, and allows a choice, a priori, of the most convenient hydraulic regime for this type of operation. It also allows the operator to choose the model used for the calculation of some parameters, namely between the Eckert/Robbins model and the Billet model for estimating the pressure drop of the gaseous phase, and between the Billet and Onda/Djebbar’s models for the mass transfer. Illustrations of the graphical interface offered are presented.
Resumo:
Resumo: Cement, as well as the remaining constituents of self-compacting mortars, must be carefully selected, in order to obtain an adequate composition with a granular mix as compact as possible and a good performance in the fresh state (self-compacting effect) and the hardened state (mechanical and durability-related behavior). Therefore in this work the possibility of incorporating nano particles in self-compacting mortars was studied. Nano materials are very reactive due mostly to their high specific surface and show a great potential to improve the properties of these mortars, both in mechanical and durability terms. In this work two nano materials were used, nano silica (nano SiO2) in colloidal state and nano titanium (nano TiO2) in amorphous state, in two types of self-compacting mortars (ratio binder:sand of 1:1 and 1:2). The self-compacting mortar mixes have the same water/cement ratio and 30% of replacement of cement with fly ashes. The influence of nano materials nano-SiO2 and nano-TiO2 on the fresh and hardened state properties of these self-compacting mortars was studied. The results show that the use of nano materials in repair and rehabilitation mortars has significant potential but still needs to be optimized. (C) 2015 Elsevier Ltd. All rights reserved.
Resumo:
Recent integrated circuit technologies have opened the possibility to design parallel architectures with hundreds of cores on a single chip. The design space of these parallel architectures is huge with many architectural options. Exploring the design space gets even more difficult if, beyond performance and area, we also consider extra metrics like performance and area efficiency, where the designer tries to design the architecture with the best performance per chip area and the best sustainable performance. In this paper we present an algorithm-oriented approach to design a many-core architecture. Instead of doing the design space exploration of the many core architecture based on the experimental execution results of a particular benchmark of algorithms, our approach is to make a formal analysis of the algorithms considering the main architectural aspects and to determine how each particular architectural aspect is related to the performance of the architecture when running an algorithm or set of algorithms. The architectural aspects considered include the number of cores, the local memory available in each core, the communication bandwidth between the many-core architecture and the external memory and the memory hierarchy. To exemplify the approach we did a theoretical analysis of a dense matrix multiplication algorithm and determined an equation that relates the number of execution cycles with the architectural parameters. Based on this equation a many-core architecture has been designed. The results obtained indicate that a 100 mm(2) integrated circuit design of the proposed architecture, using a 65 nm technology, is able to achieve 464 GFLOPs (double precision floating-point) for a memory bandwidth of 16 GB/s. This corresponds to a performance efficiency of 71 %. Considering a 45 nm technology, a 100 mm(2) chip attains 833 GFLOPs which corresponds to 84 % of peak performance These figures are better than those obtained by previous many-core architectures, except for the area efficiency which is limited by the lower memory bandwidth considered. The results achieved are also better than those of previous state-of-the-art many-cores architectures designed specifically to achieve high performance for matrix multiplication.
Resumo:
Trabalho de Projeto submetido à Escola Superior de Teatro e Cinema para cumprimento dos requisitos necessários à obtenção do grau de Mestre em TEATRO - especialização em Artes Performativas (Teatro-Música).
Resumo:
In the present paper we assess the performance of information-theoretic inspired risks functionals in multilayer perceptrons with reference to the two most popular ones, Mean Square Error and Cross-Entropy. The information-theoretic inspired risks, recently proposed, are: HS and HR2 are, respectively, the Shannon and quadratic Rényi entropies of the error; ZED is a risk reflecting the error density at zero errors; EXP is a generalized exponential risk, able to mimic a wide variety of risk functionals, including the information-thoeretic ones. The experiments were carried out with multilayer perceptrons on 35 public real-world datasets. All experiments were performed according to the same protocol. The statistical tests applied to the experimental results showed that the ubiquitous mean square error was the less interesting risk functional to be used by multilayer perceptrons. Namely, mean square error never achieved a significantly better classification performance than competing risks. Cross-entropy and EXP were the risks found by several tests to be significantly better than their competitors. Counts of significantly better and worse risks have also shown the usefulness of HS and HR2 for some datasets.