792 resultados para communication performance evaluation
Resumo:
This study reports the development and performance evaluation of prototypes of biogas-fuelled stationary power generators in the range of 1 kW. Strategies to achieve high engine efficiency namely pulsed manifold injection, electronic throttle control and dual spark plugs, have been incorporated in the prototype. A complete closed-loop control of the engine operation to maintain a steady engine speed of 3000 rpm (+/- 5%) across the entire load range while maintaining an optimum fuel-air equivalence ratio is made possible by an electronic control unit (ECU) controlling the injection duration, ignition timing and throttle position. This study specifically focuses on the response of the generator to transient loads, and the overall efficiency obtained. The results obtained from testing the prototype have been found to be satisfactory and show that biogas power generators for low power applications can be made efficient (overall efficiency of 19% at electrical load of 640 W) using the strategies of biogas fuel injection.
Resumo:
[EN]Fundación Zain is developing new built heritage assessment protocols. The goal is to objectivize and standardize the analysis and decision process that leads to determining the degree of protection of built heritage in the Basque Country. The ultimate step in this objectivization and standardization effort will be the development of an information and communication technology (ICT) tool for the assessment of built heritage. This paper presents the ground work carried out to make this tool possible: the automatic, image-based delineation of stone masonry. This is a necessary first step in the development of the tool, as the built heritage that will be assessed consists of stone masonry construction, and many of the features analyzed can be characterized according to the geometry and arrangement of the stones. Much of the assessment is carried out through visual inspection. Thus, this process will be automated by applying image processing on digital images of the elements under inspection. The principal contribution of this paper is the automatic delineation the framework proposed. The other contribution is the performance evaluation of this delineation as the input to a classifier for a geometrically characterized feature of a built heritage object. The element chosen to perform this evaluation is the stone arrangement of masonry walls. The validity of the proposed framework is assessed on real images of masonry walls.
Resumo:
Fundacion Zain is developing new built heritage assessment protocols. The goal is to objectivize and standardize the analysis and decision process that leads to determining the degree of protection of built heritage in the Basque Country. The ultimate step in this objectivization and standardization effort will be the development of an information and communication technology (ICT) tool for the assessment of built heritage. This paper presents the ground work carried out to make this tool possible: the automatic, image-based delineation of stone masonry. This is a necessary first step in the development of the tool, as the built heritage that will be assessed consists of stone masonry construction, and many of the features analyzed can be characterized according to the geometry and arrangement of the stones. Much of the assessment is carried out through visual inspection. Thus, this process will be automated by applying image processing on digital images of the elements under inspection. The principal contribution of this paper is the automatic delineation the framework proposed. The other contribution is the performance evaluation of this delineation as the input to a classifier for a geometrically characterized feature of a built heritage object. The element chosen to perform this evaluation is the stone arrangement of masonry walls. The validity of the proposed framework is assessed on real images of masonry walls.
Resumo:
A realização da Internet das Coisas (Internet of Things, IoT) requer a integração e interação de dispositivos e serviços com protocolos de comunicação heterogêneos. Os dados gerados pelos dispositivos precisam ser analisados e interpretados em concordância com um modelo de dados em comum, o que pode ser solucionado com o uso de tecnologias de modelagem semântica, processamento, raciocínio e persistência de dados. A computação ciente de contexto possui soluções para estes desafios com mecanismos que associam os dados de contexto com dados coletados pelos dispositivos. Entretanto, a IoT precisa ir além da computação ciente de contexto, sendo simultaneamente necessário soluções para aspectos de segurança, privacidade e escalabilidade. Para integração destas tecnologias é necessário o suporte de uma infraestrutura, que pode ser implementada como um middleware. No entanto, uma solução centralizada de integração de dispositivos heterogêneos pode afetar escalabilidade. Assim esta integração é delegada para agentes de software, que são responsáveis por integrar os dispositivos e serviços, encapsulando as especificidades das suas interfaces e protocolos de comunicação. Neste trabalho são explorados os aspectos de segurança, persistência e nomeação para agentes de recursos. Para este fim foi desenvolvido o ContQuest, um framework, que facilita a integração de novos recursos e o desenvolvimento de aplicações cientes de contexto para a IoT, através de uma arquitetura de serviços e um modelo de dados. O ContQuest inclui soluções consistentes para os aspectos de persistência, segurança e controle de acesso tanto para os serviços de middleware, como para os Agentes de Recursos, que encapsulam dispositivos e serviços, e aplicações-clientes. O ContQuest utiliza OWL para a modelagem dos recursos e inclui um mecanismo de geração de identificadores únicos universais nas ontologias. Um protótipo do ContQuest foi desenvolvido e validado com a integração de três Agentes de Recurso para dispositivos reais: um dispositivo Arduino, um leitor de RFID e uma rede de sensores. Foi também realizado um experimento para avaliação de desempenho dos componentes do sistema, em que se observou o impacto do mecanismo de segurança proposto no desempenho do protótipo. Os resultados da validação e do desempenho são satisfatórios
Resumo:
This paper presents the first performance evaluation of interest points on scalar volumetric data. Such data encodes 3D shape, a fundamental property of objects. The use of another such property, texture (i.e. 2D surface colouration), or appearance, for object detection, recognition and registration has been well studied; 3D shape less so. However, the increasing prevalence of depth sensors and the diminishing returns to be had from appearance alone have seen a surge in shape-based methods. In this work we investigate the performance of several detectors of interest points in volumetric data, in terms of repeatability, number and nature of interest points. Such methods form the first step in many shape-based applications. Our detailed comparison, with both quantitative and qualitative measures on synthetic and real 3D data, both point-based and volumetric, aids readers in selecting a method suitable for their application. © 2011 IEEE.
Resumo:
This thesis study the problem of work group effectiveness and group job design according to extensive literature investigation and the analysis of realistic background. The whole research consists of four parts: (1) The evaluation of work group effectiveness, the aim is to search to criteria that can describe and analysis work group, and explore the cognitive dimensions of effectiveness of Chinese subjects; (2) The study on the relationship between group job characteristic and effectiveness, the aim is to find the general correlation between work group characteristic and effectiveness, and try to search the most important core variable; (3) The study on the preference for the way of group work; (4) The study of the relationship between group composition and group effectiveness, try to examine how different approaches of personnel selection influence work results. The results indicate: (1) The evaluation of group effectiveness mainly consists of two dimensions: performance and the employee's attitude and feelings toward the group, so we can use these two dimensions and corresponding criteria as the standard of effectiveness evaluation. (2) According to the analysis of related literature, we can determine work group characteristic from five aspects: job design, the interdependence among members, group composition, organizational background, group process. (3) Experimental study find that different group job characteristics have different relationship models with effectiveness criteria. Job design, the interdependence among group members, group composition have significant correlation relationships with two kinds of effectiveness criteria; organizational background mainly has relationships with satisfaction criteria; group process mainly has relationships with performance criteria. (4) The choice for people to select the way of group work has the consistency, that is people prefer to "Self--managed work groups"; but different groups have the difference, the main group dimension is the difference between group members and group leaders. (5) Work groups which were composed accordingly to interpersonal attraction have higher levels of communication, coordination, group cohesion and job satisfaction than ability--based groups based. But the performance evaluation under these two conditions has no difference. The thesis analysis and discuss the research results, also point out several questions that needed to be explored further and the possible research directions in the future. This study has some reference value in group--level performance appraisal and reward design, the content and method of group--level job design, and personnel selection of group, etc.
Resumo:
K. Rasmani and Q. Shen. Subsethood-based Fuzzy Rule Models and their Application to Student Performance Classification. Proceedings of the 14th International Conference on Fuzzy Systems, pages 755-760, 2005.
Resumo:
The popularity of TCP/IP coupled with the premise of high speed communication using Asynchronous Transfer Mode (ATM) technology have prompted the network research community to propose a number of techniques to adapt TCP/IP to ATM network environments. ATM offers Available Bit Rate (ABR) and Unspecified Bit Rate (UBR) services for best-effort traffic, such as conventional file transfer. However, recent studies have shown that TCP/IP, when implemented using ABR or UBR, leads to serious performance degradations, especially when the utilization of network resources (such as switch buffers) is high. Proposed techniques-switch-level enhancements, for example-that attempt to patch up TCP/IP over ATMs have had limited success in alleviating this problem. The major reason for TCP/IP's poor performance over ATMs has been consistently attributed to packet fragmentation, which is the result of ATM's 53-byte cell-oriented switching architecture. In this paper, we present a new transport protocol, TCP Boston, that turns ATM's 53-byte cell-oriented switching architecture into an advantage for TCP/IP. At the core of TCP Boston is the Adaptive Information Dispersal Algorithm (AIDA), an efficient encoding technique that allows for dynamic redundancy control. AIDA makes TCP/IP's performance less sensitive to cell losses, thus ensuring a graceful degradation of TCP/IP's performance when faced with congested resources. In this paper, we introduce AIDA and overview the main features of TCP Boston. We present detailed simulation results that show the superiority of our protocol when compared to other adaptations of TCP/IP over ATMs. In particular, we show that TCP Boston improves TCP/IP's performance over ATMs for both network-centric metrics (e.g., effective throughput) and application-centric metrics (e.g., response time).
Resumo:
While ATM bandwidth-reservation techniques are able to offer the guarantees necessary for the delivery of real-time streams in many applications (e.g. live audio and video), they suffer from many disadvantages that make them inattractive (or impractical) for many others. These limitations coupled with the flexibility and popularity of TCP/IP as a best-effort transport protocol have prompted the network research community to propose and implement a number of techniques that adapt TCP/IP to the Available Bit Rate (ABR) and Unspecified Bit Rate (UBR) services in ATM network environments. This allows these environments to smoothly integrate (and make use of) currently available TCP-based applications and services without much (if any) modifications. However, recent studies have shown that TCP/IP, when implemented over ATM networks, is susceptible to serious performance limitations. In a recently completed study, we have unveiled a new transport protocol, TCP Boston, that turns ATM's 53-byte cell-oriented switching architecture into an advantage for TCP/IP. In this paper, we demonstrate the real-time features of TCP Boston that allow communication bandwidth to be traded off for timeliness. We start with an overview of the protocol. Next, we analytically characterize the dynamic redundancy control features of TCP Boston. Next, We present detailed simulation results that show the superiority of our protocol when compared to other adaptations of TCP/IP over ATMs. In particular, we show that TCP Boston improves TCP/IP's performance over ATMs for both network-centric metrics (e.g., effective throughput and percent of missed deadlines) and real-time application-centric metrics (e.g., response time and jitter).
Resumo:
High-speed networks, such as ATM networks, are expected to support diverse Quality of Service (QoS) constraints, including real-time QoS guarantees. Real-time QoS is required by many applications such as those that involve voice and video communication. To support such services, routing algorithms that allow applications to reserve the needed bandwidth over a Virtual Circuit (VC) have been proposed. Commonly, these bandwidth-reservation algorithms assign VCs to routes using the least-loaded concept, and thus result in balancing the load over the set of all candidate routes. In this paper, we show that for such reservation-based protocols|which allow for the exclusive use of a preset fraction of a resource's bandwidth for an extended period of time-load balancing is not desirable as it results in resource fragmentation, which adversely affects the likelihood of accepting new reservations. In particular, we show that load-balancing VC routing algorithms are not appropriate when the main objective of the routing protocol is to increase the probability of finding routes that satisfy incoming VC requests, as opposed to equalizing the bandwidth utilization along the various routes. We present an on-line VC routing scheme that is based on the concept of "load profiling", which allows a distribution of "available" bandwidth across a set of candidate routes to match the characteristics of incoming VC QoS requests. We show the effectiveness of our load-profiling approach when compared to traditional load-balancing and load-packing VC routing schemes.
Resumo:
To support the diverse Quality of Service (QoS) requirements of real-time (e.g. audio/video) applications in integrated services networks, several routing algorithms that allow for the reservation of the needed bandwidth over a Virtual Circuit (VC) established on one of several candidate routes have been proposed. Traditionally, such routing is done using the least-loaded concept, and thus results in balancing the load across the set of candidate routes. In a recent study, we have established the inadequacy of this load balancing practice and proposed the use of load profiling as an alternative. Load profiling techniques allow the distribution of "available" bandwidth across a set of candidate routes to match the characteristics of incoming VC QoS requests. In this paper we thoroughly characterize the performance of VC routing using load profiling and contrast it to routing using load balancing and load packing. We do so both analytically and via extensive simulations of multi-class traffic routing in Virtual Path (VP) based networks. Our findings confirm that for routing guaranteed bandwidth flows in VP networks, load balancing is not desirable as it results in VP bandwidth fragmentation, which adversely affects the likelihood of accepting new VC requests. This fragmentation is more pronounced when the granularity of VC requests is large. Typically, this occurs when a common VC is established to carry the aggregate traffic flow of many high-bandwidth real-time sources. For VP-based networks, our simulation results show that our load-profiling VC routing scheme performs better or as well as the traditional load-balancing VC routing in terms of revenue under both skewed and uniform workloads. Furthermore, load-profiling routing improves routing fairness by proactively increasing the chances of admitting high-bandwidth connections.
Resumo:
We examine the role of liquidity risk, both as a stock characteristic as well as systematic liquidity risk, in UK mutual fund performance for the first time. Using four alternative measures of stock liquidity we extract principal components across stocks in order to construct systematic or market liquidity factors. We find that on average UK mutual funds are tilted towards liquid stocks (except for small stock funds as might be expected) but that, counter-intuitively, liquidity as a stock characteristic is positively priced in the cross-section of fund performance. We find that systematic liquidity risk is positively priced in the cross-section of fund performance. Overall, our results reveal a strong role for stock liquidity level and systematic liquidity risk in fund performance evaluation models.
Resumo:
The electrical and communication performance of a 0.8-mu W UHF temperature telemeter designed for human vaginal placement is discussed; a solenoidal loop antenna was used, occupying a volume of 0.1 cm(3). In situ, measured power absorption was between 19-25 dB, resulting in an effective operating range of 10 m. Capacitive loading lowered the antenna's resonant frequency by 1.4% and there was a significant polarization change in the radiated output.
Resumo:
Per-core scratchpad memories (or local stores) allow direct inter-core communication, with latency and energy advantages over coherent cache-based communication, especially as CMP architectures become more distributed. We have designed cache-integrated network interfaces, appropriate for scalable multicores, that combine the best of two worlds – the flexibility of caches and the efficiency of scratchpad memories: on-chip SRAM is configurably shared among caching, scratchpad, and virtualized network interface (NI) functions. This paper presents our architecture, which provides local and remote scratchpad access, to either individual words or multiword blocks through RDMA copy. Furthermore, we introduce event responses, as a technique that enables software configurable communication and synchronization primitives. We present three event response mechanisms that expose NI functionality to software, for multiword transfer initiation, completion notifications for software selected sets of arbitrary size transfers, and multi-party synchronization queues. We implemented these mechanisms in a four-core FPGA prototype, and measure the logic overhead over a cache-only design for basic NI functionality to be less than 20%. We also evaluate the on-chip communication performance on the prototype, as well as the performance of synchronization functions with simulation of CMPs with up to 128 cores. We demonstrate efficient synchronization, low-overhead communication, and amortized-overhead bulk transfers, which allow parallelization gains for fine-grain tasks, and efficient exploitation of the hardware bandwidth.
Resumo:
Approximately half of the houses in Northern Ireland were built before any form of minimum thermal specification (U-value) or energy efficiency standard were available. At present, 44% of households are categorised as being in fuel poverty; spending more than 10% of the household income to heat the house to an acceptable level. This paper presents the results from long term performance monitoring of 4 case study houses that have undergone retrofits to improve energy efficiency in Northern Ireland. There is some uncertainty associated with some of the marketed retrofit measures in terms of their effectiveness in reducing energy usage and their potential to cause detrimental impacts on the internal environment of a house. Using wireless sensor technology internal conditions such as temperature and humidity were measured alongside gas and electricity usage for a year. External weather conditions were also monitored. The paper considers the effectiveness of the different retrofit measures implemented based on the long term data monitoring and short term building performance evaluation tests that were completed.