941 resultados para end user computing application streaming horizon workspace portalvmware view
Resumo:
Early definitions of Smart Building focused almost entirely on the technology aspect and did not suggest user interaction at all. Indeed, today we would attribute it more to the concept of the automated building. In this sense, control of comfort conditions inside buildings is a problem that is being well investigated, since it has a direct effect on users’ productivity and an indirect effect on energy saving. Therefore, from the users’ perspective, a typical environment can be considered comfortable, if it’s capable of providing adequate thermal comfort, visual comfort and indoor air quality conditions and acoustic comfort. In the last years, the scientific community has dealt with many challenges, especially from a technological point of view. For instance, smart sensing devices, the internet, and communication technologies have enabled a new paradigm called Edge computing that brings computation and data storage closer to the location where it is needed, to improve response times and save bandwidth. This has allowed us to improve services, sustainability and decision making. Many solutions have been implemented such as smart classrooms, controlling the thermal condition of the building, monitoring HVAC data for energy-efficient of the campus and so forth. Though these projects provide to the realization of smart campus, a framework for smart campus is yet to be determined. These new technologies have also introduced new research challenges: within this thesis work, some of the principal open challenges will be faced, proposing a new conceptual framework, technologies and tools to move forward the actual implementation of smart campuses. Keeping in mind, several problems known in the literature have been investigated: the occupancy detection, noise monitoring for acoustic comfort, context awareness inside the building, wayfinding indoor, strategic deployment for air quality and books preserving.
Resumo:
Embedding intelligence in extreme edge devices allows distilling raw data acquired from sensors into actionable information, directly on IoT end-nodes. This computing paradigm, in which end-nodes no longer depend entirely on the Cloud, offers undeniable benefits, driving a large research area (TinyML) to deploy leading Machine Learning (ML) algorithms on micro-controller class of devices. To fit the limited memory storage capability of these tiny platforms, full-precision Deep Neural Networks (DNNs) are compressed by representing their data down to byte and sub-byte formats, in the integer domain. However, the current generation of micro-controller systems can barely cope with the computing requirements of QNNs. This thesis tackles the challenge from many perspectives, presenting solutions both at software and hardware levels, exploiting parallelism, heterogeneity and software programmability to guarantee high flexibility and high energy-performance proportionality. The first contribution, PULP-NN, is an optimized software computing library for QNN inference on parallel ultra-low-power (PULP) clusters of RISC-V processors, showing one order of magnitude improvements in performance and energy efficiency, compared to current State-of-the-Art (SoA) STM32 micro-controller systems (MCUs) based on ARM Cortex-M cores. The second contribution is XpulpNN, a set of RISC-V domain specific instruction set architecture (ISA) extensions to deal with sub-byte integer arithmetic computation. The solution, including the ISA extensions and the micro-architecture to support them, achieves energy efficiency comparable with dedicated DNN accelerators and surpasses the efficiency of SoA ARM Cortex-M based MCUs, such as the low-end STM32M4 and the high-end STM32H7 devices, by up to three orders of magnitude. To overcome the Von Neumann bottleneck while guaranteeing the highest flexibility, the final contribution integrates an Analog In-Memory Computing accelerator into the PULP cluster, creating a fully programmable heterogeneous fabric that demonstrates end-to-end inference capabilities of SoA MobileNetV2 models, showing two orders of magnitude performance improvements over current SoA analog/digital solutions.
Resumo:
Continuum parallel robots (CPRs) are manipulators employing multiple flexible beams arranged in parallel and connected to a rigid end-effector. CPRs promise higher payload and accuracy than serial CRs while keeping great flexibility. As the risk of injury during accidental contacts between a human and a CPR should be reduced, CPRs may be used in large-scale collaborative tasks or assisted robotic surgery. There exist various CPR designs, but the prototype conception is rarely based on performance considerations, and the CPRs realization in mainly based on intuitions or rigid-link parallel manipulators architectures. This thesis focuses on the performance analysis of CPRs, and the tools needed for such evaluation, such as workspace computation algorithms. In particular, workspace computation strategies for CPRs are essential for the performance assessment, since the CPRs workspace may be used as a performance index or it can serve for optimal-design tools. Two new workspace computation algorithms are proposed in this manuscript, the former focusing on the workspace volume computation and the certification of its numerical results, while the latter aims at computing the workspace boundary only. Due to the elastic nature of CPRs, a key performance indicator for these robots is the stability of their equilibrium configurations. This thesis proposes the experimental validation of the equilibrium stability assessment on a real prototype, demonstrating limitations of some commonly used assumptions. Additionally, a performance index measuring the distance to instability is originally proposed in this manuscript. Differently from the majority of the existing approaches, the clear advantage of the proposed index is a sound physical meaning; accordingly, the index can be used for a more straightforward performance quantification, and to derive robot specifications.
Resumo:
Al giorno d'oggi, l'industry 4.0 è un movimento sempre più prominente che induce ad equipaggiare gli impianti industriali con avanzate infrastrutture tecnologiche digitali, le quali operano sinergicamente con l'impianto, al fine di controllare ed aumentare la produttività, monitorare e prevenire i futuri guasti, ed altro ancora. In questo ambito, gli utenti sono parte integrante della struttura produttiva, in cui ricoprono ruoli strategici e flessibili, collaborano fra loro e con le macchine, con l’obiettivo di affrontare e risolvere proattivamente una vasta gamma di problemi complessi. In particolare, la customer assistance nel settore industriale può certamente variare in relazione a molteplici elementi: il tipo di produzione e le caratteristiche del prodotto; l'organizzazione ed infrastruttura aziendale interna; la quantità di risorse disponibili che possono essere impiegate; il grado di importanza ricoperto dalla customer assistance nel settore industriale di riferimento; altri eventuali fattori appartenenti ad un dominio specifico. Per queste ragioni, si è cercato di individuare e categorizzare nel modo più accurato possibile, il lavoro svolto in questo elaborato ed il contesto nel quale è stato sviluppato. In questa tesi, viene descritta un'applicazione web per erogare assistenza al cliente in ambito di industria 4.0, attraverso il paradigma di ticketing o ticket di supporto/assistenza. Questa applicazione è integrata nel sistema Mentor, il quale è attivo già da anni nel settore industriale 4.0. Il progetto Mentor è una suite di applicazioni cloud-based creata dal gruppo Bucci Industries, una multinazionale attiva nell'industria e nell'automazione con sede a Faenza. In questo caso di studio, si presenta la progettazione ed implementazione della parte front-end del suddetto sistema di assistenza, il quale è integrato ed interconnesso con un paio di applicazioni tipiche di industria 4.0, presenti nella stessa suite di applicazioni.
Resumo:
The idea of Grid Computing originated in the nineties and found its concrete applications in contexts like the SETI@home project where a lot of computers (offered by volunteers) cooperated, performing distributed computations, inside the Grid environment analyzing radio signals trying to find extraterrestrial life. The Grid was composed of traditional personal computers but, with the emergence of the first mobile devices like Personal Digital Assistants (PDAs), researchers started theorizing the inclusion of mobile devices into Grid Computing; although impressive theoretical work was done, the idea was discarded due to the limitations (mainly technological) of mobile devices available at the time. Decades have passed, and now mobile devices are extremely more performant and numerous than before, leaving a great amount of resources available on mobile devices, such as smartphones and tablets, untapped. Here we propose a solution for performing distributed computations over a Grid Computing environment that utilizes both desktop and mobile devices, exploiting the resources from day-to-day mobile users that alternatively would end up unused. The work starts with an introduction on what Grid Computing is, the evolution of mobile devices, the idea of integrating such devices into the Grid and how to convince device owners to participate in the Grid. Then, the tone becomes more technical, starting with an explanation on how Grid Computing actually works, followed by the technical challenges of integrating mobile devices into the Grid. Next, the model, which constitutes the solution offered by this study, is explained, followed by a chapter regarding the realization of a prototype that proves the feasibility of distributed computations over a Grid composed by both mobile and desktop devices. To conclude future developments and ideas to improve this project are presented.
Resumo:
With each directed acyclic graph (this includes some D-dimensional lattices) one can associate some Abelian algebras that we call directed Abelian algebras (DAAs). On each site of the graph one attaches a generator of the algebra. These algebras depend on several parameters and are semisimple. Using any DAA, one can define a family of Hamiltonians which give the continuous time evolution of a stochastic process. The calculation of the spectra and ground-state wave functions (stationary state probability distributions) is an easy algebraic exercise. If one considers D-dimensional lattices and chooses Hamiltonians linear in the generators, in finite-size scaling the Hamiltonian spectrum is gapless with a critical dynamic exponent z=D. One possible application of the DAA is to sandpile models. In the paper we present this application, considering one- and two-dimensional lattices. In the one-dimensional case, when the DAA conserves the number of particles, the avalanches belong to the random walker universality class (critical exponent sigma(tau)=3/2). We study the local density of particles inside large avalanches, showing a depletion of particles at the source of the avalanche and an enrichment at its end. In two dimensions we did extensive Monte-Carlo simulations and found sigma(tau)=1.780 +/- 0.005.
Resumo:
The flowpaths by which water moves from watersheds to streams has important consequences for the runoff dynamics and biogeochemistry of surface waters in the Amazon Basin. The clearing of Amazon forest to cattle pasture has the potential to change runoff sources to streams by shifting runoff to more surficial flow pathways. We applied end-member mixing analysis (EMMA) to 10 small watersheds throughout the Amazon in which solute composition of streamwater and groundwater, overland flow, soil solution, throughfall and rainwater were measured, largely as part of the Large-Scale Biosphere-Atmosphere Experiment in Amazonia. We found a range in the extent to which streamwater samples fell within the mixing space determined by potential flowpath end-members, suggesting that some water sources to streams were not sampled. The contribution of overland flow as a source of stream flow was greater in pasture watersheds than in forest watersheds of comparable size. Increases in overland flow contribution to pasture streams ranged in some cases from 0% in forest to 27-28% in pasture and were broadly consistent with results from hydrometric sampling of Amazon forest and pasture watersheds that indicate 17- to 18-fold increase in the overland flow contribution to stream flow in pastures. In forest, overland flow was an important contribution to stream flow (45-57%) in ephemeral streams where flows were dominated by stormflow. Overland flow contribution to stream flow decreased in importance with increasing watershed area, from 21 to 57% in forest and 60-89% in pasture watersheds of less than 10 ha to 0% in forest and 27-28% in pastures in watersheds greater than 100 ha. Soil solution contributions to stream flow were similar across watershed area and groundwater inputs generally increased in proportion to decreases in overland flow. Application of EMMA across multiple watersheds indicated patterns across gradients of stream size and land cover that were consistent with patterns determined by detailed hydrometric sampling.
Resumo:
This paper concern the development of a stable model predictive controller (MPC) to be integrated with real time optimization (RTO) in the control structure of a process system with stable and integrating outputs. The real time process optimizer produces Optimal targets for the system inputs and for Outputs that Should be dynamically implemented by the MPC controller. This paper is based oil a previous work (Comput. Chem. Eng. 2005, 29, 1089) where a nominally stable MPC was proposed for systems with the conventional control approach where only the outputs have set points. This work is also based oil the work of Gonzalez et at. (J. Process Control 2009, 19, 110) where the zone control of stable systems is studied. The new control for is obtained by defining ail extended control objective that includes input targets and zone controller the outputs. Additional decision variables are also defined to increase the set of feasible solutions to the control problem. The hard constraints resulting from the cancellation of the integrating modes Lit the end of the control horizon are softened,, and the resulting control problem is made feasible to a large class of unknown disturbances and changes of the optimizing targets. The methods are illustrated with the simulated application of the proposed,approaches to a distillation column of the oil refining industry.
Resumo:
This work aims at proposing the use of the evolutionary computation methodology in order to jointly solve the multiuser channel estimation (MuChE) and detection problems at its maximum-likelihood, both related to the direct sequence code division multiple access (DS/CDMA). The effectiveness of the proposed heuristic approach is proven by comparing performance and complexity merit figures with that obtained by traditional methods found in literature. Simulation results considering genetic algorithm (GA) applied to multipath, DS/CDMA and MuChE and multi-user detection (MuD) show that the proposed genetic algorithm multi-user channel estimation (GAMuChE) yields a normalized mean square error estimation (nMSE) inferior to 11%, under slowly varying multipath fading channels, large range of Doppler frequencies and medium system load, it exhibits lower complexity when compared to both maximum likelihood multi-user channel estimation (MLMuChE) and gradient descent method (GrdDsc). A near-optimum multi-user detector (MuD) based on the genetic algorithm (GAMuD), also proposed in this work, provides a significant reduction in the computational complexity when compared to the optimum multi-user detector (OMuD). In addition, the complexity of the GAMuChE and GAMuD algorithms were (jointly) analyzed in terms of number of operations necessary to reach the convergence, and compared to other jointly MuChE and MuD strategies. The joint GAMuChE-GAMuD scheme can be regarded as a promising alternative for implementing third-generation (3G) and fourth-generation (4G) wireless systems in the near future. Copyright (C) 2010 John Wiley & Sons, Ltd.
Resumo:
Nutrition in bean plants and anthracnose intensity in function of silicon and copper application. The objective of this work was to evaluate the effect of calcium silicate and copper sulfate on anthracnose intensity and nutrition of bean plants. The experiment was conducted using an experimental design in randomized blocks following a 4 x 4 factorial arrangement , (four levels of calcium silicate and four levels of copper sulfate) and two additional treatments (plants without inoculation and plants sprinkled with Benomyl). Four evaluations of the incidence and severity of anthracnose were done, in addition to measuring, total leaf area. At the end of the evaluations, incidence: and data were integrated over time, obtaining the area under disease progress curve (AUDPC). Contents of N, P, K, Ca, Mg, B, Cu, Fe, Mn, Zn, Si and lignin were determined in the aerial Part. A linear decrease of the intensity AUDPC was observed with the increase of the doses of calcium silicate. The severity AUDPC was influenced by the doses of copper, obtaining a reduction of 35% on the higher dosage. The supply of silicon and copper altered the content of the K, mg, S, Zn, Ca and Si in the aerial part of the bean plants.
Resumo:
The University of Queensland, Australia has developed Fez, a world-leading user-interface and management system for Fedora-based institutional repositories, which bridges the gap between a repository and users. Christiaan Kortekaas, Andrew Bennett and Keith Webster will review this open source software that gives institutions the power to create a comprehensive repository solution without the hassle..
Resumo:
OctVCE is a cartesian cell CFD code produced especially for numerical simulations of shock and blast wave interactions with complex geometries, in particular, from explosions. Virtual Cell Embedding (VCE) was chosen as its cartesian cell kernel for its simplicity and sufficiency for practical engineering design problems. The code uses a finite-volume formulation of the unsteady Euler equations with a second order explicit Runge-Kutta Godonov (MUSCL) scheme. Gradients are calculated using a least-squares method with a minmod limiter. Flux solvers used are AUSM, AUSMDV and EFM. No fluid-structure coupling or chemical reactions are allowed, but gas models can be perfect gas and JWL or JWLB for the explosive products. This report also describes the code’s ‘octree’ mesh adaptive capability and point-inclusion query procedures for the VCE geometry engine. Finally, some space will also be devoted to describing code parallelization using the shared-memory OpenMP paradigm. The user manual to the code is to be found in the companion report 2007/13.
Resumo:
In this paper we present a model of specification-based testing of interactive systems. This model provides the basis for a framework to guide such testing. Interactive systems are traditionally decomposed into a functionality component and a user interface component; this distinction is termed dialogue separation and is the underlying basis for conceptual and architectural models of such systems. Correctness involves both proper behaviour of the user interface and proper computation by the underlying functionality. Specification-based testing is one method used to increase confidence in correctness, but it has had limited application to interactive system development to date.
Resumo:
The main problem with current approaches to quantum computing is the difficulty of establishing and maintaining entanglement. A Topological Quantum Computer (TQC) aims to overcome this by using different physical processes that are topological in nature and which are less susceptible to disturbance by the environment. In a (2+1)-dimensional system, pseudoparticles called anyons have statistics that fall somewhere between bosons and fermions. The exchange of two anyons, an effect called braiding from knot theory, can occur in two different ways. The quantum states corresponding to the two elementary braids constitute a two-state system allowing the definition of a computational basis. Quantum gates can be built up from patterns of braids and for quantum computing it is essential that the operator describing the braiding-the R-matrix-be described by a unitary operator. The physics of anyonic systems is governed by quantum groups, in particular the quasi-triangular Hopf algebras obtained from finite groups by the application of the Drinfeld quantum double construction. Their representation theory has been described in detail by Gould and Tsohantjis, and in this review article we relate the work of Gould to TQC schemes, particularly that of Kauffman.
Resumo:
Lentil is a self-pollinating diploid (2n = 14 chromosomes) annual cool season legume crop that is produced throughout the world and is highly valued as a high protein food. Several abiotic stresses are important to lentil yields world wide and include drought, heat, salt susceptibility and iron deficiency. The biotic stresses are numerous and include: susceptibility to Ascochyta blight, caused by Ascochyta lentis; Anthracnose, caused by Colletotrichum truncatum; Fusarium wilt, caused by Fusarium oxysporum; Sclerotinia white mold, caused by Sclerotinia sclerotiorum; rust, caused by Uromyces fabae; and numerous aphid transmitted viruses. Lentil is also highly susceptible to several species of Orabanche prevalent in the Mediterranean region, for which there does not appear to be much resistance in the germplasm. Plant breeders and geneticists have addressed these stresses by identifying resistant/tolerant germplasm, determining the genetics involved and the genetic map positions of the resistant genes. To this end progress has been made in mapping the lentil genome and several genetic maps are available that eventually will lead to the development of a consensus map for lentil. Marker density has been limited in the published genetic maps and there is a distinct lack of co-dominant markers that would facilitate comparisons of the available genetic maps and efficient identification of markers closely linked to genes of interest. Molecular breeding of lentil for disease resistance genes using marker assisted selection, particularly for resistance to Ascochyta blight and Anthracnose, is underway in Australia and Canada and promising results have been obtained. Comparative genomics and synteny analyses with closely related legumes promises to further advance the knowledge of the lentil genome and provide lentil breeders with additional genes and selectable markers for use in marker assisted selection. Genomic tools such as macro and micro arrays, reverse genetics and genetic transformation are emerging technologies that may eventually be available for use in lentil crop improvement.