22 resultados para setup time reduction

em Universidad Politécnica de Madrid


Relevância:

90.00% 90.00%

Publicador:

Resumo:

The Boundary Element Method (BEM) is a discretisation technique for solving partial differential equations, which offers, for certain problems, important advantages over domain techniques. Despite the high CPU time reduction that can be achieved, some 3D problems remain today untreatable because the extremely large number of degrees of freedom—dof—involved in the boundary description. Model reduction seems to be an appealing choice for both, accurate and efficient numerical simulations. However, in the BEM the reduction in the number of degrees of freedom does not imply a significant reduction in the CPU time, because in this technique the more important part of the computing time is spent in the construction of the discrete system of equations. In this way, a reduction also in the number of weighting functions, seems to be a key point to render efficient boundary element simulations.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

In this paper we will see how the efficiency of the MBS simulations can be improved in two different ways, by considering both an explicit and implicit semi-recursive formulation. The explicit method is based on a double velocity transformation that involves the solution of a redundant but compatible system of equations. The high computational cost of this operation has been drastically reduced by taking into account the sparsity pattern of the system. Regarding this, the goal of this method is the introduction of MA48, a high performance mathematical library provided by Harwell Subroutine Library. The second method proposed in this paper has the particularity that, depending on the case, between 70 and 85% of the computation time is devoted to the evaluation of forces derivatives with respect to the relative position and velocity vectors. Keeping in mind that evaluating these derivatives can be decomposed into concurrent tasks, the main goal of this paper lies on a successful and straightforward parallel implementation that have led to a substantial improvement with a speedup of 3.2 by keeping all the cores busy in a quad-core processor and distributing the workload between them, achieving on this way a huge time reduction by doing an ideal CPU usage

Relevância:

80.00% 80.00%

Publicador:

Resumo:

In recent years, the increasing sophistication of embedded multimedia systems and wireless communication technologies has promoted a widespread utilization of video streaming applications. It has been reported in 2013 that youngsters, aged between 13 and 24, spend around 16.7 hours a week watching online video through social media, business websites, and video streaming sites. Video applications have already been blended into people daily life. Traditionally, video streaming research has focused on performance improvement, namely throughput increase and response time reduction. However, most mobile devices are battery-powered, a technology that grows at a much slower pace than either multimedia or hardware developments. Since battery developments cannot satisfy expanding power demand of mobile devices, research interests on video applications technology has attracted more attention to achieve energy-efficient designs. How to efficiently use the limited battery energy budget becomes a major research challenge. In addition, next generation video standards impel to diversification and personalization. Therefore, it is desirable to have mechanisms to implement energy optimizations with greater flexibility and scalability. In this context, the main goal of this dissertation is to find an energy management and optimization mechanism to reduce the energy consumption of video decoders based on the idea of functional-oriented reconfiguration. System battery life is prolonged as the result of a trade-off between energy consumption and video quality. Functional-oriented reconfiguration takes advantage of the similarities among standards to build video decoders reconnecting existing functional units. If a feedback channel from the decoder to the encoder is available, the former can signal the latter changes in either the encoding parameters or the encoding algorithms for energy-saving adaption. The proposed energy optimization and management mechanism is carried out at the decoder end. This mechanism consists of an energy-aware manager, implemented as an additional block of the reconfiguration engine, an energy estimator, integrated into the decoder, and, if available, a feedback channel connected to the encoder end. The energy-aware manager checks the battery level, selects the new decoder description and signals to build a new decoder to the reconfiguration engine. It is worth noting that the analysis of the energy consumption is fundamental for the success of the energy management and optimization mechanism. In this thesis, an energy estimation method driven by platform event monitoring is proposed. In addition, an event filter is suggested to automate the selection of the most appropriate events that affect the energy consumption. At last, a detailed study on the influence of the training data on the model accuracy is presented. The modeling methodology of the energy estimator has been evaluated on different underlying platforms, single-core and multi-core, with different characteristics of workload. All the results show a good accuracy and low on-line computation overhead. The required modifications on the reconfiguration engine to implement the energy-aware manager have been assessed under different scenarios. The results indicate a possibility to lengthen the battery lifetime of the system in two different use-cases.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

The development of this work presents the implementation of an experimental platform, which will permit to investigate on a methodology for the design and analysis of a teleoperated system, considering the delay in the communication channel. The project has been developed in partnership with the laboratory of Automatic and Robotics of the Universidad Politécnica de Madrid and the Laboratory at the Centro de Tecnologías Avanzadas de Manufactura at the Pontificia Universidad Católica del Perú. The mechanical structure of the arm that is located in the remote side has been built and the electric servomechanism has been mounted to control their movement. The experimental test of the Teleoperation system has been developed. The PC104 card commands the power interface and sensors of the DC motor of each articulation of the arm. Has developed the drives for the management of the operations of the master and the slave: send/reception of position, speed, acceleration and current data through a CAN network. The programs for the interconnection through a LAN network, between the Windows Operating System and the Real-time Operating System (QNX), has been developed. The utility of the developed platform (hardware and software) has been demonstrated.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Contaminated soil reuse was investigated, with higher profusion, throughout the early 90’s, coinciding with the 1991 Gulf War, when efforts to amend large crude oil releases began in geotechnical assessment of contaminated soils. Isolated works referring to geotechnical testing with hydrocarbon ground contaminants are described in the state-of-the-art, which have been extended to other type of contaminated soil references. Contaminated soils by light non-aquous phase liquids (LNAPL) bearing capacity reduction has been previously investigated from a forensic point of view. To date, all the research works have been published based on the assumption of constant contaminant saturation for the entire soil mass. In contrast, the actual LNAPLs distribution plumes exhibit complex flow patterns which are subject to physical and chemical changes with time and distance travelled from the release source. This aspect has been considered along the present text. A typical Madrid arkosic soil formation is commonly known as Miga sand. Geotechnical tests have been carried out, with Miga sand specimens, in incremental series of LNAPL concentrations in order to observe the soil engineering properties variation due to a contamination increase. Results are discussed in relation with previous studies and as a matter of fact, soil mechanics parameters change in the presence of LNAPL, showing different tendencies according to each test and depending on the LNAPL content, as well as to the specimen’s initially planned relative density, dense or loose. Geotechnical practical implications are also commented on and analyzed. Variation on geotechnical properties may occur only within the external contour of contamination distribution plume. This scope has motivated the author to develop a physical model based on transparent soil technology. The model aims to reproduce the distribution of LNAPL into the ground due to an accidental release from a storage facility. Preliminary results indicate that the model is a potentially complementary tool for hydrogeological applications, site-characterization and remediation treatment testing within the framework of soil pollution events. A description of the test setup of an innovative three dimensional physical model for the flow of two or more phases, in porous media, is presented herein, along with a summary of the advantages, limitations and future applications for modeling with transparent material. En los primeros años de la década de los años 90, del siglo pasado, coincidiendo con la Guerra del Golfo en 1991, se investigó intensamente sobre la reutilización de suelos afectados por grandes volúmenes de vertidos de crudo, fomentándose la evaluación geotécnica de los suelos contaminados. Se describen, en el estado del arte de esta tésis, una serie de trabajos aislados en relación con la caracterización geotécnica de suelos contaminados con hidrocarburos, descripción ampliada mediante referencias relacionadas con otros tipos de contaminación de suelos. Existen estudios previos de patología de cimentaciones que analizan la reducción de la capacidad portante de suelos contaminados por hidrocarburos líquidos ligeros en fase no acuosa (acrónimo en inglés: LNAPL de “Liquid Non-Aquous Phase Liquid”). A fecha de redacción de la tesis, todas las publicaciones anteriores estaban basadas en la consideración de una saturación del contaminante constante en toda la extensión del terreno de cimentación. La distribución real de las plumas de contaminante muestra, por el contrario, complejas trayectorias de flujo que están sujetas a cambios físico-químicos en función del tiempo y la distancia recorrida desde su origen de vertido. Éste aspecto ha sido considerado y tratado en el presente texto. La arena de Miga es una formación geológica típica de Madrid. En el ámbito de esta tesis se han desarrollado ensayos geotécnicos con series de muestras de arena de Miga contaminadas con distintas concentraciones de LNAPL con el objeto de estimar la variación de sus propiedades geotécnicas debido a un incremento de contaminación. Se ha realizado una evaluación de resultados de los ensayos en comparación con otros estudios previamente analizados, resultando que las propiedades mecánicas del suelo, efectivamente, varían en función del contenido de LNAPL y de la densidad relativa con la que se prepare la muestra, densa o floja. Se analizan y comentan las implicaciones de carácter práctico que supone la mencionada variación de propiedades geotécnicas. El autor ha desarrollado un modelo físico basado en la tecnología de suelos transparentes, considerando que las variaciones de propiedades geotécnicas únicamente deben producirse en el ámbito interior del contorno de la pluma contaminante. El objeto del modelo es el de reproducir la distribución de un LNAPL en un terreno dado, causada por el vertido accidental de una instalación de almecenamiento de combustible. Los resultados preliminares indican que el modelo podría emplearse como una herramienta complementaria para el estudio de eventos contaminantes, permitiendo el desarrollo de aplicaciones de carácter hidrogeológico, caracterización de suelos contaminados y experimentación de tratamientos de remediación. Como aportación de carácter innovadora, se presenta y describe un modelo físico tridimensional de flujo de dos o más fases a través de un medio poroso transparente, analizándose sus ventajas e inconvenientes así como sus limitaciones y futuras aplicaciones.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This Doctoral Thesis entitled Contribution to the analysis, design and assessment of compact antenna test ranges at millimeter wavelengths aims to deepen the knowledge of a particular antenna measurement system: the compact range, operating in the frequency bands of millimeter wavelengths. The thesis has been developed at Radiation Group (GR), an antenna laboratory which belongs to the Signals, Systems and Radiocommunications department (SSR), from Technical University of Madrid (UPM). The Radiation Group owns an extensive experience on antenna measurements, running at present four facilities which operate in different configurations: Gregorian compact antenna test range, spherical near field, planar near field and semianechoic arch system. The research work performed in line with this thesis contributes the knowledge of the first measurement configuration at higher frequencies, beyond the microwaves region where Radiation Group features customer-level performance. To reach this high level purpose, a set of scientific tasks were sequentially carried out. Those are succinctly described in the subsequent paragraphs. A first step dealed with the State of Art review. The study of scientific literature dealed with the analysis of measurement practices in compact antenna test ranges in addition with the particularities of millimeter wavelength technologies. Joint study of both fields of knowledge converged, when this measurement facilities are of interest, in a series of technological challenges which become serious bottlenecks at different stages: analysis, design and assessment. Thirdly after the overview study, focus was set on Electromagnetic analysis algorithms. These formulations allow to approach certain electromagnetic features of interest, such as field distribution phase or stray signal analysis of particular structures when they interact with electromagnetic waves sources. Properly operated, a CATR facility features electromagnetic waves collimation optics which are large, in terms of wavelengths. Accordingly, the electromagnetic analysis tasks introduce an extense number of mathematic unknowns which grow with frequency, following different polynomic order laws depending on the used algorithmia. In particular, the optics configuration which was of our interest consisted on the reflection type serrated edge collimator. The analysis of these devices requires a flexible handling of almost arbitrary scattering geometries, becoming this flexibility the nucleus of the algorithmia’s ability to perform the subsequent design tasks. This thesis’ contribution to this field of knowledge consisted on reaching a formulation which was powerful at the same time when dealing with various analysis geometries and computationally speaking. Two algorithmia were developed. While based on the same principle of hybridization, they reached different order Physics performance at the cost of the computational efficiency. Inter-comparison of their CATR design capabilities was performed, reaching both qualitative as well as quantitative conclusions on their scope. In third place, interest was shifted from analysis - design tasks towards range assessment. Millimetre wavelengths imply strict mechanical tolerances and fine setup adjustment. In addition, the large number of unknowns issue already faced in the analysis stage appears as well in the on chamber field probing stage. Natural decrease of dynamic range available by semiconductor millimeter waves sources requires in addition larger integration times at each probing point. These peculiarities increase exponentially the difficulty of performing assessment processes in CATR facilities beyond microwaves. The bottleneck becomes so tight that it compromises the range characterization beyond a certain limit frequency which typically lies on the lowest segment of millimeter wavelength frequencies. However the value of range assessment moves, on the contrary, towards the highest segment. This thesis contributes this technological scenario developing quiet zone probing techniques which achieves substantial data reduction ratii. Collaterally, it increases the robustness of the results to noise, which is a virtual rise of the setup’s available dynamic range. In fourth place, the environmental sensitivity of millimeter wavelengths issue was approached. It is well known the drifts of electromagnetic experiments due to the dependance of the re sults with respect to the surrounding environment. This feature relegates many industrial practices of microwave frequencies to the experimental stage, at millimeter wavelengths. In particular, evolution of the atmosphere within acceptable conditioning bounds redounds in drift phenomena which completely mask the experimental results. The contribution of this thesis on this aspect consists on modeling electrically the indoor atmosphere existing in a CATR, as a function of environmental variables which affect the range’s performance. A simple model was developed, being able to handle high level phenomena, such as feed - probe phase drift as a function of low level magnitudes easy to be sampled: relative humidity and temperature. With this model, environmental compensation can be performed and chamber conditioning is automatically extended towards higher frequencies. Therefore, the purpose of this thesis is to go further into the knowledge of millimetre wavelengths involving compact antenna test ranges. This knowledge is dosified through the sequential stages of a CATR conception, form early low level electromagnetic analysis towards the assessment of an operative facility, stages for each one of which nowadays bottleneck phenomena exist and seriously compromise the antenna measurement practices at millimeter wavelengths.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Membrane systems are computational equivalent to Turing machines. However, their distributed and massively parallel nature obtains polynomial solutions opposite to traditional non-polynomial ones. At this point, it is very important to develop dedicated hardware and software implementations exploiting those two membrane systems features. Dealing with distributed implementations of P systems, the bottleneck communication problem has arisen. When the number of membranes grows up, the network gets congested. The purpose of distributed architectures is to reach a compromise between the massively parallel character of the system and the needed evolution step time to transit from one configuration of the system to the next one, solving the bottleneck communication problem. The goal of this paper is twofold. Firstly, to survey in a systematic and uniform way the main results regarding the way membranes can be placed on processors in order to get a software/hardware simulation of P-Systems in a distributed environment. Secondly, we improve some results about the membrane dissolution problem, prove that it is connected, and discuss the possibility of simulating this property in the distributed model. All this yields an improvement in the system parallelism implementation since it gets an increment of the parallelism of the external communication among processors. Proposed ideas improve previous architectures to tackle the communication bottleneck problem, such as reduction of the total time of an evolution step, increase of the number of membranes that could run on a processor and reduction of the number of processors.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Here, a novel and efficient strategy for moving object detection by non-parametric modeling on smart cameras is presented. Whereas the background is modeled using only color information, the foreground model combines color and spatial information. The application of a particle filter allows the update of the spatial information and provides a priori information about the areas to analyze in the following images, enabling an important reduction in the computational requirements and improving the segmentation results

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This paper presents a computer vision system that successfully discriminates between weed patches and crop rows under uncontrolled lighting in real-time. The system consists of two independent subsystems, a fast image processing delivering results in real-time (Fast Image Processing, FIP), and a slower and more accurate processing (Robust Crop Row Detection, RCRD) that is used to correct the first subsystem's mistakes. This combination produces a system that achieves very good results under a wide variety of conditions. Tested on several maize videos taken of different fields and during different years, the system successfully detects an average of 95% of weeds and 80% of crops under different illumination, soil humidity and weed/crop growth conditions. Moreover, the system has been shown to produce acceptable results even under very difficult conditions, such as in the presence of dramatic sowing errors or abrupt camera movements. The computer vision system has been developed for integration into a treatment system because the ideal setup for any weed sprayer system would include a tool that could provide information on the weeds and crops present at each point in real-time, while the tractor mounting the spraying bar is moving

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Abstraction-Carrying Code (ACC) is a framework for mobile code safety in which the code supplier provides a program together with an abstraction (or abstract model of the program) whose validity entails compliance with a predefined safety policy. The abstraction plays thus the role of safety certificate and its generation is carried out automatically by a fixed-point analyzer. The advantage of providing a (fixed-point) abstraction to the code consumer is that its validity is checked in a single pass (i.e., one iteration) of an abstract interpretation-based checker. A main challenge to make ACC useful in practice is to reduce the size of certificates as much as possible, while at the same time not increasing checking time. Intuitively, we only include in the certificate the information which the checker is unable to reproduce without iterating. We introduce the notion of reduced certifícate which characterizes the subset of the abstraction which a checker needs in order to validate (and re-construct) the full certificate in a single pass. Based on this notion, we show how to instrument a generic analysis algorithm with the necessary extensions in order to identify the information relevant to the checker.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Abstraction-Carrying Code (ACC) has recently been proposed as a framework for mobile code safety in which the code supplier provides a program together with an abstraction (or abstract model of the program) whose validity entails compliance with a predefined safety policy. The abstraction plays thus the role of safety certificate and its generation is carried out automatically by a fixpoint analyzer. The advantage of providing a (fixpoint) abstraction to the code consumer is that its validity is checked in a single pass (i.e., one iteration) of an abstract interpretation-based checker. A main challenge to make ACC useful in practice is to reduce the size of certificates as much as possible while at the same time not increasing checking time. The intuitive idea is to only include in the certificate information that the checker is unable to reproduce without iterating. We introduce the notion of reduced certificate which characterizes the subset of the abstraction which a checker needs in order to validate (and re-construct) the fall certificate in a single pass. Based on this notion, we instrument a generic analysis algorithm with the necessary extensions in order to identify the information relevant to the checker. Interestingly, the fact that the reduced certificate omits (parts of) the abstraction has implications in the design of the checker. We provide the sufficient conditions which allow us to ensure that 1) if the checker succeeds in validating the certificate, then the certificate is valid for the program (correctness) and 2) the checker will succeed for any reduced certificate which is valid (completeness). Our approach has been implemented and benchmarked within the CiaoPP system. The experimental results show t h a t our proposal is able to greatly reduce the size of certificates in practice. To appear in Theory and Practice of Logic Programming (TPLP).

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The concept of independence has been recently generalized to the constraint logic programming (CLP) paradigm. Also, several abstract domains specifically designed for CLP languages, and whose information can be used to detect the generalized independence conditions, have been recently defined. As a result we are now in a position where automatic parallelization of CLP programs is feasible. In this paper we study the task of automatically parallelizing CLP programs based on such analyses, by transforming them to explicitly concurrent programs in our parallel CC platform (CIAO) as well as to AKL. We describe the analysis and transformation process, and study its efficiency, accuracy, and effectiveness in program parallelization. The information gathered by the analyzers is evaluated not only in terms of its accuracy, i.e. its ability to determine the actual dependencies among the program variables, but also of its effectiveness, measured in terms of code reduction in the resulting parallelized programs. Given that only a few abstract domains have been already defined for CLP, and that none of them were specifically designed for dependency detection, the aim of the evaluation is not only to asses the effectiveness of the available domains, but also to study what additional information it would be desirable to infer, and what domains would be appropriate for further improving the parallelization process.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This research proposes a generic methodology for dimensionality reduction upon time-frequency representations applied to the classification of different types of biosignals. The methodology directly deals with the highly redundant and irrelevant data contained in these representations, combining a first stage of irrelevant data removal by variable selection, with a second stage of redundancy reduction using methods based on linear transformations. The study addresses two techniques that provided a similar performance: the first one is based on the selection of a set of the most relevant time?frequency points, whereas the second one selects the most relevant frequency bands. The first methodology needs a lower quantity of components, leading to a lower feature space; but the second improves the capture of the time-varying dynamics of the signal, and therefore provides a more stable performance. In order to evaluate the generalization capabilities of the methodology proposed it has been applied to two types of biosignals with different kinds of non-stationary behaviors: electroencephalographic and phonocardiographic biosignals. Even when these two databases contain samples with different degrees of complexity and a wide variety of characterizing patterns, the results demonstrate a good accuracy for the detection of pathologies, over 98%.The results open the possibility to extrapolate the methodology to the study of other biosignals.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The extreme runup is a key parameter for a shore risk analysis in which the accurate and quantitative estimation of the upper limit reached by waves is essential. Runup can be better approximated by splitting the setup and swash semi-amplitude contributions. In an experimental study recording setup becomes difficult due to infragravity motions within the surf zone, hence, it would be desirable to measure the setup with available methodologies and devices. In this research, an analysis is made of evaluated the convenience of direct estimation setup as the medium level in the swash zone for experimental runup analysis through a physical model. A physical mobile bed model was setup in a wave flume at the Laboratory for Maritime Experimentation of CEDEX. The wave flume is 36 metres long, 6.5 metres wide and 1.3 metres high. The physical model was designed to cover a reasonable range of parameters, three different slopes (1/50, 1/30 and 1/20), two sand grain sizes (D50 = 0.12 mm and 0.70 mm) and a range for the Iribarren number in deep water (ξ0) from 0.1 to 0.6. Best formulations were chosen for estimating a theoretical setup in the physical model application. Once theoretical setup had been obtained, a comparison was made with an estimation of the setup directly as a medium level of the oscillation in swash usually considered in extreme runup analyses. A good correlation was noted between both theoretical and time-averaging setup and a relation is proposed. Extreme runup is analysed through the sum of setup and semi-amplitude of swash. An equation is proposed that could be applied in strong foreshore slope-dependent reflective beaches.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Previous research studies and operational trials have shown that using the airborne Required Time of Arrival (RTA) function, an aircraft can individually achieve an assigned time to a metering or merge point accurately. This study goes a step further and investigates the application of RTA to a real sequence of arriving aircraft into Melbourne Australia. Assuming that the actual arrival times were Controlled Time of Arrivals (CTAs) assigned to each aircraft, the study examines if the airborne RTA solution would work. Three scenarios were compared: a baseline scenario being the actual flown trajectories in a two hour time-span into Melbourne, a scenario in which the sequential landing slot times of the baseline scenario were assigned as CTAs and a third scenario in which the landing slots could be freely redistributed to the inbound traffic as CTAs. The research found that pressure on the terminal area would sometimes require aircraft to lose more time than possible through the RTA capability. Using linear holding as an additional measure to absorb extensive delays, up to 500NM (5%) of total track reduction and 1300kg (3%) of total fuel consumption could be saved in the scenario with landing slots freely distributed as CTAs, compared to the baseline scenario. Assigning CTAs in an arrival sequence requires the ground system to have an accurate trajectory predictor to propose additional delay measures (path stretching, linear holding) if necessary. Reducing the achievable time window of the aircraft to add control margin to the RTA function, had a negative impact and increased the amount of intervention other than speed control required to solve the sequence. It was concluded that the RTA capability is not a complete solution but merely a tool to assist in managing the increasing complexity of air traffic.