860 resultados para Vallejo
Resumo:
Mersenne Twister (MT) uniform random number generators are key cores for hardware acceleration of Monte Carlo simulations. In this work, two different architectures are studied: besides the classical table-based architecture, a different architecture based on a circular buffer and especially targeting FPGAs is proposed. A 30% performance improvement has been obtained when compared to the fastest previous work. The applicability of the proposed MT architectures has been proven in a high performance Gaussian RNG.
Resumo:
The area of Human-Machine Interface is growing fast due to its high importance in all technological systems. The basic idea behind designing human-machine interfaces is to enrich the communication with the technology in a natural and easy way. Gesture interfaces are a good example of transparent interfaces. Such interfaces must identify properly the action the user wants to perform, so the proper gesture recognition is of the highest importance. However, most of the systems based on gesture recognition use complex methods requiring high-resource devices. In this work, we propose to model gestures capturing their temporal properties, which significantly reduce storage requirements, and use clustering techniques, namely self-organizing maps and unsupervised genetic algorithm, for their classification. We further propose to train a certain number of algorithms with different parameters and combine their decision using majority voting in order to decrease the false positive rate. The main advantage of the approach is its simplicity, which enables the implementation using devices with limited resources, and therefore low cost. The testing results demonstrate its high potential.
Resumo:
As wireless sensor networks are usually deployed in unattended areas, security policies cannot be updated in a timely fashion upon identification of new attacks. This gives enough time for attackers to cause significant damage. Thus, it is of great importance to provide protection from unknown attacks. However, existing solutions are mostly concentrated on known attacks. On the other hand, mobility can make the sensor network more resilient to failures, reactive to events, and able to support disparate missions with a common set of sensors, yet the problem of security becomes more complicated. In order to address the issue of security in networks with mobile nodes, we propose a machine learning solution for anomaly detection along with the feature extraction process that tries to detect temporal and spatial inconsistencies in the sequences of sensed values and the routing paths used to forward these values to the base station. We also propose a special way to treat mobile nodes, which is the main novelty of this work. The data produced in the presence of an attacker are treated as outliers, and detected using clustering techniques. These techniques are further coupled with a reputation system, in this way isolating compromised nodes in timely fashion. The proposal exhibits good performances at detecting and confining previously unseen attacks, including the cases when mobile nodes are compromised.
Resumo:
Teniendo como punto de partida la Convención de la Diversidad biológica de 1992, desde la segunda mitad del siglo XX las perspectivas en política y gestión medioambiental han dado un giro sin precedentes. Variables intangibles, difíciles de medir, como es la biodiversidad, son cada vez más importantes para la sociedad. En consecuencia, enmarcado en políticas y requerimientos internacionales, desde el Tercer Inventario Forestal Nacional (IFN-3) se está aplicando una nueva metodología en continuo desarrollo que trata de estimar nuevas variables relacionadas con la biodiversidad. El objetivo más importante es poder analizar mediante estos nuevos indicadores la evolución del estado de la biodiversidad forestal de nuestros bosques a lo largo de los sucesivos IFN. Para llevar a cabo este reto, la toma de datos de campo en el inventario se centra principalmente en indicadores relativos a la composición florística y a la estructura de las masas forestales. En este artículo se describe la evolución y el desarrollo de la metodología llevada a cabo hasta el momento para la estimación de la biodiversidad forestal en el IFN. Además, se detallan los resultados derivados de su aplicación en las formaciones forestales más representativas de Galicia. Finalmente, se exponen los nuevos indicadores relacionados con la biodiversidad forestal en los que se está investigando, así como nuevas perspectivas de análisis.
Resumo:
Several methods to improve multiple distant microphone (MDM) speaker diarization based on Time Delay of Arrival (TDOA) features are evaluated in this paper. All of them avoid the use of a single reference channel to calculate the TDOA values and, based on different criteria, select among all possible pairs of microphones a set of pairs that will be used to estimate the TDOA's. The evaluated methods have been named the "Dynamic Margin" (DM), the "Extreme Regions" (ER), the "Most Common" (MC), the "Cross Correlation" (XCorr) and the "Principle Component Analysis" (PCA). It is shown that all methods improve the baseline results for the development set and four of them improve also the results for the evaluation set. Improvements of 3.49% and 10.77% DER relative are obtained for DM and ER respectively for the test set. The XCorr and PCA methods achieve an improvement of 36.72% and 30.82% DER relative for the test set. Moreover, the computational cost for the XCorr method is 20% less than the baseline.
Resumo:
Variabilities associated with CMOS evolution affect the yield and performance of current digital designs. FPGAs, which are widely used for fast prototyping and implementation of digital circuits, also suffer from these issues. Proactive approaches start to appear to achieve self-awareness and dynamic adaptation of these devices. To support these techniques we propose the employment of a multi-purpose sensor network. This infrastructure, through adequate use of configuration and automation tools, is able to obtain relevant data along the life cycle of an FPGA. This is realised at a very reduced cost, not only in terms of area or other limited resources, but also regarding the design effort required to define and deploy the measuring infrastructure. Our proposal has been validated by measuring inter-die and intra-die variability in different FPGA families.
Resumo:
In the recent years the missing fourth component, the memristor, was successfully synthesized. However, the mathematical complexity and variety of the models behind this component, in addition to the existence of convergence problems in the simulations, make the design of memristor-based applications long and difficult. In this work we present a memristor model characterization framework which supports the automated generation of subcircuit files. The proposed environment allows the designer to choose and parameterize the memristor model that best suits for a given application. The framework carries out characterizing simulations in order to study the possible non-convergence problems, solving the dependence on the simulation conditions and guaranteeing the functionality and performance of the design. Additionally, the occurrence of undesirable effects related to PVT variations is also taken into account. By performing a Monte Carlo or a corner analysis, the designer is aware of the safety margins which assure the correct device operation.
Resumo:
Dynamic thermal management techniques require a collection of on-chip thermal sensors that imply a significant area and power overhead. Finding the optimum number of temperature monitors and their location on the chip surface to optimize accuracy is an NP-hard problem. In this work we improve the modeling of the problem by including area, power and networking constraints along with the consideration of three inaccuracy terms: spatial errors, sampling rate errors and monitor-inherent errors. The problem is solved by the simulated annealing algorithm. We apply the algorithm to a test case employing three different types of monitors to highlight the importance of the different metrics. Finally we present a case study of the Alpha 21364 processor under two different constraint scenarios.
Resumo:
The dHDL language has been defined to improve hardware design productivity. This is achieved through the definition of a better reuse interface (including parameters, attributes and macroports) and the creation of control structures that help the designer in the hardware generation process.
Resumo:
This work proposes an encapsulation scheme aimed at simplifying the reuse process of hardware cores. This hardware encapsulation approach has been conceived with a twofold objective. First, we look for the improvement of the reuse interface associated with the hardware core description. This is carried out in a first encapsulation level by improving the limited types and configuration options available in the conventional HDLs interface, and also providing information related to the implementation itself. Second, we have devised a more generic interface focused on describing the function avoiding details from a particular implementation, what corresponds to a second encapsulation level. This encapsulation allows the designer to define how to configure and use the design to implement a given functionality. The proposed encapsulation schemes help improving the amount of information that can be supplied with the design, and also allow to automate the process of searching, configuring and implementing diverse alternatives.
Resumo:
The tsunami deposits of the valley of Agaete (Pérez-Torrado et al., 2006), north-western Gran Canaria, attributed to the Guimar flank collapse in Tenerife, have been revisited and new data are presented here. Besides the occurrences reported by Pérez-Torrado et al. (2006) a new outcrop was found and named “La Ruina” (at 28º 05’ 47,41” N; 15º 41’ 52,04” W; 71 m asl). The above-mentioned authors suggested the possibility that more than one marine conglomerate deposit could be present in the outcrops of “Llanos de Turmán” and “Berrazales”. At “La Gasolinera” and “La Aldea 1” the conglomerates are formed by a single layer representing one depositional event; at “La Aldea 2”, the conglomerates are composed of two layers directly contacting with each other, but evidence of a time hiatus between them was not found. Although the hypothesis of stacking of two depositional units within the same episode versus deposition of two distinct layers in different time-moments is debatable at the present state of knowledge, the first possibility is favoured. The field evidence at “Llanos de Turman” and “Berrazales” unquestionably shows that terrestrial sediments (colluvia; paleosols) are present and separate two marine conglomerate deposits, indicating that at least two distinct tsunami inundations are needed to explain the stratigraphy. However, at the new “La Ruina” outcrop, besides the two deposits mentioned above, a third and older marine conglomerate was found, clearly separated in time from the ones cited above. The existence of marine conglomerates emplaced in different moments is evidenced by the occurrence of intercalated paleosols, colluvia and other subaerial materials, implying significant time intervals between the emplacement of marine conglomeratic layers. A number of gastropod operculae from the tsunamiites were sent for U-Th dating to try to further constrain the age span of these deposits. The field evidence presented above shows that the emplacement of the deposits is related to, at least, three tsunami events. The lateral correlation between different outcrops is difficult due to variable number of deposits in each outcrop, lateral discontinuity and variability, and to compositional and textural similarity between distinct tsunami sediments. The occurrence of three Pleistocene tsunami deposits in the same area points to a relatively high frequency of tsunamis (generated by landslides, surface rupturing earthquakes, fast entry of voluminous volcanic deposits into the sea or large submarine eruptions). It is possible that this recurrence of tsunami inundations may reflect multiple-phased landslides responsible for the mega-landslide scars prominent in the geomorphology of the neighbouring island of Tenerife. This is a contribution from project “Estabilidad de los edificios volcánicos en Canarias: análisis de los factores geológicos, geomecánicos y paleoclimáticos. Aplicación a los flancos N y S de la isla de Tenerife” financed by MCT, Spain.
Resumo:
The aim of this work was twofold: on the one hand, to describe a comparative study of two intelligent control techniques-fuzzy and intelligent proportional-integral (PI) control, and on the other, to try to provide an answer to an as yet unsolved topic in the automotive sector-stop-and-go control in urban environments at very low speeds. Commercial vehicles exhibit nonlinear behavior and therefore constitute an excellent platform on which to check the controllers. This paper describes the design, tuning, and evaluation of the controllers performing actions on the longitudinal control of a car-the throttle and brake pedals-to accomplish stop-and-go manoeuvres. They are tested in two steps. First, a simulation model is used to design and tune the controllers, and second, these controllers are implemented in the commercial vehicle-which has automatic driving capabilities-to check their behavior. A stop-and-go manoeuvre is implemented with the two control techniques using two cooperating vehicles.
Resumo:
There is clear evidence that investment in intelligent transportation system technologies brings major social and economic benefits. Technological advances in the area of automatic systems in particular are becoming vital for the reduction of road deaths. We here describe our approach to automation of one the riskiest autonomous manœuvres involving vehicles – overtaking. The approach is based on a stereo vision system responsible for detecting any preceding vehicle and triggering the autonomous overtaking manœuvre. To this end, a fuzzy-logic based controller was developed to emulate how humans overtake. Its input is information from the vision system and from a positioning-based system consisting of a differential global positioning system (DGPS) and an inertial measurement unit (IMU). Its output is the generation of action on the vehicle’s actuators, i.e., the steering wheel and throttle and brake pedals. The system has been incorporated into a commercial Citroën car and tested on the private driving circuit at the facilities of our research center, CAR, with different preceding vehicles – a motorbike, car, and truck – with encouraging results.
Resumo:
Four longitudinal control techniques are compared: a classical Proportional-Integral (PI) control; an advanced technique-called the i-PI-that adds an intelligent component to the PI; a fuzzy controller based on human experience; and an adaptive-network-based fuzzy inference system. The controllers were designed to tackle one of the challenging topics as yet unsolved by the automotive sector: managing autonomously a gasoline-propelled vehicle at very low speeds. The dynamics involved are highly nonlinear and constitute an excellent test-bed for newly designed controllers. A Citroën C3 Pluriel car was modified to permit autonomous action on the accelerator and the brake pedals-i.e., longitudinal control. The controllers were tested in two stages. First, the vehicle was modeled to check the controllers' feasibility. Second, the controllers were then implemented in the Citroën, and their behavior under the same conditions on an identical real circuit was compared.
Resumo:
ntelligent systems designed to reduce highway fatalities have been widely applied in the automotive sector in the last decade. Of all users of transport systems, pedestrians are the most vulnerable in crashes as they are unprotected. This paper deals with an autonomous intelligent emergency system designed to avoid collisions with pedestrians. The system consists of a fuzzy controller based on the time-to-collision estimate – obtained via a vision-based system – and the wheel-locking probability – obtained via the vehicle’s CAN bus – that generates a safe braking action. The system has been tested in a real car – a convertible Citroën C3 Pluriel – equipped with an automated electro-hydraulic braking system capable of working in parallel with the vehicle’s original braking circuit. The system is used as a last resort in the case that an unexpected pedestrian is in the lane and all the warnings have failed to produce a response from the driver.