13 resultados para Safety data recording and reporting

em Universidad Politécnica de Madrid


Relevância:

100.00% 100.00%

Publicador:

Resumo:

A sensitivity analysis has been performed to assess the influence of the elastic properties of railway vehicle suspensions on the vehicle dynamic behaviour. To do this, 144 dynamic simulations were performed modifying, one at a time, the stiffness and damping coefficients, of the primary and secondary suspensions. Three values were assigned to each parameter, corresponding to the percentiles 10, 50 and 90 of a data set stored in a database of railway vehicles.After processing the results of these simulations, the analyzed parameters were sorted by increasing influence. It was also found which of these parameters could be estimated with a lesser degree of accuracy in future simulations without appreciably affecting the simulation results. In general terms, it was concluded that the highest influences were found for the longitudinal stiffness and the lateral stiffness of the primary suspension, and the lowest influences for the vertical stiffness and the vertical damping of the primary suspension, with the parameters of the secondary suspension showing intermediate influences between them.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

We present the data structures and algorithms used in the approach for building domain ontologies from folksonomies and linked data. In this approach we extracts domain terms from folksonomies and enrich them with semantic information from the Linked Open Data cloud. As a result, we obtain a domain ontology that combines the emergent knowledge of social tagging systems with formal knowledge from Ontologies.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

PART I:Cross-section uncertainties under differentneutron spectra. PART II: Processing uncertainty libraries

Relevância:

100.00% 100.00%

Publicador:

Resumo:

El objetivo de este proyecto es diseñar un sistema capaz de controlar la velocidad de rotación de un motor DC en función del valor de temperatura obtenido de un sensor. Para ello se generará con un microcontrolador una señal PWM, cuyo ciclo de trabajo estará en función de la temperatura medida. En lo que respecta a la fase de diseño, hay dos partes claramente diferenciadas, relativas al hardware y al software. En cuanto al diseño del hardware puede hacerse a su vez una división en dos partes. En primer lugar, hubo que diseñar la circuitería necesaria para adaptar los niveles de tensión entregados por el sensor de temperatura a los niveles requeridos por ADC, requerido para digitalizar la información para su posterior procesamiento por parte del microcontrolador. Por tanto hubo que diseñar capaz de corregir el offset y la pendiente de la función tensión-temperatura del sensor, a fin de adaptarlo al rango de tensión requerido por el ADC. Por otro lado, hubo que diseñar el circuito encargado de controlar la velocidad de rotación del motor. Este circuito estará basado en un transistor MOSFET en conmutación, controlado mediante una señal PWM como se mencionó anteriormente. De esta manera, al variar el ciclo de trabajo de la señal PWM, variará de manera proporcional la tensión que cae en el motor, y por tanto su velocidad de rotación. En cuanto al diseño del software, se programó el microcontrolador para que generase una señal PWM en uno de sus pines en función del valor entregado por el ADC, a cuya entrada está conectada la tensión obtenida del circuito creado para adaptar la tensión generada por el sensor. Así mismo, se utiliza el microcontrolador para representar el valor de temperatura obtenido en una pantalla LCD. Para este proyecto se eligió una placa de desarrollo mbed, que incluye el microcontrolador integrado, debido a que facilita la tarea del prototipado. Posteriormente se procedió a la integración de ambas partes, y testeado del sistema para comprobar su correcto funcionamiento. Puesto que el resultado depende de la temperatura medida, fue necesario simular variaciones en ésta, para así comprobar los resultados obtenidos a distintas temperaturas. Para este propósito se empleó una bomba de aire caliente. Una vez comprobado el funcionamiento, como último paso se diseñó la placa de circuito impreso. Como conclusión, se consiguió desarrollar un sistema con un nivel de exactitud y precisión aceptable, en base a las limitaciones del sistema. SUMMARY: It is obvious that day by day people’s daily life depends more on technology and science. Tasks tend to be done automatically, making them simpler and as a result, user life is more comfortable. Every single task that can be controlled has an electronic system behind. In this project, a control system based on a microcontroller was designed for a fan, allowing it to go faster when temperature rises or slowing down as the environment gets colder. For this purpose, a microcontroller was programmed to generate a signal, to control the rotation speed of the fan depending on the data acquired from a temperature sensor. After testing the whole design developed in the laboratory, the next step taken was to build a prototype, which allows future improvements in the system that are discussed in the corresponding section of the thesis.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The uptake of Linked Data (LD) has promoted the proliferation of datasets and their associated ontologies for describing different domains. Par-ticular LD development characteristics such as agility and web-based architec-ture necessitate the revision, adaption, and lightening of existing methodologies for ontology development. This thesis proposes a lightweight method for ontol-ogy development in an LD context which will be based in data-driven agile de-velopments, existing resources to be reused, and the evaluation of the obtained products considering both classical ontological engineering principles and LD characteristics.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Abstract is not available.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

In this position paper, we claim that the need for time consuming data preparation and result interpretation tasks in knowledge discovery, as well as for costly expert consultation and consensus building activities required for ontology building can be reduced through exploiting the interplay of data mining and ontology engineering. The aim is to obtain in a semi-automatic way new knowledge from distributed data sources that can be used for inference and reasoning, as well as to guide the extraction of further knowledge from these data sources. The proposed approach is based on the creation of a novel knowledge discovery method relying on the combination, through an iterative ?feedbackloop?, of (a) data mining techniques to make emerge implicit models from data and (b) pattern-based ontology engineering to capture these models in reusable, conceptual and inferable artefacts.

Relevância:

100.00% 100.00%

Publicador:

Relevância:

100.00% 100.00%

Publicador:

Resumo:

While a number of virtual data-gloves have been used in stroke, there is little evidence about their use in spinal cord injury (SCI). A pilot clinical experience with nine SCI subjects was performed comparing two groups: one carried out a virtual rehabilitation training based on the use of a data glove, CyberTouch combined with traditional rehabilitation, during 30 minutes a day twice a week along two weeks; while the other made only conventional rehabilitation. Furthermore, two functional indexes were developed in order to assess the patient’s performance of the sessions: normalized trajectory lengths and repeatability. While differences between groups were not statistically significant, the data-glove group seemed to obtain better results in the muscle balance and functional parameters, and in the dexterity, coordination and fine grip tests. Related to the indexes that we implemented, normalized trajectory lengths and repeatability, every patient showed an improvement in at least one of the indexes, either along Y-axis trajectory or Z-axis trajectory. This study might be a step in investigating new ways of treatments and objective measures in order to obtain more accurate data about the patient’s evolution, allowing the clinicians to develop rehabilitation treatments, adapted to the abilities and needs of the patients.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Linked Data is the key paradigm of the Semantic Web, a new generation of the World Wide Web that promises to bring meaning (semantics) to data. A large number of both public and private organizations have published their data following the Linked Data principles, or have done so with data from other organizations. To this extent, since the generation and publication of Linked Data are intensive engineering processes that require high attention in order to achieve high quality, and since experience has shown that existing general guidelines are not always sufficient to be applied to every domain, this paper presents a set of guidelines for generating and publishing Linked Data in the context of energy consumption in buildings (one aspect of Building Information Models). These guidelines offer a comprehensive description of the tasks to perform, including a list of steps, tools that help in achieving the task, various alternatives for performing the task, and best practices and recommendations. Furthermore, this paper presents a complete example on the generation and publication of Linked Data about energy consumption in buildings, following the presented guidelines, in which the energy consumption data of council sites (e.g., buildings and lights) belonging to the Leeds City Council jurisdiction have been generated and published as Linked Data.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Real time Tritium concentrations in air in two chemical forms, HT and HTO, coming from an ITER-like fusion reactor as source were coupled the European Centre Medium Range Weather Forecast (ECMWF) numerical model with the Lagrangian Atmospheric-particle dispersion model FLEXPART. This tool was analyzed in nominal tritium discharge operational reference and selected incidental conditions affecting the Western Mediterranean Basin during 45 days during summer 2010 together with surface “wind observations” or weather data based in real hourly observations of wind direction and velocity providing a real approximation of the tritium behavior after the release to the atmosphere from a fusion reactor. From comparison with NORMTRI - a code using climatologically sequences as input - over the same area, the real time results have demonstrated an apparent overestimation of the corresponding climatologically sequence of Tritium concentrations in air outputs, at several distances from the reactor. For this purpose two development patterns were established. The first one was following a cyclonic circulation over the Mediterranean Sea and the second one was based on the plume delivered over the Interior of the Iberian Peninsula and Continental Europe by another stabilized circulation corresponding to a High Pressure System. One of the important remaining activities defined then, was the qualification tool. In order to validate the model of ECMWF/FLEXPART we have developed of a new complete data base of tritium concentrations for the months from November 2010 to March 2011 and defined a new set of four patterns of HT transport in air, in each case using real boundary conditions: stationary to the North, stationary to the South, fast and very fast displacement. Finally the differences corresponding to those four early patterns (each one in assessments 1 and 2) has been analyzed in terms of the tuning of safety related issues and taking into account the primary phase o- - f tritium modeling, from its discharge to the atmosphere to the deposition on the ground, will affect to the complete tritium environmental pathway altering the chronic dose by absorption, reemission and ingestion both from elemental tritium, HT and from the oxide of tritium, HTO

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The development of new-generation intelligent vehicle technologies will lead to a better level of road safety and CO2 emission reductions. However, the weak point of all these systems is their need for comprehensive and reliable data. For traffic data acquisition, two sources are currently available: 1) infrastructure sensors and 2) floating vehicles. The former consists of a set of fixed point detectors installed in the roads, and the latter consists of the use of mobile probe vehicles as mobile sensors. However, both systems still have some deficiencies. The infrastructure sensors retrieve information fromstatic points of the road, which are spaced, in some cases, kilometers apart. This means that the picture of the actual traffic situation is not a real one. This deficiency is corrected by floating cars, which retrieve dynamic information on the traffic situation. Unfortunately, the number of floating data vehicles currently available is too small and insufficient to give a complete picture of the road traffic. In this paper, we present a floating car data (FCD) augmentation system that combines information fromfloating data vehicles and infrastructure sensors, and that, by using neural networks, is capable of incrementing the amount of FCD with virtual information. This system has been implemented and tested on actual roads, and the results show little difference between the data supplied by the floating vehicles and the virtual vehicles.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Empirical Software Engineering (ESE) replication researchers need to store and manipulate experimental data for several purposes, in particular analysis and reporting. Current research needs call for sharing and preservation of experimental data as well. In a previous work, we analyzed Replication Data Management (RDM) needs. A novel concept, called Experimental Ecosystem, was proposed to solve current deficiencies in RDMapproaches. The empirical ecosystem provides replication researchers with a common framework that integrates transparently local heterogeneous data sources. A typical situation where the Empirical Ecosystem is applicable, is when several members of a research group, or several research groups collaborating together, need to share and access each other experimental results. However, to be able to apply the Empirical Ecosystem concept and deliver all promised benefits, it is necessary to analyze the software architectures and tools that can properly support it.