920 resultados para Semi-complete Data Synchronization


Relevância:

30.00% 30.00%

Publicador:

Resumo:

In low and middle latitudes, the Cretaceous/Tertiary boundary is marked by a sudden and pronounced decrease in d13C values of near-surface-water carbonates and a reduction in the surface-to-bottom d13C gradient. These isotopic data have been interpreted as evidence of a decline in surface-water productivity that was responsible for the extinction of many planktic foraminiferal species and other marine organisms at or near the K/T boundary. We present planktic and benthic foraminiferal isotopic data from two almost biostratigraphically complete sections at Ocean Drilling Program Site 738 in the antarctic Indian Ocean and at Nye Kløv in Denmark. These data suggest that planktic carbonate d13C values in high latitudes may not have decreased dramatically at the K/T boundary; thus, surface-water productivity may not have been reduced as much as in low and middle latitudes. Comparison of the records of Site 738 with those of ODP Sites 690 and 750 indicates a pronounced decline in d13C values of planktic and benthic foraminifera and fine-fraction/bulk carbonate ~200 000 yr after the K/T boundary. This reflects a regional shift in the carbon isotopic composition of oceanic total dissolved carbon (TDC) and correlates with a similar change in benthic foraminiferal d13C values at mid- and low-latitude Deep Sea Drilling Project Sites 527 and 577. This oceanographic event was followed by the ecosystem's global recovery ~500 000 yr after the K/T boundary. These data suggest that the environmental effects of the K/T boundary may have been less severe in the high-latitude oceans than in tropical and subtropical regions.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Deep Sea Drilling Project (DSDP) Site 215 provides an expanded section across the Paleocene/Eocene boundary, the most complete mid-latitude sequence from a Southern Hemisphere location in the Indo-Pacific area. The events of this transition occurred during a span of about 1.2 m.y. Oxygen isotope values derived from benthic foraminiferal calcite decrease by about 1.0 per mil, a decrease most likely related to warming of deep ocean waters. Turnovers of benthic foraminifera accompany d18O changes and culminate in the predominant extinction event at the end of the Paleocene Epoch. Carbon isotope ratios also shift dramatically toward lighter values near the end of the Paleocene, beginning about 0.45 m.y. after oxygen isotope values start to change. The intensity of Southern Hemisphere atmospheric circulation as recorded by grain sizes of eolian particles shows a large and rapid reduction beginning another 0.45 m.y. later. A significant reduction of zonal wind strength at the Paleocene/Eocene boundary, until now observed only at Northern Hemisphere locations, appears to have been a global phenomenon related to decreased latitudinal thermal gradients occasioned by more effective poleward heat transport via the deep ocean.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The assessment of the accuracy of parameters related to the reactor core performance (e.g., ke) and f el cycle (e.g., isotopic evolution/transmutation) due to the uncertainties in the basic nuclear data (ND) is a critical issue. Different error propagation techniques (adjoint/forward sensitivity analysis procedures and/or Monte Carlo technique) can be used to address by computational simulation the systematic propagation of uncertainties on the final parameters. To perform this uncertainty assessment, the ENDF covariance les (variance/correlation in energy and cross- reactions-isotopes correlations) are required. In this paper, we assess the impact of ND uncertainties on the isotopic prediction for a conceptual design of a modular European Facility for Industrial Transmutation (EFIT) for a discharge burnup of 150 GWd/tHM. The complete set of uncertainty data for cross sections (EAF2007/UN, SCALE6.0/COVA-44G), radioactive decay and fission yield data (JEFF-3.1.1) are processed and used in ACAB code.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In a series of attempts to research and document relevant sloshing type phenomena, a series of experiments have been conducted. The aim of this paper is to describe the setup and data processing of such experiments. A sloshing tank is subjected to angular motion. As a result pressure registers are obtained at several locations, together with the motion data, torque and a collection of image and video information. The experimental rig and the data acquisition systems are described. Useful information for experimental sloshing research practitioners is provided. This information is related to the liquids used in the experiments, the dying techniques, tank building processes, synchronization of acquisition systems, etc. A new procedure for reconstructing experimental data, that takes into account experimental uncertainties, is presented. This procedure is based on a least squares spline approximation of the data. Based on a deterministic approach to the first sloshing wave impact event in a sloshing experiment, an uncertainty analysis procedure of the associated first pressure peak value is described.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

EURATOM/CIEMAT and Technical University of Madrid (UPM) have been involved in the development of a FPSC [1] (Fast Plant System Control) prototype for ITER, based on PXIe (PCI eXtensions for Instrumentation). One of the main focuses of this project has been data acquisition and all the related issues, including scientific data archiving. Additionally, a new data archiving solution has been developed to demonstrate the obtainable performances and possible bottlenecks of scientific data archiving in Fast Plant System Control. The presented system implements a fault tolerant architecture over a GEthernet network where FPSC data are reliably archived on remote, while remaining accessible to be redistributed, within the duration of a pulse. The storing service is supported by a clustering solution to guaranty scalability, so that FPSC management and configuration may be simplified, and a unique view of all archived data provided. All the involved components have been integrated under EPICS [2] (Experimental Physics and Industrial Control System), implementing in each case the necessary extensions, state machines and configuration process variables. The prototyped solution is based on the NetCDF-4 [3] and [4] (Network Common Data Format) file format in order to incorporate important features, such as scientific data models support, huge size files management, platform independent codification, or single-writer/multiple-readers concurrency. In this contribution, a complete description of the above mentioned solution is presented, together with the most relevant results of the tests performed, while focusing in the benefits and limitations of the applied technologies.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

It has been shown that it is possible to exploit Independent/Restricted And-parallelism in logic programs while retaining the conventional "don't know" semantics of such programs. In particular, it is possible to parallelize pure Prolog programs while maintaining the semantics of the language. However, when builtin side-effects (such as write or assert) appear in the program, if an identical observable behaviour to that of sequential Prolog implementations is to be preserved, such side-effects have to be properly sequenced. Previously proposed solutions to this problem are either incomplete (lacking, for example, backtracking semantics) or they force sequentialization of significant portions of the execution graph which could otherwise run in parallel. In this paper a series of side-effect synchronization methods are proposed which incur lower overhead and allow more parallelism than those previously proposed. Most importantly, and unlike previous proposals, they have well-defined backward execution behaviour and require only a small modification to a given (And-parallel) Prolog implementation.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The development of new-generation intelligent vehicle technologies will lead to a better level of road safety and CO2 emission reductions. However, the weak point of all these systems is their need for comprehensive and reliable data. For traffic data acquisition, two sources are currently available: 1) infrastructure sensors and 2) floating vehicles. The former consists of a set of fixed point detectors installed in the roads, and the latter consists of the use of mobile probe vehicles as mobile sensors. However, both systems still have some deficiencies. The infrastructure sensors retrieve information fromstatic points of the road, which are spaced, in some cases, kilometers apart. This means that the picture of the actual traffic situation is not a real one. This deficiency is corrected by floating cars, which retrieve dynamic information on the traffic situation. Unfortunately, the number of floating data vehicles currently available is too small and insufficient to give a complete picture of the road traffic. In this paper, we present a floating car data (FCD) augmentation system that combines information fromfloating data vehicles and infrastructure sensors, and that, by using neural networks, is capable of incrementing the amount of FCD with virtual information. This system has been implemented and tested on actual roads, and the results show little difference between the data supplied by the floating vehicles and the virtual vehicles.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This paper presents a data-intensive architecture that demonstrates the ability to support applications from a wide range of application domains, and support the different types of users involved in defining, designing and executing data-intensive processing tasks. The prototype architecture is introduced, and the pivotal role of DISPEL as a canonical language is explained. The architecture promotes the exploration and exploitation of distributed and heterogeneous data and spans the complete knowledge discovery process, from data preparation, to analysis, to evaluation and reiteration. The architecture evaluation included large-scale applications from astronomy, cosmology, hydrology, functional genetics, imaging processing and seismology.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In this position paper, we claim that the need for time consuming data preparation and result interpretation tasks in knowledge discovery, as well as for costly expert consultation and consensus building activities required for ontology building can be reduced through exploiting the interplay of data mining and ontology engineering. The aim is to obtain in a semi-automatic way new knowledge from distributed data sources that can be used for inference and reasoning, as well as to guide the extraction of further knowledge from these data sources. The proposed approach is based on the creation of a novel knowledge discovery method relying on the combination, through an iterative ?feedbackloop?, of (a) data mining techniques to make emerge implicit models from data and (b) pattern-based ontology engineering to capture these models in reusable, conceptual and inferable artefacts.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

A notorious advantage of wireless transmission is a significant reduction and simplification in wiring and harness. There are a lot of applications of wireless systems, but in many occasions sensor nodes require a specific housing to protect the electronics from hush environmental conditions. Nowadays the information is scarce and nonspecific on the dynamic behaviour of WSN and RFID. Therefore the purpose of this study is to evaluate the dynamic behaviour of the sensors. A series of trials were designed and performed covering temperature steps between cold room (5 °C), room temperature (23 °C) and heated environment (35 °C). As sensor nodes: three Crossbow motes, a surface mounted Nlaza module (with sensor Sensirion located on the motherboard), an aerial mounted Nlaza where the Sensirion sensor stayed at the end of a cable), and four tags RFID Turbo Tag (T700 model with and without housing), and 702-B (with and without housing). To assess the dynamic behaviour a first order response approach is used and fitted with dedicated optimization tools programmed in Matlab that allow extracting the time response (?) and corresponding determination coefficient (r2) with regard to experimental data. The shorter response time (20.9 s) is found for the uncoated T 700 tag which encapsulated version provides a significantly higher response (107.2 s). The highest ? corresponds to the Crossbow modules (144.4 s), followed by the surface mounted Nlaza module (288.1 s), while the module with aerial mounted sensor gives a response certainly close above to the T700 without coating (42.8 s). As a conclusion, the dynamic response of temperature sensors within wireless and RFID nodes is dramatically influenced by the way they are housed (to protect them from the environment) as well as by the heat released by the node electronics itself; its characterization is basic to allow monitoring of high rate temperature changes and to certify the cold chain. Besides the time to rise and to recover is significantly different being mostly higher for the latter than for the former.