922 resultados para data storage concept


Relevância:

40.00% 40.00%

Publicador:

Resumo:

Distributed data aggregation is an important task, allowing the de- centralized determination of meaningful global properties, that can then be used to direct the execution of other applications. The resulting val- ues result from the distributed computation of functions like count, sum and average. Some application examples can found to determine the network size, total storage capacity, average load, majorities and many others. In the last decade, many di erent approaches have been pro- posed, with di erent trade-o s in terms of accuracy, reliability, message and time complexity. Due to the considerable amount and variety of ag- gregation algorithms, it can be di cult and time consuming to determine which techniques will be more appropriate to use in speci c settings, jus- tifying the existence of a survey to aid in this task. This work reviews the state of the art on distributed data aggregation algorithms, providing three main contributions. First, it formally de nes the concept of aggrega- tion, characterizing the di erent types of aggregation functions. Second, it succinctly describes the main aggregation techniques, organizing them in a taxonomy. Finally, it provides some guidelines toward the selection and use of the most relevant techniques, summarizing their principal characteristics.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

BACKGROUND: The aim of our study was to assess the feasibility of minimally invasive digestive anastomosis using a modular flexible magnetic anastomotic device made up of a set of two flexible chains of magnetic elements. The assembly possesses a non-deployed linear configuration which allows it to be introduced through a dedicated small-sized applicator into the bowel where it takes the deployed form. A centering suture allows the mating between the two parts to be controlled in order to include the viscerotomy between the two magnetic rings and the connected viscera. METHODS AND PROCEDURES: Eight pigs were involved in a 2-week survival experimental study. In five colorectal anastomoses, the proximal device was inserted by a percutaneous endoscopic technique, and the colon was divided below the magnet. The distal magnet was delivered transanally to connect with the proximal magnet. In three jejunojejunostomies, the first magnetic chain was injected in its linear configuration through a small enterotomy. Once delivered, the device self-assembled into a ring shape. A second magnet was injected more distally through the same port. The centering sutures were tied together extracorporeally and, using a knot pusher, magnets were connected. Ex vivo strain testing to determine the compression force delivered by the magnetic device, burst pressure of the anastomosis, and histology were performed. RESULTS: Mean operative time including endoscopy was 69.2 ± 21.9 min, and average time to full patency was 5 days for colorectal anastomosis. Operative times for jejunojejunostomies were 125, 80, and 35 min, respectively. The postoperative period was uneventful. Burst pressure of all anastomoses was ≥ 110 mmHg. Mean strain force to detach the devices was 6.1 ± 0.98 and 12.88 ± 1.34 N in colorectal and jejunojejunal connections, respectively. Pathology showed a mild-to-moderate inflammation score. CONCLUSIONS: The modular magnetic system showed enormous potential to create minimally invasive digestive anastomoses, and may represent an alternative to stapled anastomoses, being easy to deliver, effective, and low cost.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

In this work we analyze how patchy distributions of CO2 and brine within sand reservoirs may lead to significant attenuation and velocity dispersion effects, which in turn may have a profound impact on surface seismic data. The ultimate goal of this paper is to contribute to the understanding of these processes within the framework of the seismic monitoring of CO2 sequestration, a key strategy to mitigate global warming. We first carry out a Monte Carlo analysis to study the statistical behavior of attenuation and velocity dispersion of compressional waves traveling through rocks with properties similar to those at the Utsira Sand, Sleipner field, containing quasi-fractal patchy distributions of CO2 and brine. These results show that the mean patch size and CO2 saturation play key roles in the observed wave-induced fluid flow effects. The latter can be remarkably important when CO2 concentrations are low and mean patch sizes are relatively large. To analyze these effects on the corresponding surface seismic data, we perform numerical simulations of wave propagation considering reservoir models and CO2 accumulation patterns similar to the CO2 injection site in the Sleipner field. These numerical experiments suggest that wave-induced fluid flow effects may produce changes in the reservoir's seismic response, modifying significantly the main seismic attributes usually employed in the characterization of these environments. Consequently, the determination of the nature of the fluid distributions as well as the proper modeling of the seismic data constitute important aspects that should not be ignored in the seismic monitoring of CO2 sequestration problems.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

The Attorney General’s Consumer Protection Division receives hundreds of calls and consumer complaints every year. Follow these tips to avoid unexpected expense and disappointments. This record is about: The Drive to Destroy: Removing data from computer hard drives, storage devices & wireless phones

Relevância:

40.00% 40.00%

Publicador:

Resumo:

The concept of energy gap(s) is useful for understanding the consequence of a small daily, weekly, or monthly positive energy balance and the inconspicuous shift in weight gain ultimately leading to overweight and obesity. Energy gap is a dynamic concept: an initial positive energy gap incurred via an increase in energy intake (or a decrease in physical activity) is not constant, may fade out with time if the initial conditions are maintained, and depends on the 'efficiency' with which the readjustment of the energy imbalance gap occurs with time. The metabolic response to an energy imbalance gap and the magnitude of the energy gap(s) can be estimated by at least two methods, i.e. i) assessment by longitudinal overfeeding studies, imposing (by design) an initial positive energy imbalance gap; ii) retrospective assessment based on epidemiological surveys, whereby the accumulated endogenous energy storage per unit of time is calculated from the change in body weight and body composition. In order to illustrate the difficulty of accurately assessing an energy gap we have used, as an illustrative example, a recent epidemiological study which tracked changes in total energy intake (estimated by gross food availability) and body weight over 3 decades in the US, combined with total energy expenditure prediction from body weight using doubly labelled water data. At the population level, the study attempted to assess the cause of the energy gap purported to be entirely due to increased food intake. Based on an estimate of change in energy intake judged to be more reliable (i.e. in the same study population) and together with calculations of simple energetic indices, our analysis suggests that conclusions about the fundamental causes of obesity development in a population (excess intake vs. low physical activity or both) is clouded by a high level of uncertainty.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Background: None of the HIV T-cell vaccine candidates that have reached advanced clinical testing have been able to induce protective T cell immunity. A major reason for these failures may have been suboptimal T cell immunogen designs. Methods: To overcome this problem, we used a novel immunogen design approach that is based on functional T cell response data from more than 1,000 HIV-1 clade B and C infected individuals and which aims to direct the T cell response to the most vulnerable sites of HIV-1. Results: Our approach identified 16 regions in Gag, Pol, Vif and Nef that were relatively conserved and predominantly targeted by individuals with reduced viral loads. These regions formed the basis of the HIVACAT T-cell Immunogen (HTI) sequence which is 529 amino acids in length, includes more than 50 optimally defined CD4+ and CD8+ T-cell epitopes restricted by a wide range of HLA class I and II molecules and covers viral sites where mutations led to a dramatic reduction in viral replicative fitness. In both, C57BL/6 mice and Indian rhesus macaques immunized with an HTI-expressing DNA plasmid (DNA.HTI) induced broad and balanced T-cell responses to several segments within Gag, Pol, and Vif. DNA.HTI induced robust CD4+ and CD8+ T cell responses that were increased by a booster vaccination using modified virus Ankara (MVA.HTI), expanding the DNA.HTI induced response to up to 3.2% IFN-γ T-cells in macaques. HTI-specific T cells showed a central and effector memory phenotype with a significant fraction of the IFN-γ+ CD8+ T cells being Granzyme B+ and able to degranulate (CD107a+). Conclusions: These data demonstrate the immunogenicity of a novel HIV-1 T cell vaccine concept that induced broadly balanced responses to vulnerable sites of HIV-1 while avoiding the induction of responses to potential decoy targets that may divert effective T-cell responses towards variable and less protective viral determinants.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Presentation at Open Repositories 2014, Helsinki, Finland, June 9-13, 2014

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Product Data Management (PDM) systems have been utilized within companies since the 1980s. Mainly the PDM systems have been used by large companies. This thesis presents the premise that small and medium-sized companies can also benefit from utilizing the Product Data Management systems. Furthermore, the starting point for the thesis is that the existing PDM systems are either too expensive or do not properly respond to the requirements SMEs have. The aim of this study is to investigate what kinds of requirements and special features SMEs, operating in Finnish manufacturing industry, have towards Product Data Management. Additionally, the target is to create a conceptual model that could fulfill the specified requirements. The research has been carried out as a qualitative case study, in which the research data was collected from ten Finnish companies operating in manufacturing industry. The research data is formed by interviewing key personnel from the case companies. After this, the data formed from the interviews has been processed to comprise a generic set of information system requirements and the information system concept supporting it. The commercialization of the concept is studied in the thesis from the perspective of system development. The aim was to create a conceptual model, which would be economically feasible for both, a company utilizing the system and for a company developing it. For this reason, the thesis has sought ways to scale the system development effort for multiple simultaneous cases. The main methods found were to utilize platform-based thinking and a way to generalize the system requirements, or in other words abstracting the requirements of an information system. The results of the research highlight the special features Finnish manufacturing SMEs have towards PDM. The most significant of the special features is the usage of project model to manage the order-to-delivery –process. This differs significantly from the traditional concepts of Product Data Management presented in the literature. Furthermore, as a research result, this thesis presents a conceptual model of a PDM system, which would be viable for the case companies interviewed during the research. As a by-product, this research presents a synthesized model, found from the literature, to abstract information system requirements. In addition to this, the strategic importance and categorization of information systems within companies has been discussed from the perspective of information system customizations.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Alors que certains mécanismes pourtant jugés cruciaux pour la transformation de la pluie en débit restent peu ou mal compris, le concept de connectivité hydrologique a récemment été proposé pour expliquer pourquoi certains processus sont déclenchés de manière épisodique en fonction des caractéristiques des événements de pluie et de la teneur en eau des sols avant l’événement. L’adoption de ce nouveau concept en hydrologie reste cependant difficile puisqu’il n’y a pas de consensus sur la définition de la connectivité, sa mesure, son intégration dans les modèles hydrologiques et son comportement lors des transferts d’échelles spatiales et temporelles. Le but de ce travail doctoral est donc de préciser la définition, la mesure, l’agrégation et la prédiction des processus liés à la connectivité hydrologique en s’attardant aux questions suivantes : 1) Quel cadre méthodologique adopter pour une étude sur la connectivité hydrologique ?, 2) Comment évaluer le degré de connectivité hydrologique des bassins versants à partir de données de terrain ?, et 3) Dans quelle mesure nos connaissances sur la connectivité hydrologique doivent-elles conduire à la modification des postulats de modélisation hydrologique ? Trois approches d’étude sont différenciées, soit i) une approche de type « boite noire », basée uniquement sur l’exploitation des données de pluie et de débits sans examiner le fonctionnement interne du bassin versant ; ii) une approche de type « boite grise » reposant sur l’étude de données géochimiques ponctuelles illustrant la dynamique interne du bassin versant ; et iii) une approche de type « boite blanche » axée sur l’analyse de patrons spatiaux exhaustifs de la topographie de surface, la topographie de subsurface et l’humidité du sol. Ces trois approches sont ensuite validées expérimentalement dans le bassin versant de l’Hermine (Basses Laurentides, Québec). Quatre types de réponses hydrologiques sont distingués en fonction de leur magnitude et de leur synchronisme, sachant que leur présence relative dépend des conditions antécédentes. Les forts débits enregistrés à l’exutoire du bassin versant sont associés à une contribution accrue de certaines sources de ruissellement, ce qui témoigne d’un lien hydraulique accru et donc d’un fort degré de connectivité hydrologique entre les sources concernées et le cours d’eau. Les aires saturées couvrant des superficies supérieures à 0,85 ha sont jugées critiques pour la genèse de forts débits de crue. La preuve est aussi faite que les propriétés statistiques des patrons d’humidité du sol en milieu forestier tempéré humide sont nettement différentes de celles observées en milieu de prairie tempéré sec, d’où la nécessité d’utiliser des méthodes de calcul différentes pour dériver des métriques spatiales de connectivité dans les deux types de milieux. Enfin, la double existence de sources contributives « linéaires » et « non linéaires » est mise en évidence à l’Hermine. Ces résultats suggèrent la révision de concepts qui sous-tendent l’élaboration et l’exécution des modèles hydrologiques. L’originalité de cette thèse est le fait même de son sujet. En effet, les objectifs de recherche poursuivis sont conformes à la théorie hydrologique renouvelée qui prône l’arrêt des études de particularismes de petite échelle au profit de l’examen des propriétés émergentes des bassins versants telles que la connectivité hydrologique. La contribution majeure de cette thèse consiste ainsi en la proposition d’une définition unifiée de la connectivité, d’un cadre méthodologique, d’approches de mesure sur le terrain, d’outils techniques et de pistes de solution pour la modélisation des systèmes hydrologiques.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

We present a new algorithm called TITANIC for computing concept lattices. It is based on data mining techniques for computing frequent itemsets. The algorithm is experimentally evaluated and compared with B. Ganter's Next-Closure algorithm.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Formal Concept Analysis is an unsupervised learning technique for conceptual clustering. We introduce the notion of iceberg concept lattices and show their use in Knowledge Discovery in Databases (KDD). Iceberg lattices are designed for analyzing very large databases. In particular they serve as a condensed representation of frequent patterns as known from association rule mining. In order to show the interplay between Formal Concept Analysis and association rule mining, we discuss the algorithm TITANIC. We show that iceberg concept lattices are a starting point for computing condensed sets of association rules without loss of information, and are a visualization method for the resulting rules.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

It is generally assumed that the variability of neuronal morphology has an important effect on both the connectivity and the activity of the nervous system, but this effect has not been thoroughly investigated. Neuroanatomical archives represent a crucial tool to explore structure–function relationships in the brain. We are developing computational tools to describe, generate, store and render large sets of three–dimensional neuronal structures in a format that is compact, quantitative, accurate and readily accessible to the neuroscientist. Single–cell neuroanatomy can be characterized quantitatively at several levels. In computer–aided neuronal tracing files, a dendritic tree is described as a series of cylinders, each represented by diameter, spatial coordinates and the connectivity to other cylinders in the tree. This ‘Cartesian’ description constitutes a completely accurate mapping of dendritic morphology but it bears little intuitive information for the neuroscientist. In contrast, a classical neuroanatomical analysis characterizes neuronal dendrites on the basis of the statistical distributions of morphological parameters, e.g. maximum branching order or bifurcation asymmetry. This description is intuitively more accessible, but it only yields information on the collective anatomy of a group of dendrites, i.e. it is not complete enough to provide a precise ‘blueprint’ of the original data. We are adopting a third, intermediate level of description, which consists of the algorithmic generation of neuronal structures within a certain morphological class based on a set of ‘fundamental’, measured parameters. This description is as intuitive as a classical neuroanatomical analysis (parameters have an intuitive interpretation), and as complete as a Cartesian file (the algorithms generate and display complete neurons). The advantages of the algorithmic description of neuronal structure are immense. If an algorithm can measure the values of a handful of parameters from an experimental database and generate virtual neurons whose anatomy is statistically indistinguishable from that of their real counterparts, a great deal of data compression and amplification can be achieved. Data compression results from the quantitative and complete description of thousands of neurons with a handful of statistical distributions of parameters. Data amplification is possible because, from a set of experimental neurons, many more virtual analogues can be generated. This approach could allow one, in principle, to create and store a neuroanatomical database containing data for an entire human brain in a personal computer. We are using two programs, L–NEURON and ARBORVITAE, to investigate systematically the potential of several different algorithms for the generation of virtual neurons. Using these programs, we have generated anatomically plausible virtual neurons for several morphological classes, including guinea pig cerebellar Purkinje cells and cat spinal cord motor neurons. These virtual neurons are stored in an online electronic archive of dendritic morphology. This process highlights the potential and the limitations of the ‘computational neuroanatomy’ strategy for neuroscience databases.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

This paper reviews the literature concerning the practice of using Online Analytical Processing (OLAP) systems to recall information stored by Online Transactional Processing (OLTP) systems. Such a review provides a basis for discussion on the need for the information that are recalled through OLAP systems to maintain the contexts of transactions with the data captured by the respective OLTP system. The paper observes an industry trend involving the use of OLTP systems to process information into data, which are then stored in databases without the business rules that were used to process information and data stored in OLTP databases without associated business rules. This includes the necessitation of a practice, whereby, sets of business rules are used to extract, cleanse, transform and load data from disparate OLTP systems into OLAP databases to support the requirements for complex reporting and analytics. These sets of business rules are usually not the same as business rules used to capture data in particular OLTP systems. The paper argues that, differences between the business rules used to interpret these same data sets, risk gaps in semantics between information captured by OLTP systems and information recalled through OLAP systems. Literature concerning the modeling of business transaction information as facts with context as part of the modelling of information systems were reviewed to identify design trends that are contributing to the design quality of OLTP and OLAP systems. The paper then argues that; the quality of OLTP and OLAP systems design has a critical dependency on the capture of facts with associated context, encoding facts with contexts into data with business rules, storage and sourcing of data with business rules, decoding data with business rules into the facts with the context and recall of facts with associated contexts. The paper proposes UBIRQ, a design model to aid the co-design of data with business rules storage for OLTP and OLAP purposes. The proposed design model provides the opportunity for the implementation and use of multi-purpose databases, and business rules stores for OLTP and OLAP systems. Such implementations would enable the use of OLTP systems to record and store data with executions of business rules, which will allow for the use of OLTP and OLAP systems to query data with business rules used to capture the data. Thereby ensuring information recalled via OLAP systems preserves the contexts of transactions as per the data captured by the respective OLTP system.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

An important application of Big Data Analytics is the real-time analysis of streaming data. Streaming data imposes unique challenges to data mining algorithms, such as concept drifts, the need to analyse the data on the fly due to unbounded data streams and scalable algorithms due to potentially high throughput of data. Real-time classification algorithms that are adaptive to concept drifts and fast exist, however, most approaches are not naturally parallel and are thus limited in their scalability. This paper presents work on the Micro-Cluster Nearest Neighbour (MC-NN) classifier. MC-NN is based on an adaptive statistical data summary based on Micro-Clusters. MC-NN is very fast and adaptive to concept drift whilst maintaining the parallel properties of the base KNN classifier. Also MC-NN is competitive compared with existing data stream classifiers in terms of accuracy and speed.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Concept drift, which refers to non stationary learning problems over time, has increasing importance in machine learning and data mining. Many concept drift applications require fast response, which means an algorithm must always be (re)trained with the latest available data. But the process of data labeling is usually expensive and/or time consuming when compared to acquisition of unlabeled data, thus usually only a small fraction of the incoming data may be effectively labeled. Semi-supervised learning methods may help in this scenario, as they use both labeled and unlabeled data in the training process. However, most of them are based on assumptions that the data is static. Therefore, semi-supervised learning with concept drifts is still an open challenging task in machine learning. Recently, a particle competition and cooperation approach has been developed to realize graph-based semi-supervised learning from static data. We have extend that approach to handle data streams and concept drift. The result is a passive algorithm which uses a single classifier approach, naturally adapted to concept changes without any explicit drift detection mechanism. It has built-in mechanisms that provide a natural way of learning from new data, gradually "forgetting" older knowledge as older data items are no longer useful for the classification of newer data items. The proposed algorithm is applied to the KDD Cup 1999 Data of network intrusion, showing its effectiveness.