964 resultados para data centric storage


Relevância:

100.00% 100.00%

Publicador:

Resumo:

Many scientific workflows are data intensive where large volumes of intermediate data are generated during their execution. Some valuable intermediate data need to be stored for sharing or reuse. Traditionally, they are selectively stored according to the system storage capacity, determined manually. As doing science in the cloud has become popular nowadays, more intermediate data can be stored in scientific cloud workflows based on a pay-for-use model. In this paper, we build an intermediate data dependency graph (IDG) from the data provenance in scientific workflows. With the IDG, deleted intermediate data can be regenerated, and as such we develop a novel intermediate data storage strategy that can reduce the cost of scientific cloud workflow systems by automatically storing appropriate intermediate data sets with one cloud service provider. The strategy has significant research merits, i.e. it achieves a cost-effective trade-off of computation cost and storage cost and is not strongly impacted by the forecasting inaccuracy of data sets' usages. Meanwhile, the strategy also takes the users' tolerance of data accessing delay into consideration. We utilize Amazon's cost model and apply the strategy to general random as well as specific astrophysics pulsar searching scientific workflows for evaluation. The results show that our strategy can reduce the overall cost of scientific cloud workflow execution significantly.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Queensland University of Technology (QUT) is faced with a rapidly growing research agenda built upon a strategic research capacity-building program. This presentation will outline the results of a project that has recently investigated QUT’s research support requirements and which has developed a model for the support of eResearch across the university. QUT’s research building strategy has produced growth at the faculty level and within its research institutes. This increased research activity is pushing the need for university-wide eResearch platforms capable of providing infrastructure and support in areas such as collaboration, data, networking, authentication and authorisation, workflows and the grid. One of the driving forces behind the investigation is data-centric nature of modern research. It is now critical that researchers have access to supported infrastructure that allows the collection, analysis, aggregation and sharing of large data volumes for exploration and mining in order to gain new insights and to generate new knowledge. However, recent surveys into current research data management practices by the Australian Partnership for Sustainable Repositories (APSR) and by QUT itself, has revealed serious shortcomings in areas such as research data management, especially its long term maintenance for reuse and authoritative evidence of research findings. While these internal university pressures are building, at the same time there are external pressures that are magnifying them. For example, recent compliance guidelines from bodies such as the ARC, and NHMRC and Universities Australia indicate that institutions need to provide facilities for the safe and secure storage of research data along with a surrounding set of policies, on its retention, ownership and accessibility. The newly formed Australian National Data Service (ANDS) is developing strategies and guidelines for research data management and research institutions are a central focus, responsible for managing and storing institutional data on platforms that can be federated nationally and internationally for wider use. For some time QUT has recognised the importance of eResearch and has been active in a number of related areas: ePrints to digitally publish research papers, grid computing portals and workflows, institutional-wide provisioning and authentication systems, and legal protocols for copyright management. QUT also has two widely recognised centres focused on fundamental research into eResearch itself: The OAK LAW project (Open Access to Knowledge) which focuses upon legal issues relating eResearch and the Microsoft QUT eResearch Centre whose goal is to accelerate scientific research discovery, through new smart software. In order to better harness all of these resources and improve research outcomes, the university recently established a project to investigate how it might better organise the support of eResearch. This presentation will outline the project outcomes, which include a flexible and sustainable eResearch support service model addressing short and longer term research needs, identification of resource requirements required to establish and sustain the service, and the development of research data management policies and implementation plans.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Regional accreditors often desire the same metrics and data collected by professional program reviews, but may use different terminologies to describe this information; as a result, some schools must manually translate or even recollect data already stored. This report profiles strategies to proactively consolidate the language and policies of accreditation to avoid duplication of labor and to efficiently route program accreditation data that will be repurposed in regional review. It also suggests ways to select new technologies that can streamline data collection, storage, and presentation.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Diese Dissertation stellt das neu entwickelte SystemRelAndXML vor, das für das Management und dieSpeicherung von hypertextzentrierten XML-Dokumenten und dendazugehörenden XSL-Stylesheets spezialisiert ist. DerAnwendungsbereich sind die Vorlesungsmaterialien anUniversitäten. RelAndXML speichert die XML-formatiertenÜbungsblätter in Textbausteinen und weiterenTeilen in einer speziellen Datenbank.Die Speicherung von XML-Dokumenten in Datenbanken ist seiteinigen Jahren ein wichtiges Thema der Datenbankforschung.Ansätze dafür gliedern sich in solche fürdatenzentrierte und andere für dokumentenzentrierteDokumente. Die Dissertation präsentiert einen Ansatzzur Speicherung von hypertextzentrierten XML-Dokumenten, derAspekte von datenzentrierten und dokumentenzentriertenAnsätzen kombiniert. Der Ansatz erlaubt dieWiederverwendung von Textbausteinen und speichert dieReihenfolge dort, wo sie wichtig ist. Mit RelAndXML könnennicht nur Elemente gespeichert werden, wie mit einigenanderen Ansätzen, sondern auch Attribute, Kommentareund Processing Instructions. Algorithmen für dieFragmentierung und Rekonstruktion von Dokumenten werdenbereit gestellt.RelAndXML wurde mit Java und unter Verwendung einerobjektrelationalen Datenbank implementiert. Das System hateine graphische Benutzungsoberfläche, die das Erstellenund Verändern der XML- und XSL-Dokumente, dasEinfügen von neuen oder schon gespeichertenTextbausteinen sowie das Erzeugen von HTML-Dokumenten zurVeröffentlichung ermöglicht.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

We use electronic communication networks for more than simply traditional telecommunications: we access the news, buy goods online, file our taxes, contribute to public debate, and more. As a result, a wider array of privacy interests is implicated for users of electronic communications networks and services. . This development calls into question the scope of electronic communications privacy rules. This paper analyses the scope of these rules, taking into account the rationale and the historic background of the European electronic communications privacy framework. We develop a framework for analysing the scope of electronic communications privacy rules using three approaches: (i) a service-centric approach, (ii) a data-centric approach, and (iii) a value-centric approach. We discuss the strengths and weaknesses of each approach. The current e-Privacy Directive contains a complex blend of the three approaches, which does not seem to be based on a thorough analysis of their strengths and weaknesses. The upcoming review of the directive announced by the European Commission provides an opportunity to improve the scoping of the rules.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Background: Statistical shape models are widely used in biomedical research. They are routinely implemented for automatic image segmentation or object identification in medical images. In these fields, however, the acquisition of the large training datasets, required to develop these models, is usually a time-consuming process. Even after this effort, the collections of datasets are often lost or mishandled resulting in replication of work. Objective: To solve these problems, the Virtual Skeleton Database (VSD) is proposed as a centralized storage system where the data necessary to build statistical shape models can be stored and shared. Methods: The VSD provides an online repository system tailored to the needs of the medical research community. The processing of the most common image file types, a statistical shape model framework, and an ontology-based search provide the generic tools to store, exchange, and retrieve digital medical datasets. The hosted data are accessible to the community, and collaborative research catalyzes their productivity. Results: To illustrate the need for an online repository for medical research, three exemplary projects of the VSD are presented: (1) an international collaboration to achieve improvement in cochlear surgery and implant optimization, (2) a population-based analysis of femoral fracture risk between genders, and (3) an online application developed for the evaluation and comparison of the segmentation of brain tumors. Conclusions: The VSD is a novel system for scientific collaboration for the medical image community with a data-centric concept and semantically driven search option for anatomical structures. The repository has been proven to be a useful tool for collaborative model building, as a resource for biomechanical population studies, or to enhance segmentation algorithms.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

The manipulation and handling of an ever increasing volume of data by current data-intensive applications require novel techniques for e?cient data management. Despite recent advances in every aspect of data management (storage, access, querying, analysis, mining), future applications are expected to scale to even higher degrees, not only in terms of volumes of data handled but also in terms of users and resources, often making use of multiple, pre-existing autonomous, distributed or heterogeneous resources.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

The project has further developed two programs for the industry partners related to service life prediction and salt deposition. The program for Queensland Department of Main Roads which predicts salt deposition on different bridge structures at any point in Queensland has been further refined by looking at more variables. It was found that the height of the bridge significantly affects the salt deposition levels only when very close to the coast. However the effect of natural cleaning of salt by rainfall was incorporated into the program. The user interface allows selection of a location in Queensland, followed by a bridge component. The program then predicts the annual salt deposition rate and rates the likely severity of the environment. The service life prediction program for the Queensland Department of Public Works has been expanded to include 10 common building components, in a variety of environments. Data mining procedures have been used to develop the program and increase the usefulness of the application. A Query Based Learning System (QBLS) has been developed which is based on a data-centric model with extensions to provide support for user interaction. The program is based on number of sources of information about the service life of building components. These include the Delphi survey, the CSIRO Holistic model and a school survey. During the project, the Holistic model was modified for each building component and databases generated for the locations of all Queensland schools. Experiments were carried out to verify and provide parameters for the modelling. These included instrumentation of a downpipe, measurements on pH and chloride levels in leaf litter, EIS measurements and chromate leaching from Colorbond materials and dose tests to measure corrosion rates of new materials. A further database was also generated for inclusion in the program through a large school survey. Over 30 schools in a range of environments from tropical coastal to temperate inland were visited and the condition of the building components rated on a scale of 0-5. The data was analysed and used to calculate an average service life for each component/material combination in the environments, where sufficient examples were available.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

In this paper, we describe ongoing work on online banking customization with a particular focus on interaction. The scope of the study is confined to the Australian banking context where the lack of customization is evident. This paper puts forward the notion of using tags to facilitate personalized interactions in online banking. We argue that tags can afford simple and intuitive interactions unique to every individual in both online and mobile environments. Firstly, through a review of related literature, we frame our work in the customization domain. Secondly, we define a range of taggable resources in online banking. Thirdly, we describe our preliminary prototype implementation with respect to interaction customization types. Lastly, we conclude with a discussion on future work.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

High fidelity simulation as a teaching and learning approach is being embraced by many schools of nursing. Our school embarked on integrating high fidelity (HF) simulation into the undergraduate clinical education program in 2011. Low and medium fidelity simulation has been used for many years, but this did not simplify the integration of HF simulation. Alongside considerations of how and where HF simulation would be integrated, issues arose with: student consent and participation for observed activities; data management of video files; staff development, and conceptualising how methods for student learning could be researched. Simulation for undergraduate student nurses commenced as a formative learning activity, undertaken in groups of eight, where four students undertake the ‘doing’ role and four are structured observers, who then take a formal role in the simulation debrief. Challenges for integrating simulation into student learning included conceptualising and developing scenarios to trigger students’ decision making and application of skills, knowledge and attitudes explicit to solving clinical ‘problems’. Developing and planning scenarios for students to ‘try out’ skills and make decisions for problem solving lay beyond choosing pre-existing scenarios inbuilt with the software. The supplied scenarios were not concept based but rather knowledge, skills and technology (of the manikin) focussed. Challenges lay in using the technology for the purpose of building conceptual mastery rather than using technology simply because it was available. As we integrated use of HF simulation into the final year of the program, focus was on building skills, knowledge and attitudes that went beyond technical skill, and provided an opportunity to bridge the gap with theory-based knowledge that students often found difficult to link to clinical reality. We wished to provide opportunities to develop experiential knowledge based on application and clinical reasoning processes in team environments where problems are encountered, and to solve them, the nurse must show leadership and direction. Other challenges included students consenting for simulations to be videotaped and ethical considerations of this. For example if one student in a group of eight did not consent, did this mean they missed the opportunity to undertake simulation, or that others in the group may be disadvantaged by being unable to review their performance. This has implications for freely given consent but also for equity of access to learning opportunities for students who wished to be taped and those who did not. Alongside this issue were the details behind data management, storage and access. Developing staff with varying levels of computer skills to use software and undertake a different approach to being the ‘teacher’ required innovation where we took an experiential approach. Considering explicit learning approaches to be trialled for learning was not a difficult proposition, but considering how to enact this as research with issues of blinding, timetabling of blinded groups, and reducing bias for testing results of different learning approaches along with gaining ethical approval was problematic. This presentation presents examples of these challenges and how we overcame them.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

INEX investigates focused retrieval from structured documents by providing large test collections of structured documents, uniform evaluation measures, and a forum for organizations to compare their results. This paper reports on the INEX 2011 evaluation campaign, which consisted of a five active tracks: Books and Social Search, Data Centric, Question Answering, Relevance Feedback, and Snippet Retrieval. INEX 2011 saw a range of new tasks and tracks, such as Social Book Search, Faceted Search, Snippet Retrieval, and Tweet Contextualization.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Democratic governments raise taxes and charges and spend revenue on delivering peace, order and good government. The delivery process begins with a legislature as that can provide a framework of legally enforceable rules enacted according to the government’s constitution. These rules confer rights and obligations that allow particular people to carry on particular functions at particular places and times. Metadata standards as applied to public records contain information about the functioning of government as distinct from the non-government sector of society. Metadata standards apply to database construction. Data entry, storage, maintenance, interrogation and retrieval depend on a controlled vocabulary needed to enable accurate retrieval of suitably catalogued records in a global information environment. Queensland’s socioeconomic progress now depends in part on technical efficiency in database construction to address queries about who does what, where and when; under what legally enforceable authority; and how the evidence of those facts is recorded. The Survey and Mapping Infrastructure Act 2003 (Qld) addresses technical aspects of where questions – typically the officially recognised name of a place and a description of its boundaries. The current 10-year review of the Survey and Mapping Regulation 2004 provides a valuable opportunity to consider whether the Regulation makes sense in the context of a number of later laws concerned with management of Public Sector Information (PSI) as well as policies for ICT hardware and software procurement. Removing ambiguities about how official place names are to be regarded on a whole-of-government basis can achieve some short term goals. Longer-term goals depend on a more holistic approach to information management – and current aspirations for more open government and community engagement are unlikely to occur without such a longer-term vision.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

The development and maintenance of large and complex ontologies are often time-consuming and error-prone. Thus, automated ontology learning and revision have attracted intensive research interest. In data-centric applications where ontologies are designed or automatically learnt from the data, when new data instances are added that contradict to the ontology, it is often desirable to incrementally revise the ontology according to the added data. This problem can be intuitively formulated as the problem of revising a TBox by an ABox. In this paper we introduce a model-theoretic approach to such an ontology revision problem by using a novel alternative semantic characterisation of DL-Lite ontologies. We show some desired properties for our ontology revision. We have also developed an algorithm for reasoning with the ontology revision without computing the revision result. The algorithm is efficient as its computational complexity is in coNP in the worst case and in PTIME when the size of the new data is bounded.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

The importance of passenger experience in aviation has become well understood in the last several years. It is now generally accepted that the provision of good passenger experience is not an option, but a necessity, from an aviation profitability perspective. In this paper, we paint a picture of the future passenger experience by consolidating a number of industry and research perspectives. Using the future passenger experience as a starting point, we explore the components needed to enable this future vision. From this bottom-up approach, we identify the need to resolve data formatting and data ownership issues. The resolution of these data integration issues is necessary to enable the seamless future travel experience that is envisioned by the aviation industry. By looking at the passenger experience from this bottom-up, data centric perspective, we identify a potential shift in the way that future passenger terminals will be designed. Whereas currently the design of terminals is largely an architectural practice, in the near future, the design of the terminal building may become more of a virtual technology practice. This of course will pose a new set of challenges to designers of airport terminal environments.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

The methane hydration process is investigated in a semi-continuous stirred tank reactor. Liquid temperatures and reaction rates without stirrer are compared with those occurring with stirrer, while at the same time better stirring conditions of the methane hydration process are given by the experiments. Some basic data of fluid mechanics, for example, stirring Reynolds number, Froucle number and stirrer power, are calculated during the methane hydration process, which can be applied to evaluate stirrer capacity and provide some basic data for a scaled up reactor. Based on experiment and calculations in this work, some conclusions are drawn. First, the stirrer has great influence on the methane hydration process. Batch stirring is helpful to improve the mass transfer and heat transfer performances of the methane hydration process. Second, induction time can be shortened effectively by use of the stirrer. Third, in this paper, the appropriate stirring velocity and stirring time were 320 rpm and 30 min, respectively, at 5.0 MPa, for which the storage capacity and reaction time were 159.1 V/V and 370 min, respectively. Under the condition of the on-flow state, the initial stirring Reynolds number of the fluid and the stirring power were 12,150 and 0.54 W, respectively. Fourth, some suggestions, for example, the use of another type of stirrer or some baffles, are proposed to accelerate the methane hydration process. Comparing with literature data, higher storage capacity and hydration rate are achieved in this work. Moreover, some fluid mechanics parameters are calculated, which can provide some references to engineering application.