996 resultados para sharing features
Resumo:
A lability criterion is developed for dynamic metal binding by colloidal ligands with convective diffusion as the dominant mode of mass transport. Scanned stripping chronopotentiometric measurements of Pb(II) and Cd(II) binding by carboxylated latex core-shell particles were in good agreement with the predicted values. The dynamic features of metal ion binding by these particles illustrate that the conventional approach of assuming a smeared-out homogeneous ligand distribution overestimates the lability of a colloidal ligand system. Due to the nature of the spatial distribution of the binding sites, the change in lability of a metal species with changing ligand concentration depends on whether the ligand concentration is varied via manipulation of the pH (degree of protonation) or via the particle concentration. In the former case the local ligand density varies, whereas in the latter case it is constant. This feature provides a useful diagnostic tool for the presence of geometrically constrained binding sites.
Resumo:
Tese de doutoramento, Ciências Biomédicas (Neurociências), Universidade de Lisboa, Faculdade de Medicina, 2015
Resumo:
Senior thesis written for Oceanography 445
Resumo:
Senior thesis written for Oceanography 445
Resumo:
Researchers want to run scientific experiments focusing on their disciplines. They do not want to know how and where the experiments are executed. Science gateways hide details by coordinating the execution of experiments using different infrastructures and workflow systems. ER-flow/SHIWA and SCI-BUS project developed repositories to share artefacts such as applications, portlets, workflows, etc. inside and among research communities. Sharing artefacts in re-positories enable gateway developers to reuse them when building a new gateway and/or creating a new application.
Resumo:
E-scientists want to run their scientific experiments on Distributed Computing Infrastructures (DCI) to be able to access large pools of resources and services. To run experiments on these infrastructures requires specific expertise that e-scientists may not have. Workflows can hide resources and services as a virtualization layer providing a user interface that e-scientists can use. There are many workflow systems used by research communities but they are not interoperable. To learn a workflow system and create workflows in this workflow system may require significant efforts from e-scientists. Considering these efforts it is not reasonable to expect that research communities will learn new workflow systems if they want to run workflows developed in other workflow systems. The solution is to create workflow interoperability solutions to allow workflow sharing. The FP7 Sharing Interoperable Workflow for Large-Scale Scientific Simulation on Available DCIs (SHIWA) project developed two interoperability solutions to support workflow sharing: Coarse-Grained Interoperability (CGI) and Fine-Grained Interoperability (FGI). The project created the SHIWA Simulation Platform (SSP) to implement the Coarse-Grained Interoperability approach as a production-level service for research communities. The paper describes the CGI approach and how it enables sharing and combining existing workflows into complex applications and run them on Distributed Computing Infrastructures. The paper also outlines the architecture, components and usage scenarios of the simulation platform.
Resumo:
The task of this work is to apply thoughts from Georg Lukács’ final book, the Ontology of Social Being, for the theoretical analysis of cultural and digital labour. It discusses Lukács’ concepts of work and communication and relates them to the analysis of cultural and digital work. It also analyses his conception of the relation of labour and ideology and points out how we can make use of it for critically understanding social media ideologies. Lukács opposes the dualist separation of the realms of work and ideas. He introduces in this context the notion of teleological positing that allows us to better understand cultural and digital labour as well as associated ideologies, such as the engaging/connecting/sharing-ideology, today. The analysis shows that Lukács’ Ontology is in the age of Facebook, YouTube, and Twitter still a very relevant book, although it has thus far not received the attention that it deserves. This article also introduces the Ontology’s main ideas on work and culture, which is important because large parts of the book have not been translated from the German original into English. Lukács’ notion of teleological positing is crucial for understanding the common features of the economy and culture.
Resumo:
— In the new learning environments, built width digital technologies, the need to promote quality of education resources, commonly known as Learning Objects, which can support formal and informal distance learning, emerge as one of the biggest challenge that educational institutions will have to face. Due to the fact that is expensive, the reuse and sharing became very important issue. This article presents a Learning Object Repository which aims to store, to disseminate and maintain accessible Learning Objects.
Resumo:
The IEEE 802.15.4 standard provides appealing features to simultaneously support real-time and non realtime traffic, but it is only capable of supporting real-time communications from at most seven devices. Additionally, it cannot guarantee delay bounds lower than the superframe duration. Motivated by this problem, in this paper we propose an Explicit Guaranteed time slot Sharing and Allocation scheme (EGSA) for beacon-enabled IEEE 802.15.4 networks. This scheme is capable of providing tighter delay bounds for real-time communications by splitting the Contention Free access Period (CFP) into smaller mini time slots and by means of a new guaranteed bandwidth allocation scheme for a set of devices with periodic messages. At the same the novel bandwidth allocation scheme can maximize the duration of the CFP for non real-time communications. Performance analysis results show that the EGSA scheme works efficiently and outperforms competitor schemes both in terms of guaranteed delay and bandwidth utilization.
Resumo:
Several projects in the recent past have aimed at promoting Wireless Sensor Networks as an infrastructure technology, where several independent users can submit applications that execute concurrently across the network. Concurrent multiple applications cause significant energy-usage overhead on sensor nodes, that cannot be eliminated by traditional schemes optimized for single-application scenarios. In this paper, we outline two main optimization techniques for reducing power consumption across applications. First, we describe a compiler based approach that identifies redundant sensing requests across applications and eliminates those. Second, we cluster the radio transmissions together by concatenating packets from independent applications based on Rate-Harmonized Scheduling.
Resumo:
We present a 12(1 + 3R/(4m)) competitive algorithm for scheduling implicit-deadline sporadic tasks on a platform comprising m processors, where a task may request one of R shared resources.
Resumo:
This paper proposes a dynamic scheduler that supports the coexistence of guaranteed and non-guaranteed bandwidth servers to efficiently handle soft-tasks’ overloads by making additional capacity available from two sources: (i) residual capacity allocated but unused when jobs complete in less than their budgeted execution time; (ii) stealing capacity from inactive non-isolated servers used to schedule best-effort jobs. The effectiveness of the proposed approach in reducing the mean tardiness of periodic jobs is demonstrated through extensive simulations. The achieved results become even more significant when tasks’ computation times have a large variance.
Resumo:
This paper proposes a new strategy to integrate shared resources and precedence constraints among real-time tasks, assuming no precise information on critical sections and computation times is available. The concept of bandwidth inheritance is combined with a greedy capacity sharing and stealing policy to efficiently exchange bandwidth among tasks, minimising the degree of deviation from the ideal system's behaviour caused by inter-application blocking. The proposed capacity exchange protocol (CXP) focus on exchanging extra capacities as early, and not necessarily as fairly, as possible. This loss of optimality is worth the reduced complexity as the protocol's behaviour nevertheless tends to be fair in the long run and outperforms other solutions in highly dynamic scenarios, as demonstrated by extensive simulations.
Resumo:
Ramsey pricing has been proposed in the pharmaceutical industry as a principle to price discriminate among markets while allowing to recover the (fixed) R&D cost. However, such analyses neglect the presence of insurance or the fund raising costs for most of drug reimbursement. By incorporating these new elements, we aim at providing some building blocks towards an economic theory incorporating Ramsey pricing and insurance coverage. We show how coinsurance affects the optimal prices to pay for the R&D investment. We also show that under certain conditions, there is no strategic incentive by governments to set coinsurance rates in order to shift the financial burden of R&D. This will have important implications to the application of Ramsey pricing principles to pharmaceutical products across countries.