978 resultados para Resource sharing
Resumo:
BACKGROUND: In recent years large bibliographic databases have made much of the published literature of biology available for searches. However, the capabilities of the search engines integrated into these databases for text-based bibliographic searches are limited. To enable searches that deliver the results expected by comparative anatomists, an underlying logical structure known as an ontology is required. DEVELOPMENT AND TESTING OF THE ONTOLOGY: Here we present the Mammalian Feeding Muscle Ontology (MFMO), a multi-species ontology focused on anatomical structures that participate in feeding and other oral/pharyngeal behaviors. A unique feature of the MFMO is that a simple, computable, definition of each muscle, which includes its attachments and innervation, is true across mammals. This construction mirrors the logical foundation of comparative anatomy and permits searches using language familiar to biologists. Further, it provides a template for muscles that will be useful in extending any anatomy ontology. The MFMO is developed to support the Feeding Experiments End-User Database Project (FEED, https://feedexp.org/), a publicly-available, online repository for physiological data collected from in vivo studies of feeding (e.g., mastication, biting, swallowing) in mammals. Currently the MFMO is integrated into FEED and also into two literature-specific implementations of Textpresso, a text-mining system that facilitates powerful searches of a corpus of scientific publications. We evaluate the MFMO by asking questions that test the ability of the ontology to return appropriate answers (competency questions). We compare the results of queries of the MFMO to results from similar searches in PubMed and Google Scholar. RESULTS AND SIGNIFICANCE: Our tests demonstrate that the MFMO is competent to answer queries formed in the common language of comparative anatomy, but PubMed and Google Scholar are not. Overall, our results show that by incorporating anatomical ontologies into searches, an expanded and anatomically comprehensive set of results can be obtained. The broader scientific and publishing communities should consider taking up the challenge of semantically enabled search capabilities.
Resumo:
The factors that are driving the development and use of grids and grid computing, such as size, dynamic features, distribution and heterogeneity, are also pushing to the forefront service quality issues. These include performance, reliability and security. Although grid middleware can address some of these issues on a wider scale, it has also become imperative to ensure adequate service provision at local level. Load sharing in clusters can contribute to the provision of a high quality service, by exploiting both static and dynamic information. This paper is concerned with the presentation of a load sharing scheme, that can satisfy grid computing requirements. It follows a proactive, non preemptive and distributed approach. Load information is gathered continuously before it is needed, and a task is allocated to the most appropriate node for execution. Performance and reliability are enhanced by the decentralised nature of the scheme and the symmetric roles of the nodes. In addition, the scheme exhibits transparency characteristics that facilitate integration with the grid.
Resumo:
This paper presents a proactive approach to load sharing and describes the architecture of a scheme, Concert, based on this approach. A proactive approach is characterized by a shift of emphasis from reacting to load imbalance to avoiding its occurrence. In contrast, in a reactive load sharing scheme, activity is triggered when a processing node is either overloaded or underloaded. The main drawback of this approach is that a load imbalance is allowed to develop before costly corrective action is taken. Concert is a load sharing scheme for loosely-coupled distributed systems. Under this scheme, load and task behaviour information is collected and cached in advance of when it is needed. Concert uses Linux as a platform for development. Implemented partially in kernel space and partially in user space, it achieves transparency to users and applications whilst keeping the extent of kernel modifications to a minimum. Non-preemptive task transfers are used exclusively, motivated by lower complexity, lower overheads and faster transfers. The goal is to minimize the average response-time of tasks. Concert is compared with other schemes by considering the level of transparency it provides with respect to users, tasks and the underlying operating system.
Resumo:
The scheduling problem of minimizing the makespan for m parallel dedicated machines under single resource constraints is considered. For different variants of the problem the complexity status is established. Heuristic algorithms employing the so-called group technology approach are presented and their worst-case behavior is examined. Finally, a polynomial time approximation scheme is presented for the problem with fixed number of machines.
Resumo:
The paper considers scheduling problems for parallel dedicated machines subject to resource constraints. A fairly complete computational complexity classification is obtained, a number of polynomial-time algorithms are designed. For the problem with a fixed number of machines in which a job uses at most one resource of unit size a polynomial-time approximation scheme is offered.
Resumo:
We consider a problem of scheduling jobs on m parallel machines. The machines are dedicated, i.e., for each job the processing machine is known in advance. We mainly concentrate on the model in which at any time there is one unit of an additional resource. Any job may be assigned the resource and this reduces its processing time. A job that is given the resource uses it at each time of its processing. No two jobs are allowed to use the resource simultaneously. The objective is to minimize the makespan. We prove that the two-machine problem is NP-hard in the ordinary sense, describe a pseudopolynomial dynamic programming algorithm and convert it into an FPTAS. For the problem with an arbitrary number of machines we present an algorithm with a worst-case ratio close to 3/2, and close to 3, if a job can be given several units of the resource. For the problem with a fixed number of machines we give a PTAS. Virtually all algorithms rely on a certain variant of the linear knapsack problem (maximization, minimization, multiple-choice, bicriteria). © 2008 Wiley Periodicals, Inc. Naval Research Logistics, 2008
Resumo:
Comments on the Chancery Division decision in Clarence House Ltd v National Westminster Bank Plc on whether the alienation covenant in a lease of commercial premises had been breached by the tenant effecting a virtual assignment of it, under which all the economic benefits and burdens of the lease were transferred to a third party without there being any actually assignment of the leasehold interest or change in occupancy.
Resumo:
This paper studies a two-level supply chain consisting of components supplier and product assembly manufacturer, while the manufacturer shares the investment on shortening supply lead time. The objective of this research is to investigate the benefits of cost sharing strategy and adopting component commonality. The result of numerical analysis demonstrates that using component commonality can help reduce the total cost, especially when the manufacture shares a higher fraction of the cost of investment in shortening supply lead time.
Resumo:
RATIONALE & OBJECTIVES: The food multimix (FFM)concept states that limited food resources can be combined using scientific knowledge to meet nutrient needs of vulnerable groups at low cost utilizing the ‘nutrient strengths’ of individual or candidate foods in composite recipes within a cultural context. METHODS: The method employed the food-to-food approach for recipe development using traditional food ingredients. Recipes were subjected to proximate and micronutrient analysis and optimized to meet at tleast 40% of recommended daily intakes. End products including breads, porridge and soup were developed. RESULTS: FMM products were employed in a feeding trial among 120 healthy pregnant women in Gauteng, South Africa resulting in improvements in serum iron levels from baseline values of 14.59 (=/-7.67) umol/L and 14.02 (=/-8.13) umol/L for control and intervention groups (p=0.71), to 16.03 (=/-5.67) umol/L and 18.66 (=/-9.41) umol/L (p=0.19). The increases from baseline to post-intervention were however statistically significant within groups. Similarly Mean Cell Volume values improved from baseline as well as serum ferritin and transferritin levels. CONCLUSION: The FMM concept has potential value in feeding programs for vulnerable groups including pregnant and lactating mothers.
Resumo:
This study examines the roll-out of a collaborative information repository or 'knowledge-base' in a medium-sized UK professional services firm over a six year period. Data from usage logs provides the basis for analysis of the dynamic evolution of social networks around the depository during this time. The adoption pattern follows an 's-curve' and usage exhibits something of a power law distribution, both attributable to network effects and network opposition is associated with organisational performance on a number of indicators. But periodicity in usage is evident and the usage distribution displays an exponential cut-off. Fourier analysis provides some evidence of mathematical complexity in the periodicity. Some implications of complex patterns in social network data for research and management are discussed.
Resumo:
Deliberating on Enterprise Resource Planning (ERP) software sourcing and provision, this paper contrasts the corporate environment with the small business environment. The paper is about Enterprise Resource Planning client (ERPc) expectations and Enterprise Resource Planning vendor (ERPv) value propositions as a mutually compatible process for achieving acceptable standards of ERP software performance. It is suggested that a less-than-equitable vendor–client relationship would not contribute to the implementation of the optimum solution. Adapting selected theoretical concepts and models, the researchers analyse ERPv to ERPc relationship. This analysis is designed to discover if the provision of the very large ERP vendors who market systems such as SAP, and the provision of the smaller ERP vendors (in this instance Eshbel Technologies Ltd who market an ERP software solution called Priority) when framed as a value proposition (Walters, D. (2002) Operations Strategy. Hampshire, UK: Palgrave), is at all comparable or distinctive.
Resumo:
Here we describe a new trait-based model for cellular resource allocation that we use to investigate the relative importance of different drivers for small cell size in phytoplankton. Using the model, we show that increased investment in nonscalable structural components with decreasing cell size leads to a trade-off between cell size, nutrient and light affinity, and growth rate. Within the most extreme nutrient-limited, stratified environments, resource competition theory then predicts a trend toward larger minimum cell size with increasing depth. We demonstrate that this explains observed trends using a marine ecosystem model that represents selection and adaptation of a diverse community defined by traits for cell size and subcellular resource allocation. This framework for linking cellular physiology to environmental selection can be used to investigate the adaptive response of the marine microbial community to environmental conditions and the adaptive value of variations in cellular physiology.