26 resultados para software reuse
Resumo:
[This abstract is based on the authors' abstract.]Three new standards to be applied when adopting commercial computer off-the-shelf (COTS) software solutions are discussed. The first standard is for a COTS software life cycle, the second for a software solution user requirements life cycle, and the third is a checklist to help in completing the requirements. The standards are based on recent major COTS software solution implementations.
Resumo:
The parallelization of real-world compute intensive Fortran application codes is generally not a trivial task. If the time to complete the parallelization is to be significantly reduced then an environment is needed that will assist the programmer in the various tasks of code parallelization. In this paper the authors present a code parallelization environment where a number of tools that address the main tasks such as code parallelization, debugging and optimization are available. The ParaWise and CAPO parallelization tools are discussed which enable the near automatic parallelization of real-world scientific application codes for shared and distributed memory-based parallel systems. As user involvement in the parallelization process can introduce errors, a relative debugging tool (P2d2) is also available and can be used to perform nearly automatic relative debugging of a program that has been parallelized using the tools. A high quality interprocedural dependence analysis as well as user-tool interaction are also highlighted and are vital to the generation of efficient parallel code and in the optimization of the backtracking and speculation process used in relative debugging. Results of benchmark and real-world application codes parallelized are presented and show the benefits of using the environment
Resumo:
In this chapter we look at JOSTLE, the multilevel graph-partitioning software package, and highlight some of the key research issues that it addresses. We first outline the core algorithms and place it in the context of the multilevel refinement paradigm. We then look at issues relating to its use as a tool for parallel processing and, in particular, partitioning in parallel. Since its first release in 1995, JOSTLE has been used for many mesh-based parallel scientific computing applications and so we also outline some enhancements such as multiphase mesh-partitioning, heterogeneous mapping and partitioning to optimise subdomain shape
Resumo:
This paper describes a methodology for embedding dynamic behaviour into software components. The implications and system architecture requirements to support this adaptivity are discussed. This work is part of a European Commission funded and industry supported project to produce a reconfigurable middleware for use in automotive systems. Such systems must be trustable against illegal internal behaviour and activity with external origins, additional devices for example. Policy-based computing is used here as an example of embedded logic. A key contribution of this work is the way in which static and dynamic aspects of the system are interfaced, such that the behaviour can be changed very flexibly (even during run-time), without modification, recompilation or redeployment of the embedded application code. An implementation of these concepts is presented, focussing on achieving trust in the use of dynamic behaviour.
Resumo:
This paper describes a highly flexible component architecture, primarily designed for automotive control systems, that supports distributed dynamically- configurable context-aware behaviour. The architecture enforces a separation of design-time and run-time concerns, enabling almost all decisions concerning runtime composition and adaptation to be deferred beyond deployment. Dynamic context management contributes to flexibility. The architecture is extensible, and can embed potentially many different self-management decision technologies simultaneously. The mechanism that implements the run-time configuration has been designed to be very robust, automatically and silently handling problems arising from the evaluation of self- management logic and ensuring that in the worst case the dynamic aspects of the system collapse down to static behavior in totally predictable ways.
Resumo:
This short position paper considers issues in developing Data Architecture for the Internet of Things (IoT) through the medium of an exemplar project, Domain Expertise Capture in Authoring and Development Environments (DECADE). A brief discussion sets the background for IoT, and the development of the distinction between things and computers. The paper makes a strong argument to avoid reinvention of the wheel, and to reuse approaches to distributed heterogeneous data architectures and the lessons learned from that work, and apply them to this situation. DECADE requires an autonomous recording system, local data storage, semi-autonomous verification model, sign-off mechanism, qualitative and quantitative analysis carried out when and where required through web-service architecture, based on ontology and analytic agents, with a self-maintaining ontology model. To develop this, we describe a web-service architecture, combining a distributed data warehouse, web services for analysis agents, ontology agents and a verification engine, with a centrally verified outcome database maintained by certifying body for qualification/professional status.
Resumo:
The role intra-organizational knowledge exchanges play in innovation processes has been widely acknowledged in the organizational literature. This paper contributes to the understanding of which specific configurations knowledge networks assume during different phases of radical and incremental innovation processes. The case study we selected is a FLOSS (Free/Libre Open Source Software) community consisting of 233 developers committed to the development of a web browser application since November 2002. By harvesting the mailing list, official blog and code repository of a FLOSS community, we investigate the patterns of knowledge exchange and individual contributions of its developers. We measure structural cohesion and compare global and local network properties at different points in time. Preliminary results show that phases of radical and incremental innovation are associated with specific configurations of the knowledge network as a whole as well as with different network positions of the core developers of the software.
Resumo:
This paper presents a framework to integrate requirements management and design knowledge reuse. The research approach begins with a literature review in design reuse and requirements management to identify appropriate methods within each domain. A framework is proposed based on the identified requirements. The framework is then demonstrated using a case study example: vacuum pump design. Requirements are presented as a component of the integrated design knowledge framework. The proposed framework enables the application of requirements management as a dynamic process, including capture, analysis and recording of requirements. It takes account of the evolving requirements and the dynamic nature of the interaction between requirements and product structure through the various stages of product development.
Resumo:
This paper describes a practical approach for the investigation, assessment and design of existing soakaways. This method can be utilised for measuring the performance and capacity of the systems and examining whether the systems are suitable for reuse when information about the design and installation of the systems is not available. The requirements for field observations and the procedure for a soil infiltration test for the installed system are suggested for successful assessment. The soil infiltration rate of the system is estimated from the field test data without requiring information on the design and construction details of the system. The system's working condition is measured by a performance indicator related to the time taken to empty the soakaway. This is then employed to evaluate the potential reuse of the system. The system's drain capacity is determined by the design principles of current practice and the effect of climate change on its drain capacity is considered. Contamination of soils around the systems after long-term use of discharge service and the water present in soakaway chambers are also investigated. A detailed case study for the reuse of four installed soakaways for a new housing development demonstrates how the proposed approach provides a straightforward process for the infiltration performance and drain capacity assessment of the existing systems. The effectiveness and applicability of the proposed approach are further demonstrated from the assessments for a number of installed systems over various sites
Resumo:
Product knowledge support needs are compared in two companies with different production volumes and product complexity. Knowledge support requirements identified include: function, performance data, requirements data, common parts, regulatory guidelines and layout data. A process based data driven knowledge reuse method is evaluated in light of the identified product knowledge needs. The evaluation takes place through developing a pilot case with each company. It is found that the method provides more benefit to the high complexity design domain, in which a significant amount of work takes place at the conceptual design stages, relying on a conceptual product representation. There is not such a clear value proposition in a design environment whose main challenge is layout design and the application of standard parts and features. The method supports the requirement for conceptual product representation but does not fully support a standard parts library.
Resumo:
Deliberating on Enterprise Resource Planning (ERP) software sourcing and provision, this paper contrasts the corporate environment with the small business environment. The paper is about Enterprise Resource Planning client (ERPc) expectations and Enterprise Resource Planning vendor (ERPv) value propositions as a mutually compatible process for achieving acceptable standards of ERP software performance. It is suggested that a less-than-equitable vendor–client relationship would not contribute to the implementation of the optimum solution. Adapting selected theoretical concepts and models, the researchers analyse ERPv to ERPc relationship. This analysis is designed to discover if the provision of the very large ERP vendors who market systems such as SAP, and the provision of the smaller ERP vendors (in this instance Eshbel Technologies Ltd who market an ERP software solution called Priority) when framed as a value proposition (Walters, D. (2002) Operations Strategy. Hampshire, UK: Palgrave), is at all comparable or distinctive.