17 resultados para Non-functional requirement. Software architecture. NFR-framework. Architectural pattern


Relevância:

100.00% 100.00%

Publicador:

Resumo:

Specification of the non-functional requirements of applications and determining the required resources for their execution are activities that demand a great deal of technical knowledge, frequently resulting in an inefficient use of resources. Cloud computing is an alternative for provisioning of resources, which can be done using either the provider's own infrastructure or the infrastructure of one or more public clouds, or even a combination of both. It enables more flexibly/elastic use of resources, but does not solve the specification problem. In this paper we present an approach that uses models at runtime to facilitate the specification of non-functional requirements and resources, aiming to facilitate dynamic support for application execution in cloud computing environments with shared resources. © 2013 IEEE.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Software architecture plays an essential role in the high level description of a system design, where the structure and communication are emphasized. Despite its importance in the software engineering process, the lack of formal description and automated verification hinders the development of good software architecture models. In this paper, we present an approach to support the rigorous design and verification of software architecture models using the semantic web technology. We view software architecture models as ontology representations, where their structures and communication constraints are captured by the Web Ontology Language (OWL) and the Semantic Web Rule Language (SWRL). Specific configurations on the design are represented as concrete instances of the ontology, to which their structures and dynamic behaviors must conform. Furthermore, ontology reasoning tools can be applied to perform various automated verification on the design to ensure correctness, such as consistency checking, style recognition, and behavioral inference.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Automated negotiation is widely applied in various domains. However, the development of such systems is a complex knowledge and software engineering task. So, a methodology there will be helpful. Unfortunately, none of existing methodologies can offer sufficient, detailed support for such system development. To remove this limitation, this paper develops a new methodology made up of: (1) a generic framework (architectural pattern) for the main task, and (2) a library of modular and reusable design pattern (templates) of subtasks. Thus, it is much easier to build a negotiating agent by assembling these standardised components rather than reinventing the wheel each time. Moreover, since these patterns are identified from a wide variety of existing negotiating agents (especially high impact ones), they can also improve the quality of the final systems developed. In addition, our methodology reveals what types of domain knowledge need to be input into the negotiating agents. This in turn provides a basis for developing techniques to acquire the domain knowledge from human users. This is important because negotiation agents act faithfully on the behalf of their human users and thus the relevant domain knowledge must be acquired from the human users. Finally, our methodology is validated with one high impact system.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Due to dynamic variability, identifying the specific conditions under which non-functional requirements (NFRs) are satisfied may be only possible at runtime. Therefore, it is necessary to consider the dynamic treatment of relevant information during the requirements specifications. The associated data can be gathered by monitoring the execution of the application and its underlying environment to support reasoning about how the current application configuration is fulfilling the established requirements. This paper presents a dynamic decision-making infrastructure to support both NFRs representation and monitoring, and to reason about the degree of satisfaction of NFRs during runtime. The infrastructure is composed of: (i) an extended feature model aligned with a domain-specific language for representing NFRs to be monitored at runtime; (ii) a monitoring infrastructure to continuously assess NFRs at runtime; and (iii) a exible decision-making process to select the best available configuration based on the satisfaction degree of the NRFs. The evaluation of the approach has shown that it is able to choose application configurations that well fit user NFRs based on runtime information. The evaluation also revealed that the proposed infrastructure provided consistent indicators regarding the best application configurations that fit user NFRs. Finally, a benefit of our approach is that it allows us to quantify the level of satisfaction with respect to NFRs specification.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Bilateral corneal blindness represents a quarter of the total blind, world-wide. The artificial cornea in assorted forms, was developed to replace opaque non-functional corneas and to return sight in otherwise hopeless cases that were not amenable to corneal grafts; believed to be 2% of corneal blind. Despite technological advances in materials design and tissue engineering no artificial cornea has provided absolute, long-term success. Formidable problems exist, due to a combination of unpredictable wound healing and unmanageable pathology. To have a solid guarantee of reliable success an artificial cornea must possess three attributes: an optical window to replace the opaque cornea; a strong, long term union to surrounding ocular tissue; and the ability to induce desired host responses. A unique artificial cornea possesses all three functional attributes- the Osteo-odonto-keratoprosthesis (OOKP). The OOKP has a high success rate and can survive for up to twenty years, but it is complicated both in structure and in surgical procedure; it is expensive and not universally available. The aim of this project was to develop a synthetic substitute for the OOKP, based upon key features of the tooth and bone structure. In doing so, surgical complexity and biological complications would be reduced. Analysis of the biological effectiveness of the OOKP showed that the structure of bone was the most crucial component for implant retention. An experimental semi-rigid hydroxyapatite framework was fabricated with a complex bone-like architecture, which could be fused to the optical window. The first method for making such a framework, was pressing and sintering of hydroxyapatite powders; however, it was not possible to fabricate a void architecture with the correct sizes and uniformity of pores. Ceramers were synthesised using alternative pore forming methods, providing for improved mechanical properties and stronger attachment to the plastic optical window. Naturally occurring skeletal structures closely match the structural features of all forms of natural bone. Synthetic casts were fabricated using the replamineform process, of desirable natural artifacts, such as coral and sponges. The final method of construction by-passed ceramic fabrication in favour of pre-formed coral derivatives and focused on methods for polymer infiltration, adhesion and fabrication. Prototypes were constructed and evaluated; a fully penetrative synthetic OOKP analogue was fabricated according to the dimensions of the OOKP. Fabrication of the cornea shaped OOKP synthetic analogue was also attempted.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Modelling architectural information is particularly important because of the acknowledged crucial role of software architecture in raising the level of abstraction during development. In the MDE area, the level of abstraction of models has frequently been related to low-level design concepts. However, model-driven techniques can be further exploited to model software artefacts that take into account the architecture of the system and its changes according to variations of the environment. In this paper, we propose model-driven techniques and dynamic variability as concepts useful for modelling the dynamic fluctuation of the environment and its impact on the architecture. Using the mappings from the models to implementation, generative techniques allow the (semi) automatic generation of artefacts making the process more efficient and promoting software reuse. The automatic generation of configurations and reconfigurations from models provides the basis for safer execution. The architectural perspective offered by the models shift focus away from implementation details to the whole view of the system and its runtime change promoting high-level analysis. © 2009 Springer Berlin Heidelberg.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Traditionally, geostatistical algorithms are contained within specialist GIS and spatial statistics software. Such packages are often expensive, with relatively complex user interfaces and steep learning curves, and cannot be easily integrated into more complex process chains. In contrast, Service Oriented Architectures (SOAs) promote interoperability and loose coupling within distributed systems, typically using XML (eXtensible Markup Language) and Web services. Web services provide a mechanism for a user to discover and consume a particular process, often as part of a larger process chain, with minimal knowledge of how it works. Wrapping current geostatistical algorithms with a Web service layer would thus increase their accessibility, but raises several complex issues. This paper discusses a solution to providing interoperable, automatic geostatistical processing through the use of Web services, developed in the INTAMAP project (INTeroperability and Automated MAPping). The project builds upon Open Geospatial Consortium standards for describing observations, typically used within sensor webs, and employs Geography Markup Language (GML) to describe the spatial aspect of the problem domain. Thus the interpolation service is extremely flexible, being able to support a range of observation types, and can cope with issues such as change of support and differing error characteristics of sensors (by utilising descriptions of the observation process provided by SensorML). XML is accepted as the de facto standard for describing Web services, due to its expressive capabilities which allow automatic discovery and consumption by ‘naive’ users. Any XML schema employed must therefore be capable of describing every aspect of a service and its processes. However, no schema currently exists that can define the complex uncertainties and modelling choices that are often present within geostatistical analysis. We show a solution to this problem, developing a family of XML schemata to enable the description of a full range of uncertainty types. These types will range from simple statistics, such as the kriging mean and variances, through to a range of probability distributions and non-parametric models, such as realisations from a conditional simulation. By employing these schemata within a Web Processing Service (WPS) we show a prototype moving towards a truly interoperable geostatistical software architecture.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Designers of self-adaptive systems often formulate adaptive design decisions, making unrealistic or myopic assumptions about the system's requirements and environment. The decisions taken during this formulation are crucial for satisfying requirements. In environments which are characterized by uncertainty and dynamism, deviation from these assumptions is the norm and may trigger 'surprises'. Our method allows designers to make explicit links between the possible emergence of surprises, risks and design trade-offs. The method can be used to explore the design decisions for self-adaptive systems and choose among decisions that better fulfil (or rather partially fulfil) non-functional requirements and address their trade-offs. The analysis can also provide designers with valuable input for refining the adaptation decisions to balance, for example, resilience (i.e. Satisfiability of non-functional requirements and their trade-offs) and stability (i.e. Minimizing the frequency of adaptation). The objective is to provide designers of self adaptive systems with a basis for multi-dimensional what-if analysis to revise and improve the understanding of the environment and its effect on non-functional requirements and thereafter decision-making. We have applied the method to a wireless sensor network for flood prediction. The application shows that the method gives rise to questions that were not explicitly asked before at design-time and assists designers in the process of risk-aware, what-if and trade-off analysis.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Contemporary software systems are becoming increasingly large, heterogeneous, and decentralised. They operate in dynamic environments and their architectures exhibit complex trade-offs across dimensions of goals, time, and interaction, which emerges internally from the systems and externally from their environment. This gives rise to the vision of self-aware architecture, where design decisions and execution strategies for these concerns are dynamically analysed and seamlessly managed at run-time. Drawing on the concept of self-awareness from psychology, this paper extends the foundation of software architecture styles for self-adaptive systems to arrive at a new principled approach for architecting self-aware systems. We demonstrate the added value and applicability of the approach in the context of service provisioning to cloud-reliant service-based applications.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Context/Motivation - Different modeling techniques have been used to model requirements and decision-making of self-adaptive systems (SASs). Specifically, goal models have been prolific in supporting decision-making depending on partial and total fulfilment of functional (goals) and non-functional requirements (softgoals). Different goalrealization strategies can have different effects on softgoals which are specified with weighted contribution-links. The final decision about what strategy to use is based, among other reasons, on a utility function that takes into account the weighted sum of the different effects on softgoals. Questions/Problems - One of the main challenges about decisionmaking in self-adaptive systems is to deal with uncertainty during runtime. New techniques are needed to systematically revise the current model when empirical evidence becomes available from the deployment. Principal ideas/results - In this paper we enrich the decision-making supported by goal models by using Dynamic Decision Networks (DDNs). Goal realization strategies and their impact on softgoals have a correspondence with decision alternatives and conditional probabilities and expected utilities in the DDNs respectively. Our novel approach allows the specification of preferences over the softgoals and supports reasoning about partial satisfaction of softgoals using probabilities. We report results of the application of the approach on two different cases. Our early results suggest the decision-making process of SASs can be improved by using DDNs. © 2013 Springer-Verlag.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Computational reflection is a well-established technique that gives a program the ability to dynamically observe and possibly modify its behaviour. To date, however, reflection is mainly applied either to the software architecture or its implementation. We know of no approach that fully supports requirements reflection- that is, making requirements available as runtime objects. Although there is a body of literature on requirements monitoring, such work typically generates runtime artefacts from requirements and so the requirements themselves are not directly accessible at runtime. In this paper, we define requirements reflection and a set of research challenges. Requirements reflection is important because software systems of the future will be self-managing and will need to adapt continuously to changing environmental conditions. We argue requirements reflection can support such self-adaptive systems by making requirements first-class runtime entities, thus endowing software systems with the ability to reason about, understand, explain and modify requirements at runtime. © 2010 ACM.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Purpose: Polyvinyl alcohol (PVA) is a successful tear film stabiliser and is widely used in comfort drops and some soft contact lens materials. A PVA-containing lens, nelfilcon A has been modified to include additional (non-functional) PVA in order to provide improved comfort. This study aims to examine the clinical performance of this nelfilcon A lens with AquaRelease™ (AquaRelease). Methods: Two contralateral, investigator masked, open label, subjective and objective evaluations were conducted. The first examined the effect of adding increased molecular weight PVA to nelfilcon A (n = 5), and the second compared this AquaRelease lens to ocufilcon B (n = 34). The principal measures were non-invasive break-up time (NIBUT) and subjective comfort, which were assessed at the beginning and end of a week of daily wear, and three times throughout 1 day at 8, 12 and 16 h. Results: All subjects successfully completed the daily wearing schedule of 16 h. On initial insertion, subjective comfort and NIBUT improved for AquaRelease than original nelfilcon A lenses (p < 0.05). Initial comfort was better for AquaRelease compared to ocufilcon B lenses (p = 0.01); however, NIBUT was not statistically different (11.7 ± 15.6 s versus 8.4 ± 6.8 s; p = 0.26). Subjective comfort decreased with time (p < 0.001), but there was no significant difference between AquaRelease and ocufilcon B lenses (p = 0.16). NIBUT was not significantly affected by time (p = 0.56) or between lenses (p = 0.33). At the end of a weeks' wear, subjective initial, end-of-day, overall comfort and vision were rated significantly better with AquaRelease than ocufilcon B (p < 0.01). Conclusions: Release of additional non-functionalised PVA from the nelfilcon A lenses appears to enhance comfortable contact lens wear. © 2006 British Contact Lens Association.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The Act that established the Greater London Authority (GLA) incorporated many of New Labour's aspirations for modern governance. Among those aspirations was the notion of policy integration, or 'joining up'. The Mayor of Greater London was required to develop a number of strategies, broadly in the planning and environmental policy domains, and to ensure that those strategies meshed into a coherent overall strategy for promoting London's economic, social and environmental well-being. How would this work in practice, given the need for coordination between the GLA and a number of related functional bodies, and given the political imperative for the GLA to make an impact quickly? Through our analysis of the strategy development and integration efforts of the GLA in its first nine months, we have gleaned new insights into the highly complex and difficult process of policy integration. We argue that the high aspirations of the Act for policy integration have not been met, policy integration instead being narrowly interpreted as the coordination of strategies to the Mayor's political agenda. Finally,we reflect on the likelihood of the GLA, as currently constituted, evolving to meet the functional requirement of policy integration.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

This paper describes how dimensional variation management could be integrated throughout design, manufacture and verification, to improve quality while reducing cycle times and manufacturing cost in the Digital Factory environment. Initially variation analysis is used to optimize tolerances during product and tooling design and also results in the creation of a simplified representation of product key characteristics. This simplified representation can then be used to carry out measurability analysis and process simulation. The link established between the variation analysis model and measurement processes can subsequently be used throughout the production process to automatically update the variation analysis model in real time with measurement data. This ‘live’ simulation of variation during manufacture will allow early detection of quality issues and facilitate autonomous measurement assisted processes such as predictive shimming. A study is described showing how these principles can be demonstrated using commercially available software combined with a number of prototype applications operating as discrete modules. The commercially available modules include Catia/Delmia for product and process design, 3DCS for variation analysis and Spatial Analyzer for measurement simulation. Prototype modules are used to carry out measurability analysis and instrument selection. Realizing the full potential of Metrology in the Digital Factory will require that these modules are integrated and software architecture to facilitate this is described. Crucially this integration must facilitate the use of realtime metrology data describing the emerging assembly to update the digital model.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

The process framework comprises three phases, as follows: scope the supply chain/network; identify the options for supply system architecture and select supply system architecture. It facilitates a structured approach that analyses the supply chain/network contextual characteristics, in order to ensure alignment with the appropriate supply system architecture. The process framework was derived from comprehensive literature review and archival case study analysis. The review led to the classification of supply system architectures according to their orientation, whether integrated; partially integrated; co-ordinated or independent. The classification was combined with the characteristics that influence the selection of supply system architecture to encapsulate the conceptual framework. It builds upon existing frameworks and methodologies by focusing on structured procedure; supporting project management; facilitating participation and clarifying point of entry. The process framework was initially tested in three case study applications from the food, automobile and hand tool industries. A variety of industrial settings was chosen to illustrate transferability. The case study applications indicate that the process framework is a valid approach to the problem; however, further testing is required. In particular, the use of group support system technologies to support the process and the steps involving the participation of software vendors need further testing. However, the process framework can be followed due to the clarity of its presentation. It considers the issue of timing by including alternative decision-making techniques, dependent on the constraints. It is useful for ensuring a sound business case is developed, with supporting documentation and analysis that identifies the strategic and functional requirements of supply system architecture.