314 resultados para Computational architecture
Resumo:
Abstract—Computational Intelligence Systems (CIS) is one of advanced softwares. CIS has been important position for solving single-objective / reverse / inverse and multi-objective design problems in engineering. The paper hybridise a CIS for optimisation with the concept of Nash-Equilibrium as an optimisation pre-conditioner to accelerate the optimisation process. The hybridised CIS (Hybrid Intelligence System) coupled to the Finite Element Analysis (FEA) tool and one type of Computer Aided Design(CAD) system; GiD is applied to solve an inverse engineering design problem; reconstruction of High Lift Systems (HLS). Numerical results obtained by the hybridised CIS are compared to the results obtained by the original CIS. The benefits of using the concept of Nash-Equilibrium are clearly demonstrated in terms of solution accuracy and optimisation efficiency.
Resumo:
Notwithstanding the obvious potential advantages of information and communications technology (ICT) in the enhanced provision of healthcare services, there are some concerns associated with integration of and access to electronic health records. A security violation in health records, such as an unauthorised disclosure or unauthorised alteration of an individual's health information, can significantly undermine both healthcare providers' and consumers' confidence and trust in e-health systems. A crisis in confidence in any national level e-health system could seriously degrade the realisation of the system's potential benefits. In response to the privacy and security requirements for the protection of health information, this research project investigated national and international e-health development activities to identify the necessary requirements for the creation of a trusted health information system architecture consistent with legislative and regulatory requirements and relevant health informatics standards. The research examined the appropriateness and sustainability of the current approaches for the protection of health information. It then proposed an architecture to facilitate the viable and sustainable enforcement of privacy and security in health information systems under the project title "Open and Trusted Health Information Systems (OTHIS)". OTHIS addresses necessary security controls to protect sensitive health information when such data is at rest, during processing and in transit with three separate and achievable security function-based concepts and modules: a) Health Informatics Application Security (HIAS); b) Health Informatics Access Control (HIAC); and c) Health Informatics Network Security (HINS). The outcome of this research is a roadmap for a viable and sustainable architecture for providing robust protection and security of health information including elucidations of three achievable security control subsystem requirements within the proposed architecture. The successful completion of two proof-of-concept prototypes demonstrated the comprehensibility, feasibility and practicality of the HIAC and HIAS models for the development and assessment of trusted health systems. Meanwhile, the OTHIS architecture has provided guidance for technical and security design appropriate to the development and implementation of trusted health information systems whilst simultaneously offering guidance for ongoing research projects. The socio-economic implications of this research can be summarised in the fact that this research embraces the need for low cost security strategies against economic realities by using open-source technologies for overall test implementation. This allows the proposed architecture to be publicly accessible, providing a platform for interoperability to meet real-world application security demands. On the whole, the OTHIS architecture sets a high level of security standard for the establishment and maintenance of both current and future health information systems. This thereby increases healthcare providers‘ and consumers‘ trust in the adoption of electronic health records to realise the associated benefits.
Resumo:
Objective To determine the test-retest reliability of measurements of thickness, fascicle length (Lf) and pennation angle (θ) of the vastus lateralis (VL) and gastrocnemius medialis (GM) muscles in older adults. Participants Twenty-one healthy older adults (11 men and ten women; average age 68·1 ± 5·2 years) participated in this study. Methods Ultrasound images (probe frequency 10 MHz) of the VL at two sites (VL site 1 and 2) were obtained with participants seated with knee at 90º flexion. For GM measures, participants lay prone with ankle fixed at 15º dorsiflexion. Measures were taken on two separate occasions, 7 days apart (T1 and T2). Results The ICCs (95% CI) were: VL site 1 thickness = 0·96(0·90–0·98); VL site 2 thickness = 0·96(0·90–0·98), VL θ = 0·87(0·68–0·95), VL Lf = 0·80(0·50–0·92), GM thickness = 0·97(0·92–0·99), GM θ = 0·85(0·62–0·94) and GM Lf =0·90(0·75–0·96). The 95% ratio limits of agreement (LOAs) for all measures, calculated by multiplying the standard deviation of the ratio of the results between T1 and T2 by 1·96, ranged from 10·59 to 38·01%. Conclusion The ability of these tests to determine a real change in VL and GM muscle architecture is good on a group level but problematic on an individual level as the relatively large 95% ratio LOAs in the current study may encompass the changes in architecture observed in other training studies. Therefore, the current findings suggest that B-mode ultrasonography can be used with confidence by researchers when investigating changes in muscle architecture in groups of older adults, but its use is limited in showing changes in individuals over time.
An experimental and computational investigation of performance of Green Gully for reusing stormwater
Resumo:
A new stormwater quality improvement device (SQID) called ‘Green Gully’ has been designed and developed in this study with an aim to re-using stormwater for irrigating plants and trees. The main purpose of the Green Gully is to collect road runoff/stormwater, make it suitable for irrigation and provide an automated network system for watering roadside plants and irrigational areas. This paper presents the design and development of Green Gully along with experimental and computational investigations of the performance of Green Gully. Performance (in the form of efficiency, i.e. the percentage of water flow through the gully grate) was experimentally determined using a gully model in the laboratory first, then a three dimensional numerical model was developed and simulated to predict the efficiency of Green Gully as a function of flow rate. Computational Fluid Dynamics (CFD) code FLUENT was used for the simulation. GAMBIT was used for geometry creation and mesh generation. Experimental and simulation results are discussed and compared in this paper. The predicted efficiency was compared with the laboratory measured efficiency. It was found that the simulated results are in good agreement with the experimental results.
Resumo:
In recent years, enterprise architecture (EA) has captured a growing attention as a means to systematically consolidate and interrelate diverse business and IT artefacts in order to provide holistic decision support. The recent popularity of a service-orientation has added “service “and related constructs as a new element that requires consideration within an Enterprise Architecture. Since the emergence of the Service-Oriented Architecture (SOA), many attempts have been made to incorporate SOA artefacts in existing EA frameworks. Yet, the approaches taken to achieve this goal differ substantially for the most commonly used EA frameworks to date. SOA in the context of enterprise architecture is one of the future research challenges. Several authors argue that further research is needed in order to understand how SOA impacts prior enterprise architecture frameworks. This study explores SOA integration within EA, identifies SOA integration approaches within EA and identifies factors that impact SOA integration within Enterprise Architecture.
Resumo:
Building Information Modeling (BIM) is a modern approach to the design, documentation, delivery, and life cycle management of buildings through the use of project information databases coupled with object-based parametric modeling. BIM has the potential to revolutionize the Architecture, Engineering and Construction (AEC) industry in terms of the positive impact it may have on information flows, working relationships between project participants from different disciplines and the resulting benefits it may achieve through improvements to conventional methods. This chapter reviews the development of BIM, the extent to which BIM has been implemented in Australia, and the factors which have affected the up-take of BIM. More specifically, the objectives of this chapter are to investigate the adoption of BIM in the Australian AEC industry and factors that contribute towards the uptake (or non uptake) of BIM. These objectives are met by a review of the related literature in the first instance, followed by the presentation of the results of a 2007 postal questionnaire survey and telephone interviews of a random sample of professionals in the Australian AEC industry. The responses suggest that less than 25 percent of the sample had been involved in BIM – rather less than might be expected from reading the literature. Also, of those who have been involved with BIM, there has been very little interdisciplinary collaboration. The main barriers impeding the implementation of BIM widely across the Australian AEC industry are also identified. These were found to be primarily a lack of BIM expertise, lack of awareness and resistance to change. The benefits experienced as a result of using BIM are also discussed. These include improved design consistency, better coordination, cost savings, higher quality work, greater productivity and increased speed of delivery. In terms of conclusion, some suggestions are made concerning the underlying practical reasons for the slow up-take of BIM and the successes for those early adopters. Prospects for future improvement are discussed and proposals are also made for a large scale worldwide comparative study covering industry-wide participants
Resumo:
Virtual environments can provide, through digital games and online social interfaces, extremely exciting forms of interactive entertainment. Because of their capability in displaying and manipulating information in natural and intuitive ways, such environments have found extensive applications in decision support, education and training in the health and science domains amongst others. Currently, the burden of validating both the interactive functionality and visual consistency of a virtual environment content is entirely carried out by developers and play-testers. While considerable research has been conducted in assisting the design of virtual world content and mechanics, to date, only limited contributions have been made regarding the automatic testing of the underpinning graphics software and hardware. The aim of this thesis is to determine whether the correctness of the images generated by a virtual environment can be quantitatively defined, and automatically measured, in order to facilitate the validation of the content. In an attempt to provide an environment-independent definition of visual consistency, a number of classification approaches were developed. First, a novel model-based object description was proposed in order to enable reasoning about the color and geometry change of virtual entities during a play-session. From such an analysis, two view-based connectionist approaches were developed to map from geometry and color spaces to a single, environment-independent, geometric transformation space; we used such a mapping to predict the correct visualization of the scene. Finally, an appearance-based aliasing detector was developed to show how incorrectness too, can be quantified for debugging purposes. Since computer games heavily rely on the use of highly complex and interactive virtual worlds, they provide an excellent test bed against which to develop, calibrate and validate our techniques. Experiments were conducted on a game engine and other virtual worlds prototypes to determine the applicability and effectiveness of our algorithms. The results show that quantifying visual correctness in virtual scenes is a feasible enterprise, and that effective automatic bug detection can be performed through the techniques we have developed. We expect these techniques to find application in large 3D games and virtual world studios that require a scalable solution to testing their virtual world software and digital content.
Resumo:
We report on analysis of discussions in an online community of people with chronic illness using socio-cognitively motivated, automatically produced semantic spaces. The analysis aims to further the emerging theory of "transition" (how people can learn to incorporate the consequences of illness into their lives). An automatically derived representation of sense of self for individuals is created in the semantic space by the analysis of the email utterances of the community members. The movement over time of the sense of self is visualised, via projection, with respect to axes of "ordinariness" and "extra-ordinariness". Qualitative evaluation shows that the visualisation is paralleled by the transitions of people during the course of their illness. The research aims to progress tools for analysis of textual data to promote greater use of tacit knowledge as found in online virtual communities. We hope it also encourages further interest in representation of sense-of-self.
Resumo:
The in vitro and in vivo degradation properties of poly(lactic-co-glycolic acid) (PLGA) scaffolds produced by two different technologies - thermally induced phase separation (TIPS), and solvent casting and particulate leaching (SCPL) were compared. Over 6 weeks, in vitro degradation produced changes in SCPL scaffold dimension, mass, internal architecture and mechanical properties. TIPS scaffolds produced far less changes in these parameters providing significant advantages over SCPL. In vivo results were based on a microsurgically created arteriovenous (AV) loop sandwiched between two TIPS scaffolds placed in a polycarbonate chamber under rat groin skin. Histologically, a predominant foreign body giant cell response and reduced vascularity was evident in tissue ingrowth between 2 and 8 weeks in TIPS scaffolds. Tissue death occurred at 8 weeks in the smallest pores. Morphometric comparison of TIPS and SCPL scaffolds indicated slightly better tissue ingrowth but greater loss of scaffold structure in SCPL scaffolds. Although advantageous in vitro, large surface area:volume ratios and varying pore sizes in PLGA TIPS scaffolds mean that effective in vivo (AV loop) utilization will only be achieved if the foreign body response can be significantly reduced so as to allow successful vascularisation, and hence sustained tissue growth, in pores less than 300 μm. © 2005 Elsevier Ltd. All rights reserved.
Resumo:
This project investigates machine listening and improvisation in interactive music systems with the goal of improvising musically appropriate accompaniment to an audio stream in real-time. The input audio may be from a live musical ensemble, or playback of a recording for use by a DJ. I present a collection of robust techniques for machine listening in the context of Western popular dance music genres, and strategies of improvisation to allow for intuitive and musically salient interaction in live performance. The findings are embodied in a computational agent – the Jambot – capable of real-time musical improvisation in an ensemble setting. Conceptually the agent’s functionality is split into three domains: reception, analysis and generation. The project has resulted in novel techniques for addressing a range of issues in each of these domains. In the reception domain I present a novel suite of onset detection algorithms for real-time detection and classification of percussive onsets. This suite achieves reasonable discrimination between the kick, snare and hi-hat attacks of a standard drum-kit, with sufficiently low-latency to allow perceptually simultaneous triggering of accompaniment notes. The onset detection algorithms are designed to operate in the context of complex polyphonic audio. In the analysis domain I present novel beat-tracking and metre-induction algorithms that operate in real-time and are responsive to change in a live setting. I also present a novel analytic model of rhythm, based on musically salient features. This model informs the generation process, affording intuitive parametric control and allowing for the creation of a broad range of interesting rhythms. In the generation domain I present a novel improvisatory architecture drawing on theories of music perception, which provides a mechanism for the real-time generation of complementary accompaniment in an ensemble setting. All of these innovations have been combined into a computational agent – the Jambot, which is capable of producing improvised percussive musical accompaniment to an audio stream in real-time. I situate the architectural philosophy of the Jambot within contemporary debate regarding the nature of cognition and artificial intelligence, and argue for an approach to algorithmic improvisation that privileges the minimisation of cognitive dissonance in human-computer interaction. This thesis contains extensive written discussions of the Jambot and its component algorithms, along with some comparative analyses of aspects of its operation and aesthetic evaluations of its output. The accompanying CD contains the Jambot software, along with video documentation of experiments and performances conducted during the project.
Resumo:
The "vernacular" housing tradition of southeast Queensland is easily identifiable. Its history is more complex. This study seeks to challenge two popular conceptions of the "Queenslander" history by showing that they actually provide contradictory explanations. The aim is to produce a more complex account of local architecture and its historical explanation so that both its past and its present practices can be better understood as a distinctly subtropical idiom. This discussion shows that such practices may respond to common concerns but that are also ever-changing.
Resumo:
This thesis investigates the radically uncertain formal, business, and industrial environment of current entertainment creators. It researches how a novel communication technology, the Internet, leads to novel entertainment forms, how these lead to novel kinds of businesses that lead to novel industries; and in what way established entertainment forms, businesses, and industries are part of that process. This last aspect is addressed by focusing on one exemplary es-tablished form: movies. Using a transdisciplinary approach and a combination of historical analysis, industry interviews, and an innovative mode of ‘immersive’ textual analysis, a coherent and comprehensive conceptual framework for the creation of and re-search into a specific emerging entertainment form is proposed. That form, products based on it, and the conceptual framework describing it are all re-ferred to as Entertainment Architecture (‘entarch,’ for short). The thesis charac-terises this novel form as Internet-native transmedia entertainment, meaning it fully utilises the unique communicative characteristics of the Internet, and is spread across media. The thesis isolates four constitutive elements within Entertainment Architec-ture: story, play, ‘dance,’ and ‘glue.’ That is, entarch tells a story; offers playful interaction; invites social interaction between producer and consumer, and amongst consumers (‘dance’); and all components of it can be spread across many media, but are so well interconnected and mutually dependent that they are perceived as one product instead of many (‘glue’). This sets entarch apart from current media franchises like Star Wars or Halo, which are perceived as many products spread across many media. Entarch thus embraces the commu-nicative behaviour of Internet-native consumers instead of forcing them to de-sist from it, it harnesses the strengths of various media while avoiding some of their weaknesses, and it can sustain viable businesses. The entarch framework is an innovative contribution to scholarship that al-lows researchers to investigate this emerging entertainment form in a structured way. The thesis demonstrates this by using it to survey business models appro-priate to the entarch environment. The framework can also be used by enter-tainment creators — exemplified in the thesis by moviemakers — to delimit the room for manoeuvre available to them in a changing environment.
Resumo:
Enterprise architecture management (EAM) has become an intensively discussed approach to manage enterprise transformations. While many organizations employ EAM, a notable insecurity about the value of EAM remains. In this paper, we propose a model to measure the realization of benefits from EAM. We identify EAM success factors and EAM benefits through a comprehensive literature review and eleven explorative expert interviews. Based on our findings, we integrate the EAM success factors and benefits with the established DeLone & McLean IS success model resulting in a model that explains the realization of EAM benefits. This model aids organizations as a benchmark and framework for identifying and assessing the setup of their EAM initiatives and whether and how EAM benefits are materialized. We see our model also as a first step to gain insights in and start a discussion on the theory of EAM benefit realization.
Resumo:
Enterprise architecture (EA) management has become an intensively discussed approach to manage enterprise transformations. Despite the popularity and potential of EA, both researchers and practitioners lament a lack of knowledge about the realization of benefits from EA. To determine the benefits from EA, we explore the various dimensions of EA benefit realization and report on the development of a validated and robust measurement instrument. In this paper, we test the reliability and construct validity of the EA benefit realization model (EABRM), which we have designed based on the DeLone & McLean IS success model and findings from exploratory interviews. A confirmatory factor analysis confirms the existence of an impact of five distinct and individually important dimensions on the benefits derived from EA: EA artefact quality, EA infrastructure quality, EA service quality, EA culture, and EA use. The analysis presented in this paper shows that the EA benefit realization model is an instrument that demonstrates strong reliability and validity.