23 resultados para Aboriginal knowledge domain
em Aston University Research Archive
Resumo:
This paper describes the knowledge elicitation and knowledge representation aspects of a system being developed to help with the design and maintenance of relational data bases. The size algorithmic components. In addition, the domain contains multiple experts, but any given expert's knowledge of this large domain is only partial. The paper discusses the methods and techniques used for knowledge elicitation, which was based on a "broad and shallow" approach at first, moving to a "narrow and deep" one later, and describes the models used for knowledge representation, which were based on a layered "generic and variants" approach. © 1995.
Resumo:
Motivation: In molecular biology, molecular events describe observable alterations of biomolecules, such as binding of proteins or RNA production. These events might be responsible for drug reactions or development of certain diseases. As such, biomedical event extraction, the process of automatically detecting description of molecular interactions in research articles, attracted substantial research interest recently. Event trigger identification, detecting the words describing the event types, is a crucial and prerequisite step in the pipeline process of biomedical event extraction. Taking the event types as classes, event trigger identification can be viewed as a classification task. For each word in a sentence, a trained classifier predicts whether the word corresponds to an event type and which event type based on the context features. Therefore, a well-designed feature set with a good level of discrimination and generalization is crucial for the performance of event trigger identification. Results: In this article, we propose a novel framework for event trigger identification. In particular, we learn biomedical domain knowledge from a large text corpus built from Medline and embed it into word features using neural language modeling. The embedded features are then combined with the syntactic and semantic context features using the multiple kernel learning method. The combined feature set is used for training the event trigger classifier. Experimental results on the golden standard corpus show that >2.5% improvement on F-score is achieved by the proposed framework when compared with the state-of-the-art approach, demonstrating the effectiveness of the proposed framework. © 2014 The Author 2014. The source code for the proposed framework is freely available and can be downloaded at http://cse.seu.edu.cn/people/zhoudeyu/ETI_Sourcecode.zip.
Resumo:
Ontologies have become a key component in the Semantic Web and Knowledge management. One accepted goal is to construct ontologies from a domain specific set of texts. An ontology reflects the background knowledge used in writing and reading a text. However, a text is an act of knowledge maintenance, in that it re-enforces the background assumptions, alters links and associations in the ontology, and adds new concepts. This means that background knowledge is rarely expressed in a machine interpretable manner. When it is, it is usually in the conceptual boundaries of the domain, e.g. in textbooks or when ideas are borrowed into other domains. We argue that a partial solution to this lies in searching external resources such as specialized glossaries and the internet. We show that a random selection of concept pairs from the Gene Ontology do not occur in a relevant corpus of texts from the journal Nature. In contrast, a significant proportion can be found on the internet. Thus, we conclude that sources external to the domain corpus are necessary for the automatic construction of ontologies.
Resumo:
Automatic ontology building is a vital issue in many fields where they are currently built manually. This paper presents a user-centred methodology for ontology construction based on the use of Machine Learning and Natural Language Processing. In our approach, the user selects a corpus of texts and sketches a preliminary ontology (or selects an existing one) for a domain with a preliminary vocabulary associated to the elements in the ontology (lexicalisations). Examples of sentences involving such lexicalisation (e.g. ISA relation) in the corpus are automatically retrieved by the system. Retrieved examples are validated by the user and used by an adaptive Information Extraction system to generate patterns that discover other lexicalisations of the same objects in the ontology, possibly identifying new concepts or relations. New instances are added to the existing ontology or used to tune it. This process is repeated until a satisfactory ontology is obtained. The methodology largely automates the ontology construction process and the output is an ontology with an associated trained leaner to be used for further ontology modifications.
Resumo:
A review of the literature suggests that various IS knowledge and skills should be grouped into a construct because many IS tasks require a combination of diverse disciplines. Many studies argue that today's IS executives should be multidisciplinary, while the organizational skills are becoming increasingly important. However, the outstanding issue is what is the appropriate balance of disciplines between the organizational domain and technical domain. For understanding such a balance, it is easier to refer to quantitative proportions than to qualitative descriptions. Therefore, this study suggests a way to quantitatively identify an IS management knowledge and skills construct - a combination that comprises six related knowledge and skills categories suggested by the literature. By analyzing the data obtained in a 1996's survey, two such constructs for current proficiency and expected level have been quantitatively identified. To people aiming at a senior IS executive career, the deficiencies between current and expected constructs suggest not only the direction but also the extent should be enhanced, and the constructs indicate the balance among various disciplines should be maintained.
Resumo:
Problem-structuring techniques are an integral aspect of 'Soft-OR'. SSM, SAST, Strategic Choice, and JOURNEY Making, all depend for their success on a group developing a shared view of a problem through some form of explicit modelling. The negotiated problem structure becomes the basis for problem resolution. Implicit to this process is an assumption that members of the group share and build their knowledge about the problem domain. This paper explores the extent to which this assumption is reasonable. The research is based on detailed records from the use of JOURNEY Making, where it has used special purpose Group Support software to aid the group problem structuring. This software continuously tracks the contributions of each member of the group and thus the extent to which they appear to be 'connecting' and augmenting their own knowledge with that of other members of the group. Software records of problem resolution in real organisational settings are used to explore the sharing of knowledge among senior managers. These explorations suggest a typology of knowledge sharing. The implications of this typology for problem structuring and an agenda for future research are considered.
Resumo:
The International Cooperation Agency (identified in this article as IDEA) working in Colombia is one of the most important in Colombian society with programs that support gender rights, human rights, justice and peace, scholarships, aboriginal population, youth, afro descendants population, economic development in communities, and environmental development. The identified problem is based on the diversified offer of services, collaboration and social intervention which requires diverse groups of people with multiple agendas, ways to support their mandates, disciplines, and professional competences. Knowledge creation and the growth and sustainability of the organization can be in danger because of a silo culture and the resulting reduced leverage of the separate group capabilities. Organizational memory is generally formed by the tacit knowledge of the organization members, given the value of accumulated experience that this kind of social work implies. Its loss is therefore a strategic and operational risk when most problem interventions rely on direct work in the socio-economic field and living real experiences with communities. The knowledge management solution presented in this article starts first, with the identification of the people and groups concerned and the creation of a knowledge map as a means to strengthen the ties between organizational members; second, by introducing a content management system designed to support the documentation process and knowledge sharing process; and third, introducing a methodology for the adaptation of a Balanced Scorecard based on the knowledge management processes. These three main steps lead to a knowledge management “solution” that has been implemented in the organization, comprising three components: a knowledge management system, training support and promotion of cultural change.
Resumo:
In a certain automobile factory, batch-painting of the body types in colours is controlled by an allocation system. This tries to balance production with orders, whilst making optimally-sized batches of colours. Sequences of cars entering painting cannot be optimised for easy selection of colour and batch size. `Over-production' is not allowed, in order to reduce buffer stocks of unsold vehicles. Paint quality is degraded by random effects. This thesis describes a toolkit which supports IKBS in an object-centred formalism. The intended domain of use for the toolkit is flexible manufacturing. A sizeable application program was developed, using the toolkit, to test the validity of the IKBS approach in solving the real manufacturing problem above, for which an existing conventional program was already being used. A detailed statistical analysis of the operating circumstances of the program was made to evaluate the likely need for the more flexible type of program for which the toolkit was intended. The IKBS program captures the many disparate and conflicting constraints in the scheduling knowledge and emulates the behaviour of the program installed in the factory. In the factory system, many possible, newly-discovered, heuristics would be awkward to represent and it would be impossible to make many new extensions. The representation scheme is capable of admitting changes to the knowledge, relying on the inherent encapsulating properties of object-centres programming to protect and isolate data. The object-centred scheme is supported by an enhancement of the `C' programming language and runs under BSD 4.2 UNIX. The structuring technique, using objects, provides a mechanism for separating control of expression of rule-based knowledge from the knowledge itself and allowing explicit `contexts', within which appropriate expression of knowledge can be done. Facilities are provided for acquisition of knowledge in a consistent manner.
Resumo:
Investigation of the different approaches used by Expert Systems researchers to solve problems in the domain of Mechanical Design and Expert Systems was carried out. The techniques used for conventional formal logic programming were compared with those used when applying Expert Systems concepts. A literature survey of design processes was also conducted with a view to adopting a suitable model of the design process. A model, comprising a variation on two established ones, was developed and applied to a problem within what are described as class 3 design tasks. The research explored the application of these concepts to Mechanical Engineering Design problems and their implementation on a microcomputer using an Expert System building tool. It was necessary to explore the use of Expert Systems in this manner so as to bridge the gap between their use as a control structure and for detailed analytical design. The former application is well researched into and this thesis discusses the latter. Some Expert System building tools available to the author at the beginning of his work were evaluated specifically for their suitability for Mechanical Engineering design problems. Microsynics was found to be the most suitable on which to implement a design problem because of its simple but powerful Semantic Net Knowledge Representation structure and the ability to use other types of representation schemes. Two major implementations were carried out. The first involved a design program for a Helical compression spring and the second a gearpair system design. Two concepts were proposed in the thesis for the modelling and implementation of design systems involving many equations. The method proposed enables equation manipulation and analysis using a combination of frames, semantic nets and production rules. The use of semantic nets for purposes other than for psychology and natural language interpretation, is quite new and represents one of the major contributions to knowledge by the author. The development of a purpose built shell program for this type of design problems was recommended as an extension of the research. Microsynics may usefully be used as a platform for this development.
Resumo:
The thesis describes the work carried out to develop a prototype knowledge-based system 'KBS-SETUPP' to generate process plans for the manufacture of seamless tubes. The work is specifically related to a plant in which hollows are made from solid billets using a rotary piercing process and then reduced to required size and finished properties using the fixed plug cold drawing process. The thesis first discusses various methods of tube production in order to give a general background of tube manufacture. Then a review of the automation of the process planning function is presented in terms of its basic sub-tasks and the techniques and suitability of a knowledge-based system is established. In the light of such a review and a case study, the process planning problem is formulated in the domain of seamless tube manufacture, its basic sub-tasks are identified and capabilities and constraints of the available equipment in the specific plant are established. The task of collecting and collating the process planning knowledge in seamless tube manufacture is discussed and is mostly fulfilled from domain experts, analysing of existing manufacturing records specific to plant, textbooks and applicable Standards. For the cold drawing mill, tube-drawing schedules have been rationalised to correspond with practice. The validation of such schedules has been achieved by computing the process parameters and then comparing these with the drawbench capacity to avoid over-loading. The existing models cannot be simulated in the computer program as such, therefore a mathematical model has been proposed which estimates the process parameters which are in close agreement with experimental values established by other researchers. To implement the concepts, a Knowledge-Based System 'KBS- SETUPP' has been developed on Personal Computer using Turbo- Prolog. The system is capable of generating process plans, production schedules and some additional capabilities to supplement process planning. The system generated process plans have been compared with the actual plans of the company and it has been shown that the results are satisfactory and encouraging and that the system has the capabilities which are useful.
Resumo:
This research aims to investigate knowledge acquisition and concept formation in the domain of economics and business studies through a foreign language, English, from the very initial to the very final stage of development in the context of Higher Education in Turkey. It traces both the processes as well as the product of acquisition in order to provide a detailed picture of how knowledge acquisition occurs. It aims to explore ways in which the acquisition process can be facilitated and promoted while prospective students of the Department of Economics and Business Administration receive a language training programme, following the completion of which they will join their academic community which offers part of its courses through the English language. The study draws upon (some) theories of mental representation of knowledge, such as schema, frame and script. The concept of discourse community with its characteristics is investigated, enculturation of prospective students to acquire knowledge of their domain through L2 is explored, and the crucial role of the constructivist theory in relation to knowledge acquisition is highlighted. The present study was conducted through a process of enculturation taking place partly at the language centre of Çukurova University and partly at the target discourse community. The data utilised for initiating knowledge acquisition was obtained by establishing a corpus of economics and business texts, which the learners are expected to read during their academic courses utilising computerised technology. The method of think aloud protocols was used to analyse processes taking place in knowledge acquisition, while the product of what was acquired was investigated by means of written recall protocols. It has been discovered that knowledge acquisition operates on the basis of analogical and to a certain extent metaphorical reasoning. The evidence obtained from the think aloud protocols showed that neophytes were able to acquire fundamental concepts of their future domain by reaching the level of shared understanding with the members of their target community of the faculty. Diaries and questionnaire analyses demonstrated that enculturation facilitated learners' transition from the language centre into the target community. Analyses of the written recall protocols and examinations from the post-enculturation stage of the research showed that neophytes' academic performances in their target community were much higher than those of their non-enculturated counterparts. Processes learners go through and strategies they spontaneously make use of, especially while acquiring knowledge of a specific domain through L2 have so far remained unexplored research areas. The present research makes a potential contribution to the language and knowledge acquisition theories by examining closely and systematically the language and the strategies they employ in acquiring such knowledge. The research findings offer useful implications to English language teaching at language schools. Language teachers are provided with useful guidelines as to how they can provide prospective students of a particular academic community with an experience of acquiring fundamental concepts of their discipline before they become members of their target community.
Resumo:
The work described was carried out as part of a collaborative Alvey software engineering project (project number SE057). The project collaborators were the Inter-Disciplinary Higher Degrees Scheme of the University of Aston in Birmingham, BIS Applied Systems Ltd. (BIS) and the British Steel Corporation. The aim of the project was to investigate the potential application of knowledge-based systems (KBSs) to the design of commercial data processing (DP) systems. The work was primarily concerned with BIS's Structured Systems Design (SSD) methodology for DP systems development and how users of this methodology could be supported using KBS tools. The problems encountered by users of SSD are discussed and potential forms of computer-based support for inexpert designers are identified. The architecture for a support environment for SSD is proposed based on the integration of KBS and non-KBS tools for individual design tasks within SSD - The Intellipse system. The Intellipse system has two modes of operation - Advisor and Designer. The design, implementation and user-evaluation of Advisor are discussed. The results of a Designer feasibility study, the aim of which was to analyse major design tasks in SSD to assess their suitability for KBS support, are reported. The potential role of KBS tools in the domain of database design is discussed. The project involved extensive knowledge engineering sessions with expert DP systems designers. Some practical lessons in relation to KBS development are derived from this experience. The nature of the expertise possessed by expert designers is discussed. The need for operational KBSs to be built to the same standards as other commercial and industrial software is identified. A comparison between current KBS and conventional DP systems development is made. On the basis of this analysis, a structured development method for KBSs in proposed - the POLITE model. Some initial results of applying this method to KBS development are discussed. Several areas for further research and development are identified.
Resumo:
Mental-health risk assessment practice in the UK is mainly paper-based, with little standardisation in the tools that are used across the Services. The tools that are available tend to rely on minimal sets of items and unsophisticated scoring methods to identify at-risk individuals. This means the reasoning by which an outcome has been determined remains uncertain. Consequently, there is little provision for: including the patient as an active party in the assessment process, identifying underlying causes of risk, and eecting shared decision-making. This thesis develops a tool-chain for the formulation and deployment of a computerised clinical decision support system for mental-health risk assessment. The resultant tool, GRiST, will be based on consensual domain expert knowledge that will be validated as part of the research, and will incorporate a proven psychological model of classication for risk computation. GRiST will have an ambitious remit of being a platform that can be used over the Internet, by both the clinician and the layperson, in multiple settings, and in the assessment of patients with varying demographics. Flexibility will therefore be a guiding principle in the development of the platform, to the extent that GRiST will present an assessment environment that is tailored to the circumstances in which it nds itself. XML and XSLT will be the key technologies that help deliver this exibility.
Resumo:
In Fourier domain optical coherence tomography (FD-OCT), a large amount of interference data needs to be resampled from the wavelength domain to the wavenumber domain prior to Fourier transformation. We present an approach to optimize this data processing, using a graphics processing unit (GPU) and parallel processing algorithms. We demonstrate an increased processing and rendering rate over that previously reported by using GPU paged memory to render data in the GPU rather than copying back to the CPU. This avoids unnecessary and slow data transfer, enabling a processing and display rate of well over 524,000 A-scan/s for a single frame. To the best of our knowledge this is the fastest processing demonstrated to date and the first time that FD-OCT processing and rendering has been demonstrated entirely on a GPU.
Resumo:
Automated negotiation is widely applied in various domains. However, the development of such systems is a complex knowledge and software engineering task. So, a methodology there will be helpful. Unfortunately, none of existing methodologies can offer sufficient, detailed support for such system development. To remove this limitation, this paper develops a new methodology made up of: (1) a generic framework (architectural pattern) for the main task, and (2) a library of modular and reusable design pattern (templates) of subtasks. Thus, it is much easier to build a negotiating agent by assembling these standardised components rather than reinventing the wheel each time. Moreover, since these patterns are identified from a wide variety of existing negotiating agents (especially high impact ones), they can also improve the quality of the final systems developed. In addition, our methodology reveals what types of domain knowledge need to be input into the negotiating agents. This in turn provides a basis for developing techniques to acquire the domain knowledge from human users. This is important because negotiation agents act faithfully on the behalf of their human users and thus the relevant domain knowledge must be acquired from the human users. Finally, our methodology is validated with one high impact system.