915 resultados para 2-STATE MARKOV-PROCESSES
Resumo:
Cherbourg State School is some 300 kilometres northwest of Brisbane. It is situated in an Aboriginal community at Cherbourg with approximately 250 students, all of whom are Indigenous Australian children. Cherbourg State School aims to generate good academic outcomes for its students from kindergarten to Year 7 and nurture a strong and positive sense of what it means to be Aboriginal in today's society. In a context where the community continues to grapple with many social issues born of the historical processes of dispossession and disempowerment, Cherbourg State School is determined that its children can and will learn to become 'Strong and Smart'. It is a journey that has been charted by Chris Sarra, the school's first Aboriginal principal, in his paper Young and Black and Deadly: Strategies for Improving Outcomes for Indigenous Students, which describes how pride and expectations were engendered in the school over a four-year period from 1998. In this article the author discusses the historical context of the school and its impact on the Indigenous people of Cherbourg. The aim is to consider the historical, political, social and cultural context around the creation of Cherbourg State School. The author critically examines the historical records of the role of the State Government and the white settlers in the setting up and creation of the Aboriginal Reserve and later the primary school. Throughout the author addresses an absence � a voice missing from history � the voice of the Aboriginal people. This exercise in collective memory was designed to provide an opportunity for those who have seldom been given the opportunity to tell their story. Instead of the official view of Cherbourg School it provides a narrative which restores the victims of history to a place of dignity and indeed humanity.
Resumo:
This paper proposes a method which aims at increasing the efficiency of enterprise system implementations. First, we argue that existing process modeling languages that feature different degrees of abstraction for different user groups exist and are used for different purposes which makes it necessary to integrate them. We describe how to do this using the meta models of the involved languages. Second, we argue that an integrated process model based on the integrated meta model needs to be configurable and elaborate on the enabling mechanisms. We introduce a business example using SAP modeling techniques to illustrate the proposed method.
Resumo:
In this paper, technology is described as involving processes whereby resources are utilised to satisfy human needs or to take advantage of opportunities, to develop practical solutions to problems. This study, set within one type of technology context, information technology, investigated how, through a one semester undergraduate university course, elements of technological processes were made explicit to students. While it was acknowledged in the development and implementation of this course that students needed to learn technical skills, technological skills and knowledge, including design, were seen as vital also, to enable students to think about information technology from a perspective that was not confined and limited to `technology as hardware and software'. This paper describes how the course, set within a three year program of study, was aimed at helping students to develop their thinking and their knowledge about design processes in an explicit way. An interpretive research approach was used and data sources included a repertory grid `survey'; student interviews; video recordings of classroom interactions, audio recordings of lectures, observations of classroom interactions made by researchers; and artefacts which included students' journals and portfolios. The development of students' knowledge about design practices is discussed and reflections upon student knowledge development in conjunction with their learning experiences are made. Implications for ensuring explicitness of design practice within information technology contexts are presented, and the need to identify what constitutes design knowledge is argued.
Resumo:
Mainstream business process modelling techniques promote a design paradigm wherein the activities to be performed within a case, together with their usual execution order, form the backbone of a process model, on top of which other aspects are anchored. This paradigm, while eective in standardised and production-oriented domains, shows some limitations when confronted with processes where case-by-case variations and exceptions are the norm. In this thesis we develop the idea that the eective design of exible process models calls for an alternative modelling paradigm, one in which process models are modularised along key business objects, rather than along activity decompositions. The research follows a design science method, starting from the formulation of a research problem expressed in terms of requirements, and culminating in a set of artifacts that have been devised to satisfy these requirements. The main contributions of the thesis are: (i) a meta-model for object-centric process modelling incorporating constructs for capturing exible processes; (ii) a transformation from this meta-model to an existing activity-centric process modelling language, namely YAWL, showing the relation between object-centric and activity-centric process modelling approaches; and (iii) a Coloured Petri Net that captures the semantics of the proposed meta-model. The meta-model has been evaluated using a framework consisting of a set of work ow patterns. Moreover, the meta-model has been embodied in a modelling tool that has been used to capture two industrial scenarios.
Resumo:
This thesis describes outcomes of a research study conducted to investigate the nutrient build-up and wash-off processes on urban impervious surfaces. The data needed for the study was generated through a series of field investigations and laboratory test procedures. The study sites were selected in urbanised catchments to represent typical characteristics of residential, industrial and commercial land uses. The build-up and wash-off samples were collected from road surfaces in the selected study sites. A specially designed vacuum collection system and a rainfall simulator were used for sample collection. According to the data analysis, the solids build-up on road surfaces was significantly finer with more than 80% of the particles below 150 ìm for all the land uses. Nutrients were mostly associated with the particle size range below 150 ìm in both build-up and wash-off samples irrespective of type of land use. Therefore, the finer fraction of solids was the most important for the nutrient build-up and particulate nutrient wash-off processes. Consequently, the design of stormwater quality mitigation measures should target particles less than 150 ìm for the removal of nutrients irrespective of type of land use. Total kjeldahl nitrogen (TKN) was the most dominant form of nitrogen species in build-up on road surfaces. Phosphorus build-up on road surfaces was mainly in inorganic form and phosphate (PO4 3-) was the most dominant form. The nutrient wash-off process was found to be dependent on rainfall intensity and duration. Concentration of both total nitrogen and phosphorus was higher at the beginning of the rain event and decreased with the increase in rainfall duration. Consequently, in the design of stormwater quality mitigation strategies for nutrients removal, it is important to target the initial period of rain events. The variability of wash-off of nitrogen with rainfall intensity was significantly different to phosphorus wash-off. The concentration of nitrogen was higher in the wash-off for low intensity rain events compared to the wash-off for high intensity rain events. On the other hand, the concentration of phosphorus in the wash-off was high for high intensity rain events compared to low intensity rain events. Consequently, the nitrogen washoff can be defined as a source limiting process and phosphorus wash-off as a transport limiting process. This highlights the importance of taking into consideration the wash-off of low intensity rain events in the design of stormwater quality mitigation strategies targeting the nitrogen removal. All the nitrogen species in wash-off are primarily in dissolved form whereas phosphorus is in particulate form. The differences in the nitrogen and phosphorus wash-off processes is principally due to the degree of solubility, attachment to particulates, composition of total nitrogen and total phosphorus and the degree of adherence of the solids particles to the surface to which nutrients are attached. The particulate nitrogen available for wash-off is removed readily as these are mobilised as free solids particles on the surface. Phosphorus is washed-off mostly with the solids particles which are strongly adhered to the surface or as the fixed solids load. Investigation of the nitrogen wash-off process using bulk wash-off samples was in close agreement with the investigation of dissolved fraction of wash-off solids. This was primarily due to the predominant nature of dissolved nitrogen. However, the investigation of the processes which underpin phosphorus wash-off using bulk washoff samples could lead to loss of information. This is due to the composition of total phosphorus in wash-off solids and the inherent variability of the wash-off process for the different particle size ranges. This variability should preferably be taken into consideration as phosphorus wash-off is predominantly in particulate form. Therefore, care needs to be taken in the investigation of the phosphorus wash-off process using bulk wash-off samples to ensure that there is no loss of information and hence result in misleading outcomes. The investigation of different particle size ranges of wash-off solids is preferable in the interest of designing effective stormwater quality management strategies targeting phosphorus removal.
Resumo:
Aims : The aim of this study was to conduct an exploratory investigation into the in-session processes and behaviours that occur between therapists and young people in online counseling. Method: The Consensual Qualitative Research method was employed to identify in-session behaviours and a coding instrument was developed to determine their frequency of use and assess whether nuances carried in the meaning of text messages have an influential effect during sessions. Eighty-five single-session transcripts were examined in total by two independent coders. Results: Sample statistics revealed that, on average, rapport-building processes were used more consistently across cases with both types of processes having a moderately strong positive effect on young people. However, closer examination of these processes revealed weaker positive effects for in-session behaviours that rely more heavily on verbal and non-verbal cues to be accurately interpreted. Implications for Practice and Future Research: These findings imply that therapists may focus more on building rapport than accomplishing tasks with young people during online counselling sessions due to the absence of verbal and non-verbal information when communicating via text messages.
Resumo:
In this paper, the problems of three carrier phase ambiguity resolution (TCAR) and position estimation (PE) are generalized as real time GNSS data processing problems for a continuously observing network on large scale. In order to describe these problems, a general linear equation system is presented to uniform various geometry-free, geometry-based and geometry-constrained TCAR models, along with state transition questions between observation times. With this general formulation, generalized TCAR solutions are given to cover different real time GNSS data processing scenarios, and various simplified integer solutions, such as geometry-free rounding and geometry-based LAMBDA solutions with single and multiple-epoch measurements. In fact, various ambiguity resolution (AR) solutions differ in the floating ambiguity estimation and integer ambiguity search processes, but their theoretical equivalence remains under the same observational systems models and statistical assumptions. TCAR performance benefits as outlined from the data analyses in some recent literatures are reviewed, showing profound implications for the future GNSS development from both technology and application perspectives.
Resumo:
Adiabatic compression testing of components in gaseous oxygen is a test method that is utilized worldwide and is commonly required to qualify a component for ignition tolerance under its intended service. This testing is required by many industry standards organizations and government agencies. This paper traces the background of adiabatic compression testing in the oxygen community and discusses the thermodynamic and fluid dynamic processes that occur during rapid pressure surges. This paper is the first of several papers by the authors on the subject of adiabatic compression testing and is presented as a non-comprehensive background and introduction.
Resumo:
This paper examines accountability in the context of New Zealand's state-owned enterprises (SOEs), and considers an accountability framework which extends beyond traditional (upward and outward) dimensions. While traditional accountability dimensions are clearly detailed in the legislation, less clear are the dilemmas and contradictions faced by SOEs, balancing multiple interests, and at times, conflicting objectives. These issues are explored through interviews conducted in two phases over a two year period with senior executives from 12 of the then 17 SOEs operating in New Zealand. Findings reveal a number of dimensions of accountability that extend beyond current conceptualisations of accountability in the public sector, often including multiple directions of accountability. Implications suggest the need for a more explicit framework of accountability relevant to the SOE context, which may serve useful in identifying potential pathways to conflict mitigation.
Resumo:
In this work, we investigate an alternative bootstrap approach based on a result of Ramsey [F.L. Ramsey, Characterization of the partial autocorrelation function, Ann. Statist. 2 (1974), pp. 1296-1301] and on the Durbin-Levinson algorithm to obtain a surrogate series from linear Gaussian processes with long range dependence. We compare this bootstrap method with other existing procedures in a wide Monte Carlo experiment by estimating, parametrically and semi-parametrically, the memory parameter d. We consider Gaussian and non-Gaussian processes to prove the robustness of the method to deviations from normality. The approach is also useful to estimate confidence intervals for the memory parameter d by improving the coverage level of the interval.
Resumo:
On the surface the subjects of Corporate Social Responsibility (CSR) and Critical Management Studies (CMS) seem to be closely related. Both are concerned with reflecting on the impact of management and organisation on employees, the wider community and the environment. Both suggest that there may be a need for organisations to take responsibility for and account of people other than shareholders and both have used the concept of accountability to suggest that organisations may need to do more than just comply with the legal framework.
Resumo:
In a recent decision by Mr Justice Laddie, a patent was held anticipated by, inter alia, prior use of a device which fell within the claims of the patent in suit, even though its circuitry was enclosed in resin. The anticipating invention had been "made available to the public" within the terms of section 2 (2) of the Patents Act 1977 because its essential integers would have been revealed by an interesting character, the "skilled forensic engineer".
Resumo:
Background: The quality of stormwater runoff from ports is significant as it can be an important source of pollution to the marine environment. This is also a significant issue for the Port of Brisbane as it is located in an area of high environmental values. Therefore, it is imperative to develop an in-depth understanding of stormwater runoff quality to ensure that appropriate strategies are in place for quality improvement, where necessary. To this end, the Port of Brisbane Corporation aimed to develop a port specific stormwater model for the Fisherman Islands facility. The need has to be considered in the context of the proposed future developments of the Port area. ----------------- The Project: The research project is an outcome of the collaborative Partnership between the Port of Brisbane Corporation (POBC) and Queensland University of Technology (QUT). A key feature of this Partnership is that it seeks to undertake research to assist the Port in strengthening the environmental custodianship of the Port area through ‘cutting edge’ research and its translation into practical application. ------------------ The project was separated into two stages. The first stage developed a quantitative understanding of the generation potential of pollutant loads in the existing land uses. This knowledge was then used as input for the stormwater quality model developed in the subsequent stage. The aim is to expand this model across the yet to be developed port expansion area. This is in order to predict pollutant loads associated with stormwater flows from this area with the longer term objective of contributing to the development of ecological risk mitigation strategies for future expansion scenarios. ----------------- Study approach: Stage 1 of the overall study confirmed that Port land uses are unique in terms of the anthropogenic activities occurring on them. This uniqueness in land use results in distinctive stormwater quality characteristics different to other conventional urban land uses. Therefore, it was not scientifically valid to consider the Port as belonging to a single land use category or to consider as being similar to any typical urban land use. The approach adopted in this study was very different to conventional modelling studies where modelling parameters are developed using calibration. The field investigations undertaken in Stage 1 of the overall study helped to create fundamental knowledge on pollutant build-up and wash-off in different Port land uses. This knowledge was then used in computer modelling so that the specific characteristics of pollutant build-up and wash-off can be replicated. This meant that no calibration processes were involved due to the use of measured parameters for build-up and wash-off. ---------------- Conclusions: Stage 2 of the study was primarily undertaken using the SWMM stormwater quality model. It is a physically based model which replicates natural processes as closely as possible. The time step used and catchment variability considered was adequate to accommodate the temporal and spatial variability of input parameters and the parameters used in the modelling reflect the true nature of rainfall-runoff and pollutant processes to the best of currently available knowledge. In this study, the initial loss values adopted for the impervious surfaces are relatively high compared to values noted in research literature. However, given the scientifically valid approach used for the field investigations, it is appropriate to adopt the initial losses derived from this study for future modelling of Port land uses. The relatively high initial losses will reduce the runoff volume generated as well as the frequency of runoff events significantly. Apart from initial losses, most of the other parameters used in SWMM modelling are generic to most modelling studies. Development of parameters for MUSIC model source nodes was one of the primary objectives of this study. MUSIC, uses the mean and standard deviation of pollutant parameters based on a normal distribution. However, based on the values generated in this study, the variation of Event Mean Concentrations (EMCs) for Port land uses within the given investigation period does not fit a normal distribution. This is possibly due to the fact that only one specific location was considered, namely the Port of Brisbane unlike in the case of the MUSIC model where a range of areas with different geographic and climatic conditions were investigated. Consequently, the assumptions used in MUSIC are not totally applicable for the analysis of water quality in Port land uses. Therefore, in using the parameters included in this report for MUSIC modelling, it is important to note that it may result in under or over estimations of annual pollutant loads. It is recommended that the annual pollutant load values given in the report should be used as a guide to assess the accuracy of the modelling outcomes. A step by step guide for using the knowledge generated from this study for MUSIC modelling is given in Table 4.6. ------------------ Recommendations: The following recommendations are provided to further strengthen the cutting edge nature of the work undertaken: * It is important to further validate the approach recommended for stormwater quality modelling at the Port. Validation will require data collection in relation to rainfall, runoff and water quality from the selected Port land uses. Additionally, the recommended modelling approach could be applied to a soon-to-be-developed area to assess ‘before’ and ‘after’ scenarios. * In the modelling study, TSS was adopted as the surrogate parameter for other pollutants. This approach was based on other urban water quality research undertaken at QUT. The validity of this approach should be further assessed for Port land uses. * The adoption of TSS as a surrogate parameter for other pollutants and the confirmation that the <150 m particle size range was predominant in suspended solids for pollutant wash-off gives rise to a number of important considerations. The ability of the existing structural stormwater mitigation measures to remove the <150 m particle size range need to be assessed. The feasibility of introducing source control measures as opposed to end-of-pipe measures for stormwater quality improvement may also need to be considered.
Resumo:
Uninhabited aerial vehicles (UAVs) are a cutting-edge technology that is at the forefront of aviation/aerospace research and development worldwide. Many consider their current military and defence applications as just a token of their enormous potential. Unlocking and fully exploiting this potential will see UAVs in a multitude of civilian applications and routinely operating alongside piloted aircraft. The key to realising the full potential of UAVs lies in addressing a host of regulatory, public relation, and technological challenges never encountered be- fore. Aircraft collision avoidance is considered to be one of the most important issues to be addressed, given its safety critical nature. The collision avoidance problem can be roughly organised into three areas: 1) Sense; 2) Detect; and 3) Avoid. Sensing is concerned with obtaining accurate and reliable information about other aircraft in the air; detection involves identifying potential collision threats based on available information; avoidance deals with the formulation and execution of appropriate manoeuvres to maintain safe separation. This thesis tackles the detection aspect of collision avoidance, via the development of a target detection algorithm that is capable of real-time operation onboard a UAV platform. One of the key challenges of the detection problem is the need to provide early warning. This translates to detecting potential threats whilst they are still far away, when their presence is likely to be obscured and hidden by noise. Another important consideration is the choice of sensors to capture target information, which has implications for the design and practical implementation of the detection algorithm. The main contributions of the thesis are: 1) the proposal of a dim target detection algorithm combining image morphology and hidden Markov model (HMM) filtering approaches; 2) the novel use of relative entropy rate (RER) concepts for HMM filter design; 3) the characterisation of algorithm detection performance based on simulated data as well as real in-flight target image data; and 4) the demonstration of the proposed algorithm's capacity for real-time target detection. We also consider the extension of HMM filtering techniques and the application of RER concepts for target heading angle estimation. In this thesis we propose a computer-vision based detection solution, due to the commercial-off-the-shelf (COTS) availability of camera hardware and the hardware's relatively low cost, power, and size requirements. The proposed target detection algorithm adopts a two-stage processing paradigm that begins with an image enhancement pre-processing stage followed by a track-before-detect (TBD) temporal processing stage that has been shown to be effective in dim target detection. We compare the performance of two candidate morphological filters for the image pre-processing stage, and propose a multiple hidden Markov model (MHMM) filter for the TBD temporal processing stage. The role of the morphological pre-processing stage is to exploit the spatial features of potential collision threats, while the MHMM filter serves to exploit the temporal characteristics or dynamics. The problem of optimising our proposed MHMM filter has been examined in detail. Our investigation has produced a novel design process for the MHMM filter that exploits information theory and entropy related concepts. The filter design process is posed as a mini-max optimisation problem based on a joint RER cost criterion. We provide proof that this joint RER cost criterion provides a bound on the conditional mean estimate (CME) performance of our MHMM filter, and this in turn establishes a strong theoretical basis connecting our filter design process to filter performance. Through this connection we can intelligently compare and optimise candidate filter models at the design stage, rather than having to resort to time consuming Monte Carlo simulations to gauge the relative performance of candidate designs. Moreover, the underlying entropy concepts are not constrained to any particular model type. This suggests that the RER concepts established here may be generalised to provide a useful design criterion for multiple model filtering approaches outside the class of HMM filters. In this thesis we also evaluate the performance of our proposed target detection algorithm under realistic operation conditions, and give consideration to the practical deployment of the detection algorithm onboard a UAV platform. Two fixed-wing UAVs were engaged to recreate various collision-course scenarios to capture highly realistic vision (from an onboard camera perspective) of the moments leading up to a collision. Based on this collected data, our proposed detection approach was able to detect targets out to distances ranging from about 400m to 900m. These distances, (with some assumptions about closing speeds and aircraft trajectories) translate to an advanced warning ahead of impact that approaches the 12.5 second response time recommended for human pilots. Furthermore, readily available graphic processing unit (GPU) based hardware is exploited for its parallel computing capabilities to demonstrate the practical feasibility of the proposed target detection algorithm. A prototype hardware-in- the-loop system has been found to be capable of achieving data processing rates sufficient for real-time operation. There is also scope for further improvement in performance through code optimisations. Overall, our proposed image-based target detection algorithm offers UAVs a cost-effective real-time target detection capability that is a step forward in ad- dressing the collision avoidance issue that is currently one of the most significant obstacles preventing widespread civilian applications of uninhabited aircraft. We also highlight that the algorithm development process has led to the discovery of a powerful multiple HMM filtering approach and a novel RER-based multiple filter design process. The utility of our multiple HMM filtering approach and RER concepts, however, extend beyond the target detection problem. This is demonstrated by our application of HMM filters and RER concepts to a heading angle estimation problem.
Resumo:
As Web searching becomes more prolific for information access worldwide, we need to better understand users’ Web searching behaviour and develop better models of their interaction with Web search systems. Web search modelling is a significant and important area of Web research. Searching on the Web is an integral element of information behaviour and human–computer interaction. Web searching includes multitasking processes, the allocation of cognitive resources among several tasks, and shifts in cognitive, problem and knowledge states. In addition to multitasking, cognitive coordination and cognitive shifts are also important, but are under-explored aspects of Web searching. During the Web searching process, beyond physical actions, users experience various cognitive activities. Interactive Web searching involves many users’ cognitive shifts at different information behaviour levels. Cognitive coordination allows users to trade off the dependences among multiple information tasks and the resources available. Much research has been conducted into Web searching. However, few studies have modelled the nature of and relationship between multitasking, cognitive coordination and cognitive shifts in the Web search context. Modelling how Web users interact with Web search systems is vital for the development of more effective Web IR systems. This study aims to model the relationship between multitasking, cognitive coordination and cognitive shifts during Web searching. A preliminary theoretical model is presented based on previous studies. The research is designed to validate the preliminary model. Forty-two study participants were involved in the empirical study. A combination of data collection instruments, including pre- and post-questionnaires, think-aloud protocols, search logs, observations and interviews were employed to obtain users’ comprehensive data during Web search interactions. Based on the grounded theory approach, qualitative analysis methods including content analysis and verbal protocol analysis were used to analyse the data. The findings were inferred through an analysis of questionnaires, a transcription of think-aloud protocols, the Web search logs, and notes on observations and interviews. Five key findings emerged. (1) Multitasking during Web searching was demonstrated as a two-dimensional behaviour. The first dimension was represented as multiple information problems searching by task switching. Users’ Web searching behaviour was a process of multiple tasks switching, that is, from searching on one information problem to searching another. The second dimension of multitasking behaviour was represented as an information problem searching within multiple Web search sessions. Users usually conducted Web searching on a complex information problem by submitting multiple queries, using several Web search systems and opening multiple windows/tabs. (2) Cognitive shifts were the brain’s internal response to external stimuli. Cognitive shifts were found as an essential element of searching interactions and users’ Web searching behaviour. The study revealed two kinds of cognitive shifts. The first kind, the holistic shift, included users’ perception on the information problem and overall information evaluation before and after Web searching. The second kind, the state shift, reflected users’ changes in focus between the different cognitive states during the course of Web searching. Cognitive states included users’ focus on the states of topic, strategy, evaluation, view and overview. (3) Three levels of cognitive coordination behaviour were identified: the information task coordination level, the coordination mechanism level, and the strategy coordination level. The three levels of cognitive coordination behaviour interplayed to support multiple information tasks switching. (4) An important relationship existed between multitasking, cognitive coordination and cognitive shifts during Web searching. Cognitive coordination as a management mechanism bound together other cognitive processes, including multitasking and cognitive shifts, in order to move through users’ Web searching process. (5) Web search interaction was shown to be a multitasking process which included information problems ordering, task switching and task and mental coordinating; also, at a deeper level, cognitive shifts took place. Cognitive coordination was the hinge behaviour linking multitasking and cognitive shifts. Without cognitive coordination, neither multitasking Web searching behaviour nor the complicated mental process of cognitive shifting could occur. The preliminary model was revisited with these empirical findings. A revised theoretical model (MCC Model) was built to illustrate the relationship between multitasking, cognitive coordination and cognitive shifts during Web searching. Implications and limitations of the study are also discussed, along with future research work.