910 resultados para Hough transform
Resumo:
This paper introduces the Corporate Culture Change Cycle: a continuous innovation methodology for transforming the psychological contract in an organisational context. The eight step process is based on the action learning model. The purpose of this methodology is to benchmark the psychological contract against eight changing values of the employment relationship as a basis for facilitating a process of aligning the changing needs of employer and employee. Both the Corporate Culture Change Cycle and the New Employment Relationship Model have been validated in several organisational settings and subsequently refined. This continuous innovation methodology addresses gaps in the psychological contract, change management and continuous innovation research literatures. The approach therefore should be of interest to researchers in these fields of study and from a practical perspective for managers wishing to transform their workplace cultures.
Resumo:
In moving from lowest cost adversarial based traditional procurement towards value driven methodologies the challenges range from re-engineering the process, to metrics and team alignment. This paper describes research into methodologies which encourage alignment of project partners towards achieving mutually beneficial goals. The research identifies nine variables which influence the achievement of successful projects delivering value. Results from case studies illustrate that not all parties can achieve value for themselves which directs attention to the balance between deliverables and the interests of team members. Re- valuing construction demands refocusing towards the delivery of operational assets and their place in the value system whilst recognising the need to manage the delivery process and the team to align the value to the parties. The objective of the project was to develop tools and recommendations for reform of project delivery in the building and construction industry to transform business-as-usual performance into exceptional performance. Benefits flow not only to the construction industry, but to the community as a whole because a more sophisticated industry can deliver more effective use of assets, financing, operating and maintenance of facilities to suit the community’s needs. This research was funded by the Australian Cooperative Research Centre for Construction Innovation.
Resumo:
Readers and writers use a variety of modes of inscription – print, oral and multimedia – to understand, analyze, critique and transform their social, cultural and political worlds. Beginning from Freire (1970), ‘critical literacy’ has become a theoretically diverse educational project, drawing from reader response theory, linguistic and grammatical analysis from critical linguistics, feminist, poststructuralist, postcolonial and critical race theory, and cultural and media studies. In the UK, Australia, Canada, South Africa, New Zealand and the US different approaches to critical literacy have been developed in curriculum and schools. These focus on social and cultural analysis and on how print and digital texts and discourses work, with a necessary and delicate tension between classroom emphasis on student and community cultural ‘voice’ and social analysis – and on explicit engagement with the technical features and social uses of written and multimodal texts.
Resumo:
Automatic detection of suspicious activities in CCTV camera feeds is crucial to the success of video surveillance systems. Such a capability can help transform the dumb CCTV cameras into smart surveillance tools for fighting crime and terror. Learning and classification of basic human actions is a precursor to detecting suspicious activities. Most of the current approaches rely on a non-realistic assumption that a complete dataset of normal human actions is available. This paper presents a different approach to deal with the problem of understanding human actions in video when no prior information is available. This is achieved by working with an incomplete dataset of basic actions which are continuously updated. Initially, all video segments are represented by Bags-Of-Words (BOW) method using only Term Frequency-Inverse Document Frequency (TF-IDF) features. Then, a data-stream clustering algorithm is applied for updating the system's knowledge from the incoming video feeds. Finally, all the actions are classified into different sets. Experiments and comparisons are conducted on the well known Weizmann and KTH datasets to show the efficacy of the proposed approach.
Resumo:
Glass transition temperature of spaghetti sample was measured by thermal and rheological methods as a function of water content.
Resumo:
There is no specific self-efficacy measure that has been developed primarily for problem drinkers seeking a moderation drinking goal. In this article, we report the factor structure of a 20-item Controlled Drinking Self-Efficacy Scale (CDSES; Sitharthan et al., 1996; Sitharthan et al., 1997). The results indicate that the CDSES is highly reliable, and the factor analysis using the full sample identified four factors: negative affect, positive mood/social context, frequency of drinking, and consumption quantity. A similar factor structure was obtained for the subsample of men. In contrast, only three factors emerged in the analysis of data on female participants. Compared to women, men had low self-efficacy to control their drinking in situations relating to positive mood/social context, and subjects with high alcohol dependence had low self-efficacy for situations relating to negative affect, social situations, and drinking less frequently. The CDSES can be a useful measure in treatment programs providing a moderation drinking goal.
Resumo:
Context The School of Information Technology at QUT has recently undertaken a major restructuring of their Bachelor of Information Technology (BIT) course. Some of the aims of this restructuring include a reduction in first year attrition and to provide an attractive degree course that meets both student and industry expectations. Emphasis has been placed on the first semester in the context of retaining students by introducing a set of four units that complement one another and provide introductory material on technology, programming and related skills, and generic skills that will aid the students throughout their undergraduate course and in their careers. This discussion relates to one of these four fist semester units, namely Building IT Systems. The aim of this unit is to create small Information Technology (IT) systems that use programming or scripting, databases as either standalone applications or web applications. In the prior history of teaching introductory computer programming at QUT, programming has been taught as a stand alone subject and integration of computer applications with other systems such as databases and networks was not undertaken until students had been given a thorough grounding in those topics as well. Feedback has indicated that students do not believe that working with a database requires programming skills. In fact, the teaching of the building blocks of computer applications have been compartmentalized and taught in isolation from each other. The teaching of introductory computer programming has been an industry requirement of IT degree courses as many jobs require at least some knowledge of the topic. Yet, computer programming is not a skill that all students have equal capabilities of learning (Bruce et al., 2004) and this is clearly shown by the volume of publications dedicated to this topic in the literature over a broad period of time (Eckerdal & Berglund, 2005; Mayer, 1981; Winslow, 1996). The teaching of this introductory material has been done pretty much the same way over the past thirty years. During this period of time that introductory computer programming courses have been taught at QUT, a number of different programming languages and programming paradigms have been used and different approaches to teaching and learning have been attempted in an effort to find the golden thread that would allow students to learn this complex topic. Unfortunately, computer programming is not a skill that can be learnt in one semester. Some basics can be learnt but it can take many years to master (Norvig, 2001). Faculty data typically has shown a bimodal distribution of results for students undertaking introductory programming courses with a high proportion of students receiving a high mark and a high proportion of students receiving a low or failing mark. This indicates that there are students who understand and excel with the introductory material while there is another group who struggle to understand the concepts and practices required to be able to translate a specification or problem statement into a computer program that achieves what is being requested. The consequence of a large group of students failing the introductory programming course has been a high level of attrition amongst first year students. This attrition level does not provide good continuity in student numbers in later years of the degree program and the current approach is not seen as sustainable.
Resumo:
Structural health monitoring (SHM) is the term applied to the procedure of monitoring a structure’s performance, assessing its condition and carrying out appropriate retrofitting so that it performs reliably, safely and efficiently. Bridges form an important part of a nation’s infrastructure. They deteriorate due to age and changing load patterns and hence early detection of damage helps in prolonging the lives and preventing catastrophic failures. Monitoring of bridges has been traditionally done by means of visual inspection. With recent developments in sensor technology and availability of advanced computing resources, newer techniques have emerged for SHM. Acoustic emission (AE) is one such technology that is attracting attention of engineers and researchers all around the world. This paper discusses the use of AE technology in health monitoring of bridge structures, with a special focus on analysis of recorded data. AE waves are stress waves generated by mechanical deformation of material and can be recorded by means of sensors attached to the surface of the structure. Analysis of the AE signals provides vital information regarding the nature of the source of emission. Signal processing of the AE waveform data can be carried out in several ways and is predominantly based on time and frequency domains. Short time Fourier transform and wavelet analysis have proved to be superior alternatives to traditional frequency based analysis in extracting information from recorded waveform. Some of the preliminary results of the application of these analysis tools in signal processing of recorded AE data will be presented in this paper.
Resumo:
With service interaction modelling, it is customary to distinguish between two types of models: choreographies and orchestrations. A choreography describes interactions within a collection of services from a global perspective, where no service plays a privileged role. Instead, services interact in a peer-to-peer manner. In contrast, an orchestration describes the interactions between one particular service, the orchestrator, and a number of partner services. The main proposition of this work is an approach to bridge these two modelling viewpoints by synthesising orchestrators from choreographies. To start with, choreographies are defined using a simple behaviour description language based on communicating finite state machines. From such a model, orchestrators are initially synthesised in the form of state machines. It turns out that state machines are not suitable for orchestration modelling, because orchestrators generally need to engage in concurrent interactions. To address this issue, a technique is proposed to transform state machines into process models in the Business Process Modelling Notation (BPMN). Orchestrations represented in BPMN can then be augmented with additional business logic to achieve value-adding mediation. In addition, techniques exist for refining BPMN models into executable process definitions. The transformation from state machines to BPMN relies on Petri nets as an intermediary representation and leverages techniques from theory of regions to identify concurrency in the initial Petri net. Once concurrency has been identified, the resulting Petri net is transformed into a BPMN model. The original contributions of this work are: an algorithm to synthesise orchestrators from choreographies and a rules-based transformation from Petri nets into BPMN.
Resumo:
Event-specific scales commonly have greater power than generalized scales in prediction of specific disorders and in testing mediator models for predicting such disorders. Therefore, in a preliminary study, a 6-item Alcohol Helplessness Scale was constructed and found to be reliable for a sample of 98 problem drinkers. Hierarchical multiple regression and its derivative path analysis were used to test whether helplessness and self-efficacy moderate or mediate the link between alcohol dependence and depression, A test of a moderation model was not supported, whereas a test of a mediation model was supported. Helplessness and self-efficacy both significantly and independently mediated between alcohol dependence and depression. Nevertheless, a significant direct effect of alcohol dependence on depression also remained.
Resumo:
Silylated layered double hydroxides (LDHs) were synthesized through a surfactant-free method involving an in situ condensation of silane with the surface hydroxyl group of LDHs during its reconstruction in carbonate solution. X-ray diffraction (XRD) patterns showed the silylation reaction occurred on the external surfaces of LDHs layers. The successful silylation was evidenced by 29Si cross-polarization magic-angle spinning nuclear magnetic resonance (29Si CP/MAS NMR) spectroscopy, attenuated total reflection Fourier transform infrared (ATR FTIR) spectroscopy, and infrared emission spectroscopy (IES). The ribbon shaped crystallites with a “rodlike” aggregation were observed through transmission electron microscopy (TEM) images. The aggregation was explained by the T2 and T3 types of linkage between adjacent silane molecules as indicated in the 29Si NMR spectrum. In addition, the silylated products show high thermal stability by maintained Si related bands even when the temperature was increased to 1000 °C as observed in IES spectra.
Resumo:
Intuitively, any `bag of words' approach in IR should benefit from taking term dependencies into account. Unfortunately, for years the results of exploiting such dependencies have been mixed or inconclusive. To improve the situation, this paper shows how the natural language properties of the target documents can be used to transform and enrich the term dependencies to more useful statistics. This is done in three steps. The term co-occurrence statistics of queries and documents are each represented by a Markov chain. The paper proves that such a chain is ergodic, and therefore its asymptotic behavior is unique, stationary, and independent of the initial state. Next, the stationary distribution is taken to model queries and documents, rather than their initial distri- butions. Finally, ranking is achieved following the customary language modeling paradigm. The main contribution of this paper is to argue why the asymptotic behavior of the document model is a better representation then just the document's initial distribution. A secondary contribution is to investigate the practical application of this representation in case the queries become increasingly verbose. In the experiments (based on Lemur's search engine substrate) the default query model was replaced by the stable distribution of the query. Just modeling the query this way already resulted in significant improvements over a standard language model baseline. The results were on a par or better than more sophisticated algorithms that use fine-tuned parameters or extensive training. Moreover, the more verbose the query, the more effective the approach seems to become.
Resumo:
Matrix function approximation is a current focus of worldwide interest and finds application in a variety of areas of applied mathematics and statistics. In this thesis we focus on the approximation of A^(-α/2)b, where A ∈ ℝ^(n×n) is a large, sparse symmetric positive definite matrix and b ∈ ℝ^n is a vector. In particular, we will focus on matrix function techniques for sampling from Gaussian Markov random fields in applied statistics and the solution of fractional-in-space partial differential equations. Gaussian Markov random fields (GMRFs) are multivariate normal random variables characterised by a sparse precision (inverse covariance) matrix. GMRFs are popular models in computational spatial statistics as the sparse structure can be exploited, typically through the use of the sparse Cholesky decomposition, to construct fast sampling methods. It is well known, however, that for sufficiently large problems, iterative methods for solving linear systems outperform direct methods. Fractional-in-space partial differential equations arise in models of processes undergoing anomalous diffusion. Unfortunately, as the fractional Laplacian is a non-local operator, numerical methods based on the direct discretisation of these equations typically requires the solution of dense linear systems, which is impractical for fine discretisations. In this thesis, novel applications of Krylov subspace approximations to matrix functions for both of these problems are investigated. Matrix functions arise when sampling from a GMRF by noting that the Cholesky decomposition A = LL^T is, essentially, a `square root' of the precision matrix A. Therefore, we can replace the usual sampling method, which forms x = L^(-T)z, with x = A^(-1/2)z, where z is a vector of independent and identically distributed standard normal random variables. Similarly, the matrix transfer technique can be used to build solutions to the fractional Poisson equation of the form ϕn = A^(-α/2)b, where A is the finite difference approximation to the Laplacian. Hence both applications require the approximation of f(A)b, where f(t) = t^(-α/2) and A is sparse. In this thesis we will compare the Lanczos approximation, the shift-and-invert Lanczos approximation, the extended Krylov subspace method, rational approximations and the restarted Lanczos approximation for approximating matrix functions of this form. A number of new and novel results are presented in this thesis. Firstly, we prove the convergence of the matrix transfer technique for the solution of the fractional Poisson equation and we give conditions by which the finite difference discretisation can be replaced by other methods for discretising the Laplacian. We then investigate a number of methods for approximating matrix functions of the form A^(-α/2)b and investigate stopping criteria for these methods. In particular, we derive a new method for restarting the Lanczos approximation to f(A)b. We then apply these techniques to the problem of sampling from a GMRF and construct a full suite of methods for sampling conditioned on linear constraints and approximating the likelihood. Finally, we consider the problem of sampling from a generalised Matern random field, which combines our techniques for solving fractional-in-space partial differential equations with our method for sampling from GMRFs.
Resumo:
In this paper, we consider the numerical solution of a fractional partial differential equation with Riesz space fractional derivatives (FPDE-RSFD) on a finite domain. Two types of FPDE-RSFD are considered: the Riesz fractional diffusion equation (RFDE) and the Riesz fractional advection–dispersion equation (RFADE). The RFDE is obtained from the standard diffusion equation by replacing the second-order space derivative with the Riesz fractional derivative of order αset membership, variant(1,2]. The RFADE is obtained from the standard advection–dispersion equation by replacing the first-order and second-order space derivatives with the Riesz fractional derivatives of order βset membership, variant(0,1) and of order αset membership, variant(1,2], respectively. Firstly, analytic solutions of both the RFDE and RFADE are derived. Secondly, three numerical methods are provided to deal with the Riesz space fractional derivatives, namely, the L1/L2-approximation method, the standard/shifted Grünwald method, and the matrix transform method (MTM). Thirdly, the RFDE and RFADE are transformed into a system of ordinary differential equations, which is then solved by the method of lines. Finally, numerical results are given, which demonstrate the effectiveness and convergence of the three numerical methods.
Resumo:
The TraSe (Transform-Select) algorithm has been developed to investigate the morphing of electronic music through automatically applying a series of deterministic compositional transformations to the source, guided towards a target by similarity metrics. This is in contrast to other morphing techniques such as interpolation or parameters or probabilistic variation. TraSe allows control over stylistic elements of the music through user-defined weighting of numerous compositional transformations. The formal evaluation of TraSe was mostly qualitative and occurred through nine participants completing an online questionnaire. The music generated by TraSe was generally felt to be less coherent than a human composed benchmark but in some cases judged as more creative.