410 resultados para Analysis of literature
em Queensland University of Technology - ePrints Archive
Resumo:
The importance of a thorough and systematic literature review has long been recognised across academic domains as critical to the foundation of new knowledge and theory evolution. Driven by an exponentially growing body of knowledge in the IS discipline, there has been a recent influx of guidance on how to conduct a literature review. As literature reviews are emerging as a standalone research method in itself, increasingly these method focused guidelines are of great interest, receiving acceptance at top tier IS publication outlets. Nevertheless, the finer details which offer justification for the selected content, and the effective presentation of supporting data has not been widely discussed in these method papers to date. This paper addresses this gap by exploring the concept of ‘literature profiling’ while arguing that it is a key aspect of a comprehensive literature review. The study establishes the importance of profiling for managing aspects such as quality assurance, transparency and the mitigation of selection bias. And then discusses how profiling can provide a valid basis for data analysis based on the attributes of selected literature. In essence, this study has conducted an archival analysis of literature (predominately from the IS domain) to present its main argument; the value for literature profiling, with supporting exemplary illustrations.
Resumo:
GLBTI (gay, lesbian, bisexual, transgender, intersex) police liaison programs have been an important part of policing these communities for a number of decades now. In fact, this model appears to dominate approaches as the preferred way to manage relationships between GLBTI communities and police. Interestingly, while this model dominates, research on the effectiveness of this model, and the services that align with it, is limited. To date, only few studies have asked critical questions about the effectiveness of GLBTI police liaison services. For instance, we know that over 70% of GLBTI communities are aware of police liaison services, but only 4% of those victimised access them (Berman & Robinson, 2010). This paper critically examines existing literature about GLBTI police liaison services to demonstrate key themes and omissions. It argues that police liaison services as a preferred model has been taken for granted as the correct way of engaging with GLBTI communities in ways that may exclude other forms of engagement, and suggests that further research into these liaison programs is necessary if the relationships between GLBTI communities and the police are to be strengthened.
Resumo:
This article presents the first narrative analysis of the areas of research that have developed within the destination marketing field since its commencement in 1973. Given the broad extent of the field, and the absence of any previous reviews in four decades, a key challenge is in providing a focus for such a disparate body of knowledge. The review is structured around one principal question: ‘To what extent is the Destination Marketing Organisation (DMO) responsible for the competitiveness of the destination’? In pursuit of this underlying question, we address a number of themes including nomenclature and the DMO, the evolution of the destination marketing literature, competitiveness as the DMO reason d’être, and DMO effectiveness including issues of branding and positioning, and future research themes in the field.
Resumo:
A significant gap in the tourism and travel literature exists in the area of tourism destination branding. Although brands have been used as sources of differentiation in consumer goods markets for over a century, academic research attention towards destination branding has only been reported since the late 1990s. Three important components of the brand construct are brand identity, brand position and brand image. While interest in applications of brand theory to practise in tourism is increasing, there is a paucity of published research in the literature to guide destination marketing organisations (DMOs). In particular there have been few reported analyses of destination brand positioning slogans. The focus of this paper is on destination brand position slogans, which represent the interface between brand identity and brand image. Part of a wider investigation of DMO slogans worldwide, and in keeping with the conference location, the paper focuses on analysis of slogans used by New Zealand RTOs. The slogans are examined in terms of the extent to which they have been limited to ephemeral indifference. In other words, have they stood the test of time and do they effectively differentiate through a meaningful proposition? Analysis of the slogans indicates very few could be characterised as memorably distinctive. This reflects the complexity involved in capturing the essence of a multi-attributed destination in a succinct and focused positioning slogan, in a way that is both meaningful to the target audience and effectively differentiates the destination from competitors offering the same benefits.
Resumo:
Over the past several years, there has been resurgent interest in regional planning in North America, Europe and Australasia. Spurred by issues such as metropolitan growth, transportation infrastructure, environmental management and economic development, many states and metropolitan regions are undertaking new planning initiatives. These regional efforts have also raised significant question about governance structures, accountability and measures of effectiveness.n this paper, the authors conducted an international review of ten case studies from the United States, Canada, England, Belgium, New Zealand and Australia to explore several critical questions. Using qualitative data template, the research team reviewed plans, documents, web sites and published literature to address three questions. First, what are the governance arrangements for delivering regional planning? Second, what are the mechanisms linking regional plans with state plans (when relevant) and local plans? Third, what means and mechanisms do these regional plans use to evaluate and measure effectiveness? The case study analysis revealed several common themes. First, there is an increasing focus on goverance at the regional level, which is being driven by a range of trends, including regional spatial development initiatives in Europe, regional transportation issues in the US, and the growth of metropolitan regions generally. However, there is considerable variation in how regional governance arrangements are being played out. Similarly, there is a range of processes being used at the regional level to guide planning that range from broad ranging (thick) processes to narrow and limited (thin) approaches. Finally, evaluation and monitoring of regional planning efforts are compiling data on inputs, processes, outputs and outcomes. Although there is increased attention being paid to indicators and monitoring, most of it falls into outcome evaluations such as Agenda 21 or sustainability reporting. Based on our review we suggest there is a need for increased attention on input, process and output indicators and clearer linkages of these indicators in monitoring and evaluation frameworks. The focus on outcome indicators, such as sustainability indicators, creates feedback systems that are too long-term and remote for effective monitoring and feedback. Although we found some examples of where these kinds of monitoring frameworks are linked into a system of governance, there is a need for clearer conceptual development for both theory and practice.
Resumo:
Although the benefits of service orientation are prevalent in literature, a review, analysis, and evaluation of the 30 existing service analysis approaches presented in this paper have shown that a comprehensive approach to the identification and analysis of both business and supporting software services is missing. Based on this evaluation of existing approaches and additional sources, we close this gap by proposing an integrated, consolidated approach to business and software service analysis that combines and extends the strengths of the examined methodologies.
Resumo:
Machine downtime, whether planned or unplanned, is intuitively costly to manufacturing organisations, but is often very difficult to quantify. The available literature showed that costing processes are rarely undertaken within manufacturing organisations. Where cost analyses have been undertaken, they generally have only valued a small proportion of the affected costs, leading to an overly conservative estimate. This thesis aimed to develop a cost of downtime model, with particular emphasis on the application of the model to Australia Post’s Flat Mail Optical Character Reader (FMOCR). The costing analysis determined a cost of downtime of $5,700,000 per annum, or an average cost of $138 per operational hour. The second section of this work focused on the use of the cost of downtime to objectively determine areas of opportunity for cost reduction on the FMOCR. This was the first time within Post that maintenance costs were considered along side of downtime for determining machine performance. Because of this, the results of the analysis revealed areas which have historically not been targeted for cost reduction. Further exploratory work was undertaken on the Flats Lift Module (FLM) and Auto Induction Station (AIS) Deceleration Belts through the comparison of the results against two additional FMOCR analysis programs. This research has demonstrated the development of a methodical and quantifiable cost of downtime for the FMOCR. This has been the first time that Post has endeavoured to examine the cost of downtime. It is also one of the very few methodologies for valuing downtime costs that has been proposed in literature. The work undertaken has also demonstrated how the cost of downtime can be incorporated into machine performance analysis with specific application to identifying high costs modules. The outcome of this report has both been the methodology for costing downtime, as well as a list of areas for cost reduction. In doing so, this thesis has outlined the two key deliverables presented at the outset of the research.
Resumo:
This paper aims to develop an effective numerical simulation technique for the dynamic deflection analysis of nanotubes-based nanoswitches. The nanoswitch is simplified to a continuum structure, and some key material parameters are extracted from typical molecular dynamics (MD). An advanced local meshless formulation is applied to obtain the discretized dynamic equations for the numerical solution. The developed numerical technique is firstly validated by the static deflection analyses of nanoswitches, and then, the fundamental dynamic properties of nanoswitches are analyzed. A parametric comparison with the results in the literature and from experiments shows that the developed modelling approach is accurate, efficient and effective.
Resumo:
Malcolm Shepherd Knowles was a key writer and theorist in the field of adult education in the United States. He died in 1997 and left a large legacy of books and journal articles. This thesis traced the development of his thinking over the 46-year period from 1950 to 1995. It examined the 25 works authored, co-authored, edited, reissued and revised by him during that period. The writings were scrutinised using a literature research methodology to expose the theoretical content, and a history of thought lens to identify and account for the development of major ideas. The methodology enabled a gradual unfolding of the history. A broadly-consistent and sequential pattern of thought focusing on the notion of andragogy emerged. The study revealed that after the initial phases of exploratory thinking, Knowles developed a practical-theoretical framework he believed could function as a comprehensive theory of adult learning. As his thinking progressed, his theory developed into a unified framework for human resource development and, later, into a model for the development of self-directed lifelong learners. The study traced the development of Knowles’ thinking through the phases of thought, identified the writings that belonged within each phase and produced a series of diagrammatic representations showing the evolution of his conceptual framework. The production of a history of the development of Knowles’ thought is the major outcome of the study. In addition to plotting the narrative sequence of thought-events, the history helps to explicate the factors and conditions that influenced Knowles’ thinking and to show the interrelationships between ideas. The study should help practitioners in their use and appreciation of Knowles’ works.
Resumo:
A growing literature seeks to explain differences in individuals' self-reported satisfaction with their jobs. The evidence so far has mainly been based on cross-sectional data and when panel data have been used, individual unobserved heterogeneity has been modelled as an ordered probit model with random effects. This article makes use of longitudinal data for Denmark, taken from the waves 1995-1999 of the European Community Household Panel, and estimates fixed effects ordered logit models using the estimation methods proposed by Ferrer-i-Carbonel and Frijters (2004) and Das and van Soest (1999). For comparison and testing purposes a random effects ordered probit is also estimated. Estimations are carried out separately on the samples of men and women for individuals' overall satisfaction with the jobs they hold. We find that using the fixed effects approach (that clearly rejects the random effects specification), considerably reduces the number of key explanatory variables. The impact of central economic factors is the same as in previous studies, though. Moreover, the determinants of job satisfaction differ considerably between the genders, in particular once individual fixed effects are allowed for.
Resumo:
The main objective of this PhD was to further develop Bayesian spatio-temporal models (specifically the Conditional Autoregressive (CAR) class of models), for the analysis of sparse disease outcomes such as birth defects. The motivation for the thesis arose from problems encountered when analyzing a large birth defect registry in New South Wales. The specific components and related research objectives of the thesis were developed from gaps in the literature on current formulations of the CAR model, and health service planning requirements. Data from a large probabilistically-linked database from 1990 to 2004, consisting of fields from two separate registries: the Birth Defect Registry (BDR) and Midwives Data Collection (MDC) were used in the analyses in this thesis. The main objective was split into smaller goals. The first goal was to determine how the specification of the neighbourhood weight matrix will affect the smoothing properties of the CAR model, and this is the focus of chapter 6. Secondly, I hoped to evaluate the usefulness of incorporating a zero-inflated Poisson (ZIP) component as well as a shared-component model in terms of modeling a sparse outcome, and this is carried out in chapter 7. The third goal was to identify optimal sampling and sample size schemes designed to select individual level data for a hybrid ecological spatial model, and this is done in chapter 8. Finally, I wanted to put together the earlier improvements to the CAR model, and along with demographic projections, provide forecasts for birth defects at the SLA level. Chapter 9 describes how this is done. For the first objective, I examined a series of neighbourhood weight matrices, and showed how smoothing the relative risk estimates according to similarity by an important covariate (i.e. maternal age) helped improve the model’s ability to recover the underlying risk, as compared to the traditional adjacency (specifically the Queen) method of applying weights. Next, to address the sparseness and excess zeros commonly encountered in the analysis of rare outcomes such as birth defects, I compared a few models, including an extension of the usual Poisson model to encompass excess zeros in the data. This was achieved via a mixture model, which also encompassed the shared component model to improve on the estimation of sparse counts through borrowing strength across a shared component (e.g. latent risk factor/s) with the referent outcome (caesarean section was used in this example). Using the Deviance Information Criteria (DIC), I showed how the proposed model performed better than the usual models, but only when both outcomes shared a strong spatial correlation. The next objective involved identifying the optimal sampling and sample size strategy for incorporating individual-level data with areal covariates in a hybrid study design. I performed extensive simulation studies, evaluating thirteen different sampling schemes along with variations in sample size. This was done in the context of an ecological regression model that incorporated spatial correlation in the outcomes, as well as accommodating both individual and areal measures of covariates. Using the Average Mean Squared Error (AMSE), I showed how a simple random sample of 20% of the SLAs, followed by selecting all cases in the SLAs chosen, along with an equal number of controls, provided the lowest AMSE. The final objective involved combining the improved spatio-temporal CAR model with population (i.e. women) forecasts, to provide 30-year annual estimates of birth defects at the Statistical Local Area (SLA) level in New South Wales, Australia. The projections were illustrated using sixteen different SLAs, representing the various areal measures of socio-economic status and remoteness. A sensitivity analysis of the assumptions used in the projection was also undertaken. By the end of the thesis, I will show how challenges in the spatial analysis of rare diseases such as birth defects can be addressed, by specifically formulating the neighbourhood weight matrix to smooth according to a key covariate (i.e. maternal age), incorporating a ZIP component to model excess zeros in outcomes and borrowing strength from a referent outcome (i.e. caesarean counts). An efficient strategy to sample individual-level data and sample size considerations for rare disease will also be presented. Finally, projections in birth defect categories at the SLA level will be made.
Resumo:
Boards of directors are thought to provide access to a wealth of knowledge and resources for the companies they serve, and are considered important to corporate governance. Under the Resource Based View (RBV) of the firm (Wernerfelt, 1984) boards are viewed as a strategic resource available to firms. As a consequence there has been a significant research effort aimed at establishing a link between board attributes and company performance. In this thesis I explore and extend the study of interlocking directorships (Mizruchi, 1996; Scott 1991a) by examining the links between directors’ opportunity networks and firm performance. Specifically, I use resource dependence theory (Pfeffer & Salancik, 1978) and social capital theory (Burt, 1980b; Coleman, 1988) as the basis for a new measure of a board’s opportunity network. I contend that both directors’ formal company ties and their social ties determine a director’s opportunity network through which they are able to access and mobilise resources for their firms. This approach is based on recent studies that suggest the measurement of interlocks at the director level, rather than at the firm level, may be a more reliable indicator of this phenomenon. This research uses publicly available data drawn from Australia’s top-105 listed companies and their directors in 1999. I employ Social Network Analysis (SNA) (Scott, 1991b) using the UCINET software to analyse the individual director’s formal and social networks. SNA is used to measure a the number of ties a director has to other directors in the top-105 company director network at both one and two degrees of separation, that is, direct ties and indirect (or ‘friend of a friend’) ties. These individual measures of director connectedness are aggregated to produce a board-level network metric for comparison with measures of a firm’s performance using multiple regression analysis. Performance is measured with accounting-based and market-based measures. Findings indicate that better-connected boards are associated with higher market-based company performance (measured by Tobin’s q). However, weaker and mostly unreliable associations were found for accounting-based performance measure ROA. Furthermore, formal (or corporate) network ties are a stronger predictor of market performance than total network ties (comprising social and corporate ties). Similarly, strong ties (connectedness at degree-1) are better predictors of performance than weak ties (connectedness at degree-2). My research makes four contributions to the literature on director interlocks. First, it extends a new way of measuring a board’s opportunity network based on the director rather than the company as the unit of interlock. Second, it establishes evidence of a relationship between market-based measures of firm performance and the connectedness of that firm’s board. Third, it establishes that director’s formal corporate ties matter more to market-based firm performance than their social ties. Fourth, it establishes that director’s strong direct ties are more important to market-based performance than weak ties. The thesis concludes with implications for research and practice, including a more speculative interpretation of these results. In particular, I raise the possibility of reverse causality – that is networked directors seek to join high-performing companies. Thus, the relationship may be a result of symbolic action by companies seeking to increase the legitimacy of their firms rather than a reflection of the social capital available to the companies. This is an important consideration worthy of future investigation.
Resumo:
Spectrum sensing is considered to be one of the most important tasks in cognitive radio. Many sensing detectors have been proposed in the literature, with the common assumption that the primary user is either fully present or completely absent within the window of observation. In reality, there are scenarios where the primary user signal only occupies a fraction of the observed window. This paper aims to analyse the effect of the primary user duty cycle on spectrum sensing performance through the analysis of a few common detectors. Simulations show that the probability of detection degrades severely with reduced duty cycle regardless of the detection method. Furthermore we show that reducing the duty cycle has a greater degradation on performance than lowering the signal strength.