950 resultados para Statistical Model


Relevância:

30.00% 30.00%

Publicador:

Resumo:

Likelihood computation in spatial statistics requires accurate and efficient calculation of the normalizing constant (i.e. partition function) of the Gibbs distribution of the model. Two available methods to calculate the normalizing constant by Markov chain Monte Carlo methods are compared by simulation experiments for an Ising model, a Gaussian Markov field model and a pairwise interaction point field model.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Spam is commonly defined as unsolicited email messages and the goal of spam categorization is to distinguish between spam and legitimate email messages. Many researchers have been trying to separate spam from legitimate emails using machine learning algorithms based on statistical learning methods. In this paper, an innovative and intelligent spam filtering model has been proposed based on support vector machine (SVM). This model combines both linear and nonlinear SVM techniques where linear SVM performs better for text based spam classification that share similar characteristics. The proposed model considers both text and image based email messages for classification by selecting an appropriate kernel function for information transformation.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Researchers worldwide have been actively seeking for the most robust and powerful solutions to detect and classify key events (or highlights) in various sports domains. Most approaches have employed manual heuristics that model the typical pattern of audio-visual features within particular sport events To avoid manual observation and knowledge, machine-learning can be used as an alternative approach. To bridge the gaps between these two alternatives, an attempt is made to integrate statistics into heuristic models during highlight detection in our investigation. The models can be designed with a modest amount of domain-knowledge, making them less subjective and more robust for different sports. We have also successfully used a universal scope of detection and a standard set of features that can be applied for different sports that include soccer, basketball and Australian football. An experiment on a large dataset of sport videos, with a total of around 15 hours, has demonstrated the effectiveness and robustness of our
aIlgorithms.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

A new model selection criterion, termed as the “quasi-likelihood under the independence model criterion” (QIC), was proposed by Pan (2001) for GEE models. Cui (2007) developed a general computing program to implement the QIC method for a range of statistical distributions. However, only a special case of the negative binomial distribution was considered in Cui (2007), where the dispersion parameter equals to unity. This article introduces a new computing program that can be applied for the general negative binomial model, where the dispersion parameter can be any fixed value. An example is also given in this article.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The current work used discrete event simulation techniques to model the economics of quality within an actual automotive stamping plant. Automotive stamping is a complex, capital intensive process requiring part-specific tooling and specialised machinery. Quality control and quality improvement is difficult in the stamping environment due to the general lack of process understanding and the large number to interacting variables. These factors have prevented the widespread use of statistical process control. In this work, a model of the quality control techniques used at the Ford Geelong Stamping plant is developed and indirectly validated against results from production. To date, most discrete event models are of systems where the quality control process is clearly defined by the rules of statistical process control. However, the quality control technique used within the stamping plant is for the operator to perform a 100% visual inspection while unloading the finished panels. In the developed model, control is enacted after a cumulative count of defective items is observed, thereby approximating the operator who allows a number of defective panels to accumulate before resetting the line. Analysis of this model found that the cost sensitivity to inspection error is dependent upon the level of control and that the level of control determines line utilisation. Additional analysis of this model demonstrated that additional inspection processes would lead to more stable cost structures but these structures many not necessarily be lower cost. The model was subsequently applied to investigate the economics of quality improvement. The quality problem of panel blemishes, induced by slivers (small metal fragments), was chosen as a case stuffy. Errors of 20-30% were observed during direct validation of the cost model and it was concluded that the use of discrete event simulation models for applications requiring high accuracy would not be possible unless the production system was of low complexity. However, the model could be used to evaluate the sensitivity of input factors and investigating the effects of a number of potential improvement opportunities. Therefore, the research concluded that it is possible to use discrete event simulation to determine the quality economics of an actual stamping plant. However, limitations imposed by inability of the model to consider a number of external factors, such as continuous improvement, operator working conditions or wear and the lack of reliable quality data, result in low cost accuracy. Despite this, it still can be demonstrated that discrete event simulation has significant benefits over the alternate modelling methods.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The overarching goal of this dissertation was to evaluate the contextual components of instructional strategies for the acquisition of complex programming concepts. A meta-knowledge processing model is proposed, on the basis of the research findings, thereby facilitating the selection of media treatment for electronic courseware. When implemented, this model extends the work of Smith (1998), as a front-end methodology, for his glass-box interpreter called Bradman, for teaching novice programmers. Technology now provides the means to produce individualized instructional packages with relative ease. Multimedia and Web courseware development accentuate a highly graphical (or visual) approach to instructional formats. Typically, little consideration is given to the effectiveness of screen-based visual stimuli, and curiously, students are expected to be visually literate, despite the complexity of human-computer interaction. Visual literacy is much harder for some people to acquire than for others! (see Chapter Four: Conditions-of-the-Learner) An innovative research programme was devised to investigate the interactive effect of instructional strategies, enhanced with text-plus-textual metaphors or text-plus-graphical metaphors, and cognitive style, on the acquisition of a special category of abstract (process) programming concept. This type of concept was chosen to focus on the role of analogic knowledge involved in computer programming. The results are discussed within the context of the internal/external exchange process, drawing on Ritchey's (1980) concepts of within-item and between-item encoding elaborations. The methodology developed for the doctoral project integrates earlier research knowledge in a novel, interdisciplinary, conceptual framework, including: from instructional science in the USA, for the concept learning models; British cognitive psychology and human memory research, for defining the cognitive style construct; and Australian educational research, to provide the measurement tools for instructional outcomes. The experimental design consisted of a screening test to determine cognitive style, a pretest to determine prior domain knowledge in abstract programming knowledge elements, the instruction period, and a post-test to measure improved performance. This research design provides a three-level discovery process to articulate: 1) the fusion of strategic knowledge required by the novice learner for dealing with contexts within instructional strategies 2) acquisition of knowledge using measurable instructional outcome and learner characteristics 3) knowledge of the innate environmental factors which influence the instructional outcomes This research has successfully identified the interactive effect of instructional strategy, within an individual's cognitive style construct, in their acquisition of complex programming concepts. However, the significance of the three-level discovery process lies in the scope of the methodology to inform the design of a meta-knowledge processing model for instructional science. Firstly, the British cognitive style testing procedure, is a low cost, user friendly, computer application that effectively measures an individual's position on the two cognitive style continua (Riding & Cheema,1991). Secondly, the QUEST Interactive Test Analysis System (Izard,1995), allows for a probabilistic determination of an individual's knowledge level, relative to other participants, and relative to test-item difficulties. Test-items can be related to skill levels, and consequently, can be used by instructional scientists to measure knowledge acquisition. Finally, an Effect Size Analysis (Cohen,1977) allows for a direct comparison between treatment groups, giving a statistical measurement of how large an effect the independent variables have on the dependent outcomes. Combined with QUEST's hierarchical positioning of participants, this tool can assist in identifying preferred learning conditions for the evaluation of treatment groups. By combining these three assessment analysis tools into instructional research, a computerized learning shell, customised for individuals' cognitive constructs can be created (McKay & Garner,1999). While this approach has widespread application, individual researchers/trainers would nonetheless, need to validate with an extensive pilot study programme (McKay,1999a; McKay,1999b), the interactive effects within their specific learning domain. Furthermore, the instructional material does not need to be limited to a textual/graphical comparison, but could be applied to any two or more instructional treatments of any kind. For instance: a structured versus exploratory strategy. The possibilities and combinations are believed to be endless, provided the focus is maintained on linking of the front-end identification of cognitive style with an improved performance outcome. My in-depth analysis provides a better understanding of the interactive effects of the cognitive style construct and instructional format on the acquisition of abstract concepts, involving spatial relations and logical reasoning. In providing the basis for a meta-knowledge processing model, this research is expected to be of interest to educators, cognitive psychologists, communications engineers and computer scientists specialising in computer-human interactions.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The Generalized Estimating Equations (GEE) method is one of the most commonly used statistical methods for the analysis of longitudinal data in epidemiological studies. A working correlation structure for the repeated measures of the outcome variable of a subject needs to be specified by this method. However, statistical criteria for selecting the best correlation structure and the best subset of explanatory variables in GEE are only available recently because the GEE method is developed on the basis of quasi-likelihood theory. Maximum likelihood based model selection methods, such as the widely used Akaike Information Criterion (AIC), are not applicable to GEE directly. Pan (2001) proposed a selection method called QIC which can be used to select the best correlation structure and the best subset of explanatory variables. Based on the QIC method, we developed a computing program to calculate the QIC value for a range of different distributions, link functions and correlation structures. This program was written in Stata software. In this article, we introduce this program and demonstrate how to use it to select the most parsimonious model in GEE analyses of longitudinal data through several representative examples.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Recent developments in ecological statistics have reached behavioral ecology, and an increasing number of studies now apply analytical tools that incorporate alternatives to the conventional null hypothesis testing based on significance levels. However, these approaches continue to receive mixed support in our field. Because our statistical choices can influence research design and the interpretation of data, there is a compelling case for reaching consensus on statistical philosophy and practice. Here, we provide a brief overview of the recently proposed approaches and open an online forum for future discussion (https://bestat.ecoinformatics.org/). From the perspective of practicing behavioral ecologists relying on either correlative or experimental data, we review the most relevant features of information theoretic approaches, Bayesian inference, and effect size statistics. We also discuss concerns about data quality, missing data, and repeatability. We emphasize the necessity of moving away from a heavy reliance on statistical significance while focusing attention on biological relevance and effect sizes, with the recognition that uncertainty is an inherent feature of biological data. Furthermore, we point to the importance of integrating previous knowledge in the current analysis, for which novel approaches offer a variety of tools. We note, however, that the drawbacks and benefits of these approaches have yet to be carefully examined in association with behavioral data. Therefore, we encourage a philosophical change in the interpretation of statistical outcomes, whereas we still retain a pluralistic perspective for making objective statistical choices given the uncertainties around different approaches in behavioral ecology. We provide recommendations on how these concepts could be made apparent in the presentation of statistical outputs in scientific papers.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Background elimination models are widely used in motion tracking systems. Our aim is to develop a system that performs reliably under adverse lighting conditions. In particular, this includes indoor scenes lit partly or entirely by diffuse natural light. We present a modified "median value" model in which the detection threshold adapts to global changes in illumination. The responses of several models are compared, demonstrating the effectiveness of the new model.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Nitric oxide is implicated in the pathogenesis of various neuropathologies characterized by oxidative stress. Although nitric oxide has been reported to be involved in the exacerbation of oxidative stress observed in several neuropathologies, existent data fail to provide a holistic description of how nitrergic pathobiology elicits neuronal injury. Here we provide a comprehensive description of mechanisms contributing to nitric oxide induced neuronal injury by global transcriptomic profiling. Microarray analyses were undertaken on RNA from murine primary cortical neurons treated with the nitric oxide generator DETA-NONOate (NOC-18, 0.5 mM) for 8–24 hrs. Biological pathway analysis focused upon 3672 gene probes which demonstrated at least a ±1.5-fold expression in a minimum of one out of three time-points and passed statistical analysis (one-way anova, P < 0.05). Numerous enriched processes potentially determining nitric oxide mediated neuronal injury were identified from the transcriptomic profile: cell death, developmental growth and survival, cell cycle, calcium ion homeostasis, endoplasmic reticulum stress, oxidative stress, mitochondrial homeostasis, ubiquitin-mediated proteolysis, and GSH and nitric oxide metabolism. Our detailed time-course study of nitric oxide induced neuronal injury allowed us to provide the first time a holistic description of the temporal sequence of cellular events contributing to nitrergic injury. These data form a foundation for the development of screening platforms and define targets for intervention in nitric oxide neuropathologies where nitric oxide mediated injury is causative.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

We propose a novel framework for large-scale scene understanding in static camera surveillance. Our techniques combine fast rank-1 constrained robust PCA to compute the foreground, with non-parametric Bayesian models for inference. Clusters are extracted in foreground patterns using a joint multinomial+Gaussian Dirichlet process model (DPM). Since the multinomial distribution is normalized, the Gaussian mixture distinguishes between similar spatial patterns but different activity levels (eg. car vs bike). We propose a modification of the decayed MCMC technique for incremental inference, providing the ability to discover theoretically unlimited patterns in unbounded video streams. A promising by-product of our framework is online, abnormal activity detection. A benchmark video and two surveillance videos, with the longest being 140 hours long are used in our experiments. The patterns discovered are as informative as existing scene understanding algorithms. However, unlike existing work, we achieve near real-time execution and encouraging performance in abnormal activity detection.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Previous experience and research indicated that the Pareto Principle (80/20 Principle) has been widely used in many industries to achieve more with less. The study described in this paper concurs that this principle can be applied to improve the estimating accuracy and efficiency, especially in design development stage of projects. In fact, establishing an effective cost estimating model to improve accuracy and efficiency in design development stage has been a subject, which has attracted many research attentions over several decades. For over almost 40 years, research studies indicate that using the 80/20 Principle is one of the approaches. However, most of these studies were built by assumption, theoretical analysis or questionnaire survey. The objective of this research is to explore a logical and systematic method to establish a cost estimating model based on the Pareto Principle. This paper includes extensive literatures review on cost estimating accuracy and efficiency in the construction industry that points out the current gap of knowledge area and understanding of the topical. These reviews assist in developing the direction for the research and explore the potential methodology of using the Pareto Principle in the new cost estimating model. The findings of this paper suggest that combining the Pareto Principle with statistical analysis could be used as the technique to improve the accuracy and efficiency of current estimating methods in design development stage.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This article describes the implementation of machine learning techniques that assist cycling experts in the crucial decision-making processes for athlete selection and strategic planning in the track cycling omnium. The omnium is a multi-event competition that was included in the Olympic Games for the first time in 2012. Presently, selectors and cycling coaches make decisions based on experience and opinion. They rarely have access to knowledge that helps predict athletic performances. The omnium presents a unique and complex decision-making challenge as it is not clear what type of athlete is best suited to the omnium (e.g., sprint or endurance specialist) and tactical decisions made by the coach and athlete during the event will have significant effects on the overall performance of the athlete. In the present work, a variety of machine learning techniques were used to analyze omnium competition data from the World Championships since 2007. The analysis indicates that sprint events have slightly more influence in determining the medalists, than endurance-based events. Using a probabilistic analysis, we created a model of performance prediction that provides an unprecedented level of supporting information that assists coaches with strategic and tactical decisions during the omnium.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

A study on the pedestrian's steering behaviour through a built environment in normal circumstances is presented in this paper. The study focuses on the relationship between the environment and the pedestrian's walking trajectory. Owing to the ambiguity and vagueness of the relationship between the pedestrians and the surrounding environment, a genetic fuzzy system is proposed for modelling and simulation of the pedestrian's walking trajectory confronting the environmental stimuli. We apply the genetic algorithm to search for the optimum membership function parameters of the fuzzy model. The proposed system receives the pedestrian's perceived stimuli from the environment as the inputs, and provides the angular change of direction in each step as the output. The environmental stimuli are quantified using the Helbing social force model. Attractive and repulsive forces within the environment represent various environmental stimuli that influence the pedestrian's walking trajectory at each point of the space. To evaluate the effectiveness of the proposed model, three experiments are conducted. The first experimental results are validated against real walking trajectories of participants within a corridor. The second and third experimental results are validated against simulated walking trajectories collected from the AnyLogic® software. Analysis and statistical measurement of the results indicate that the genetic fuzzy system with optimised membership functions produces more accurate and stable prediction of heterogeneous pedestrians' walking trajectories than those from the original fuzzy model. © 2014 Elsevier B.V. All rights reserved.