960 resultados para cumulative sum (CUSUM)


Relevância:

100.00% 100.00%

Publicador:

Resumo:

In previous research (Chung et al., 2009), the potential of the continuous risk profile (CRP) to proactively detect the systematic deterioration of freeway safety levels was presented. In this paper, this potential is investigated further, and an algorithm is proposed for proactively detecting sites where the collision rate is not sufficiently high to be classified as a high collision concentration location but where a systematic deterioration of safety level is observed. The approach proposed compares the weighted CRP across different years and uses the cumulative sum (CUSUM) algorithm to detect the sites where changes in collision rate are observed. The CRPs of the detected sites are then compared for reproducibility. When high reproducibility is observed, a growth factor is used for sequential hypothesis testing to determine if the collision profiles are increasing over time. Findings from applying the proposed method using empirical data are documented in the paper together with a detailed description of the method.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The quick detection of abrupt (unknown) parameter changes in an observed hidden Markov model (HMM) is important in several applications. Motivated by the recent application of relative entropy concepts in the robust sequential change detection problem (and the related model selection problem), this paper proposes a sequential unknown change detection algorithm based on a relative entropy based HMM parameter estimator. Our proposed approach is able to overcome the lack of knowledge of post-change parameters, and is illustrated to have similar performance to the popular cumulative sum (CUSUM) algorithm (which requires knowledge of the post-change parameter values) when examined, on both simulated and real data, in a vision-based aircraft manoeuvre detection problem.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The quick detection of an abrupt unknown change in the conditional distribution of a dependent stochastic process has numerous applications. In this paper, we pose a minimax robust quickest change detection problem for cases where there is uncertainty about the post-change conditional distribution. Our minimax robust formulation is based on the popular Lorden criteria of optimal quickest change detection. Under a condition on the set of possible post-change distributions, we show that the widely known cumulative sum (CUSUM) rule is asymptotically minimax robust under our Lorden minimax robust formulation as a false alarm constraint becomes more strict. We also establish general asymptotic bounds on the detection delay of misspecified CUSUM rules (i.e. CUSUM rules that are designed with post- change distributions that differ from those of the observed sequence). We exploit these bounds to compare the delay performance of asymptotically minimax robust, asymptotically optimal, and other misspecified CUSUM rules. In simulation examples, we illustrate that asymptotically minimax robust CUSUM rules can provide better detection delay performance at greatly reduced computation effort compared to competing generalised likelihood ratio procedures.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The abundance of many commercially important fish stocks are declining and this has led to widespread concern on the performance of traditional approach in fisheries management. Quantitative models are used for obtaining estimates of population abundance and the management advice is based on annual harvest levels (TAC), where only a certain amount of catch is allowed from specific fish stocks. However, these models are data intensive and less useful when stocks have limited historical information. This study examined whether empirical stock indicators can be used to manage fisheries. The relationship between indicators and the underlying stock abundance is not direct and hence can be affected by disturbances that may account for both transient and persistent effects. Methods from Statistical Process Control (SPC) theory such as the Cumulative Sum (CUSUM) control charts are useful in classifying these effects and hence they can be used to trigger management response only when a significant impact occurs to the stock biomass. This thesis explores how empirical indicators along with CUSUM can be used for monitoring, assessment and management of fish stocks. I begin my thesis by exploring various age based catch indicators, to identify those which are potentially useful in tracking the state of fish stocks. The sensitivity and response of these indicators towards changes in Spawning Stock Biomass (SSB) showed that indicators based on age groups that are fully selected to the fishing gear or Large Fish Indicators (LFIs) are most useful and robust across the range of scenarios considered. The Decision-Interval (DI-CUSUM) and Self-Starting (SS-CUSUM) forms are the two types of control charts used in this study. In contrast to the DI-CUSUM, the SS-CUSUM can be initiated without specifying a target reference point (‘control mean’) to detect out-of-control (significant impact) situations. The sensitivity and specificity of SS-CUSUM showed that the performances are robust when LFIs are used. Once an out-of-control situation is detected, the next step is to determine how much shift has occurred in the underlying stock biomass. If an estimate of this shift is available, they can be used to update TAC by incorporation into Harvest Control Rules (HCRs). Various methods from Engineering Process Control (EPC) theory were tested to determine which method can measure the shift size in stock biomass with the highest accuracy. Results showed that methods based on Grubb’s harmonic rule gave reliable shift size estimates. The accuracy of these estimates can be improved by monitoring a combined indicator metric of stock-recruitment and LFI because this may account for impacts independent of fishing. The procedure of integrating both SPC and EPC is known as Statistical Process Adjustment (SPA). A HCR based on SPA was designed for DI-CUSUM and the scheme was successful in bringing out-of-control fish stocks back to its in-control state. The HCR was also tested using SS-CUSUM in the context of data poor fish stocks. Results showed that the scheme will be useful for sustaining the initial in-control state of the fish stock until more observations become available for quantitative assessments.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Climatic variability on the European Continental Shelf is dominated by events over the North Atlantic Ocean, and in particular by the North Atlantic Oscillation (NAO). The NAO is essentially a winter phenomenon, and its effects will be felt most strongly by populations for which winter conditions are critical. One example is the copepod Calanus finmarchicus, whose northern North Sea populations overwinter at depth in the North Atlantic. Its annual abundance in this region is strongly dependent on water transports at the end of the winter, and hence on the NAO index. Variations in the NAO give rise to changes in the circulation of the North Atlantic Ocean, with additional perturbations arising from El Ni (n) over tildeo - Southern Oscillation (ENSO) events in the Pacific, and these changes can be delayed by several years because of the adjustment time of the ocean circulation. One measure of the circulation is the latitude of the north wall of the Gulf Stream (GSNW index). Interannual variations in the plankton of the Shelf Seas show strong correlations with the fluctuations of the GSNW index, which are the result of Atlantic-wide atmospheric processes. These associations imply that the interannual variations are climatically induced rather than due to natural fluctuations of the marine ecosystem, and that the zooplankton populations have not been significantly affected by anthropogenic processes such as nutrient enrichment or fishing pressure. While the GSNW index represents a response to atmospheric changes over two or more years, the zooplankton populations correlated with it have generation times of a few weeks. The simplest explanation for the associations between the zooplankton and the GSNW index is that the plankton are responding to weather patterns propagating downstream from the Gulf Stream system. It seems that these meteorological processes operate in the spring. Although it has been suggested that there was a regime shift in the North Sea in the late 1980s, examination of the time-series by the cumulative sum (CUSUM) technique shows that any changes in the zooplankton of the central and northern North Sea are consistent with the background climatic variability. The abundance of total copepods increased during this period but this change does not represent a dramatic change in ecosystem processes. It is possible some change may have occurred at the end of the time-series in the years 1997 and 1998.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

BACKGROUND: Robotic-assisted laparoscopic surgery (RALS) is evolving as an important surgical approach in the field of colorectal surgery. We aimed to evaluate the learning curve for RALS procedures involving resections of the rectum and rectosigmoid. METHODS: A series of 50 consecutive RALS procedures were performed between August 2008 and September 2009. Data were entered into a retrospective database and later abstracted for analysis. The surgical procedures included abdominoperineal resection (APR), anterior rectosigmoidectomy (AR), low anterior resection (LAR), and rectopexy (RP). Demographic data and intraoperative parameters including docking time (DT), surgeon console time (SCT), and total operative time (OT) were analyzed. The learning curve was evaluated using the cumulative sum (CUSUM) method. RESULTS: The procedures performed for 50 patients (54% male) included 25 AR (50%), 15 LAR (30%), 6 APR (12%), and 4 RP (8%). The mean age of the patients was 54.4 years, the mean BMI was 27.8 kg/m(2), and the median American Society of Anesthesiologists (ASA) classification was 2. The series had a mean DT of 14 min, a mean SCT of 115.1 min, and a mean OT of 246.1 min. The DT and SCT accounted for 6.3% and 46.8% of the OT, respectively. The SCT learning curve was analyzed. The CUSUM(SCT) learning curve was best modeled as a parabola, with equation CUSUM(SCT) in minutes equal to 0.73 × case number(2) - 31.54 × case number - 107.72 (R = 0.93). The learning curve consisted of three unique phases: phase 1 (the initial 15 cases), phase 2 (the middle 10 cases), and phase 3 (the subsequent cases). Phase 1 represented the initial learning curve, which spanned 15 cases. The phase 2 plateau represented increased competence with the robotic technology. Phase 3 was achieved after 25 cases and represented the mastery phase in which more challenging cases were managed. CONCLUSIONS: The three phases identified with CUSUM analysis of surgeon console time represented characteristic stages of the learning curve for robotic colorectal procedures. The data suggest that the learning phase was achieved after 15 to 25 cases.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The challenge of detecting a change in the distribution of data is a sequential decision problem that is relevant to many engineering solutions, including quality control and machine and process monitoring. This dissertation develops techniques for exact solution of change-detection problems with discrete time and discrete observations. Change-detection problems are classified as Bayes or minimax based on the availability of information on the change-time distribution. A Bayes optimal solution uses prior information about the distribution of the change time to minimize the expected cost, whereas a minimax optimal solution minimizes the cost under the worst-case change-time distribution. Both types of problems are addressed. The most important result of the dissertation is the development of a polynomial-time algorithm for the solution of important classes of Markov Bayes change-detection problems. Existing techniques for epsilon-exact solution of partially observable Markov decision processes have complexity exponential in the number of observation symbols. A new algorithm, called constellation induction, exploits the concavity and Lipschitz continuity of the value function, and has complexity polynomial in the number of observation symbols. It is shown that change-detection problems with a geometric change-time distribution and identically- and independently-distributed observations before and after the change are solvable in polynomial time. Also, change-detection problems on hidden Markov models with a fixed number of recurrent states are solvable in polynomial time. A detailed implementation and analysis of the constellation-induction algorithm are provided. Exact solution methods are also established for several types of minimax change-detection problems. Finite-horizon problems with arbitrary observation distributions are modeled as extensive-form games and solved using linear programs. Infinite-horizon problems with linear penalty for detection delay and identically- and independently-distributed observations can be solved in polynomial time via epsilon-optimal parameterization of a cumulative-sum procedure. Finally, the properties of policies for change-detection problems are described and analyzed. Simple classes of formal languages are shown to be sufficient for epsilon-exact solution of change-detection problems, and methods for finding minimally sized policy representations are described.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

This thesis is concerned with change point analysis for time series, i.e. with detection of structural breaks in time-ordered, random data. This long-standing research field regained popularity over the last few years and is still undergoing, as statistical analysis in general, a transformation to high-dimensional problems. We focus on the fundamental »change in the mean« problem and provide extensions of the classical non-parametric Darling-Erdős-type cumulative sum (CUSUM) testing and estimation theory within highdimensional Hilbert space settings. In the first part we contribute to (long run) principal component based testing methods for Hilbert space valued time series under a rather broad (abrupt, epidemic, gradual, multiple) change setting and under dependence. For the dependence structure we consider either traditional m-dependence assumptions or more recently developed m-approximability conditions which cover, e.g., MA, AR and ARCH models. We derive Gumbel and Brownian bridge type approximations of the distribution of the test statistic under the null hypothesis of no change and consistency conditions under the alternative. A new formulation of the test statistic using projections on subspaces allows us to simplify the standard proof techniques and to weaken common assumptions on the covariance structure. Furthermore, we propose to adjust the principal components by an implicit estimation of a (possible) change direction. This approach adds flexibility to projection based methods, weakens typical technical conditions and provides better consistency properties under the alternative. In the second part we contribute to estimation methods for common changes in the means of panels of Hilbert space valued time series. We analyze weighted CUSUM estimates within a recently proposed »high-dimensional low sample size (HDLSS)« framework, where the sample size is fixed but the number of panels increases. We derive sharp conditions on »pointwise asymptotic accuracy« or »uniform asymptotic accuracy« of those estimates in terms of the weighting function. Particularly, we prove that a covariance-based correction of Darling-Erdős-type CUSUM estimates is required to guarantee uniform asymptotic accuracy under moderate dependence conditions within panels and that these conditions are fulfilled, e.g., by any MA(1) time series. As a counterexample we show that for AR(1) time series, close to the non-stationary case, the dependence is too strong and uniform asymptotic accuracy cannot be ensured. Finally, we conduct simulations to demonstrate that our results are practically applicable and that our methodological suggestions are advantageous.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Change point estimation is recognized as an essential tool of root cause analyses within quality control programs as it enables clinical experts to search for potential causes of change in hospital outcomes more effectively. In this paper, we consider estimation of the time when a linear trend disturbance has occurred in survival time following an in-control clinical intervention in the presence of variable patient mix. To model the process and change point, a linear trend in the survival time of patients who underwent cardiac surgery is formulated using hierarchical models in a Bayesian framework. The data are right censored since the monitoring is conducted over a limited follow-up period. We capture the effect of risk factors prior to the surgery using a Weibull accelerated failure time regression model. We use Markov Chain Monte Carlo to obtain posterior distributions of the change point parameters including the location and the slope size of the trend and also corresponding probabilistic intervals and inferences. The performance of the Bayesian estimator is investigated through simulations and the result shows that precise estimates can be obtained when they are used in conjunction with the risk-adjusted survival time cumulative sum control chart (CUSUM) control charts for different trend scenarios. In comparison with the alternatives, step change point model and built-in CUSUM estimator, more accurate and precise estimates are obtained by the proposed Bayesian estimator over linear trends. These superiorities are enhanced when probability quantification, flexibility and generalizability of the Bayesian change point detection model are also considered.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

This paper considers cooperative spectrum sensing in Cognitive Radios. In our previous work we have developed DualSPRT, a distributed algorithm for cooperative spectrum sensing using Sequential Probability Ratio Test (SPRT) at the Cognitive Radios as well as at the fusion center. This algorithm works well, but is not optimal. In this paper we propose an improved algorithm- SPRT-CSPRT, which is motivated from Cumulative Sum Procedures (CUSUM). We analyse it theoretically. We also modify this algorithm to handle uncertainties in SNR's and fading.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Il Test di Risposta Termica (Thermal Response Test-TRT) (Mogenson,1983) è il test esistente con il più alto grado di accuratezza per la caratterizzazione del reservoir geotermico superficiale. Il test consiste in una simulazione in situ del funzionamento di un sistema a circuito chiuso di sfruttamento dell’energia geotermica, per un periodo limitato di tempo, attraverso l’iniezione o estrazione di calore a potenza costante all’interno del geo-scambiatore (Borehole Heat Exchanger-BHE). Dall’analisi della variazione delle temperature del fluido circolante, è possibile avere una stima delle proprietà termiche medie del volume del reservoir geotermico interessato dal test. Le grandezze principali per la caratterizzazione di un serbatoio geotermico sono la conduttività termica (λ), la capacità termica volumetrica (c), la temperatura indisturbata del suolo (Tg) e la resistenza termica del pozzo (Rb); la loro determinazione è necessaria per il corretto progettazione degli geo-scambiatori. I risultati del TRT sono tuttavia sensibili alle condizioni al contorno spazio-temporali quali ad es.: variazione della temperatura del terreno, movimento d’acqua di falda, condizioni metereologiche, eventi stagionali, ecc. Questo lavoro vuole: i) introdurre uno studio sui problemi di caratterizzazione del reservoir geotermico superficiale, in particolare analizzando l’effetto che il movimento d’acqua di falda ha sui parametri termici; ii) analizzare la sensitività dei risultati del test alle variabilità dei parametri caratteristici del funzionamento delle attrezzature. Parte del lavoro della mia tesi è stata svolta in azienda per un periodo di 4 mesi presso la “Groenholland Geo Energy systems” che ha sede ad Amsterdam in Olanda. Tre diversi esperimenti sono stati realizzati sullo stesso sito (stratigrafia nota del terreno: argilla, sabbia fine e sabbia grossa) usando una sonda profonda 30 metri e diversi pozzi per l’estrazione d’acqua e per monitorare gli effetti in prossimità del geo scambiatore. I risultati degli esperimenti sono stati molto diversi tra di loro, non solo in termini di dati registrati (temperature del fluido termovettore), ma in termini dei valori dei parametri ottenuti elaborando i dati. In particolare non è sufficiente adottare il modello classico della sorgente lineare infinita (Infinite Line Source Solution- ILS) (Ingersoll and Plass, 1948), il quale descrive il trasferimento di calore per conduzione in un mezzo omogeneo indefinito a temperatura costante. Infatti, lo scambio di calore avviene anche tramite convezione causata dal movimento d’acqua di falda, non identificabile mediante gli approcci classici tipo CUSUM test (Cumulative Sum test) (Brown e altri,1975) Lo studio della tesi vuole dare un quadro di riferimento per correlare la variabilità dei risultati con la variabilità delle condizioni al contorno. L’analisi integra le metodologie classiche (ILS) con un approccio geostatistico utile a comprendere i fenomeni e fluttuazioni che caratterizzano il test. Lo studio delle principali variabili e parametri del test, quali temperatura in ingresso e uscita del fluido termovettore, portata del fluido e potenza iniettata o estratta, è stato sviluppato mediante: il variogramma temporale, ovvero la semivarianza dell’accrescimento, che esprime il tipo di autocorrelazione temporale della variabile in esame; la covarianza incrociata temporale, ovvero la covarianza fra due variabili del sistema, che ne definisce quantitativamente il grado di correlazione in funzionamento del loro sfasamento temporale. L’approccio geostatistico proposto considera la temperatura del fluido Tf come una funzione aleatoria (FA) non stazionaria nel tempo (Chiles, 1999), il cui trend è formalmente definito, ma deve essere identificato numericamente. Si considera quindi un classico modello a residuo; in cui la FA è modellizzata come la somma di un termine deterministico, la media (il valore atteso) m(t),coincidente col modello descritto dalla teoria della sorgente lineare infinità, e di un termine aleatorio, la fluttuazione, Y(t). Le variabili portata e potenza sono invece considerate delle funzioni aleatorie stazionarie nel tempo, ovvero a media costante. Da questo studio di Tesi si sono raggiunte delle conclusioni molto importanti per lo studio del TRT: Confronto tra gli esperimenti in estrazione di calore, con e senza movimento d’acqua di falda: si studia l’effetto indotto dalla falda sul TRT. E’ possibile caratterizzare quantitativamente l’incremento della conducibilità termica equivalente legata a fenomeni convettivi dovuti al movimento d’acqua di falda. Inoltre, i variogrammi sperimentali evidenziano periodicità simili nei due casi e legate al funzionamento della pompa di calore e della componentistica associata ed alla circolazione del fluido termovettore all’interno della sonda. Tuttavia, la componente advettiva ha un effetto di smorzamento sulle piccole periodicità dei variogrammi, ma di aumento dell’ampiezza delle periodicità maggiori a causa del funzionamento della pompa di calore che deve fornire maggiore energia al sistema per bilanciare le dispersioni dovute al movimento d’acqua di falda. Confronto fra estrazione ed iniezione di calore, con movimento d’acqua di falda: si studia la significatività dei risultati nei due casi. L’analisi delle variografie evidenzia significative differenze nella struttura dei variogrammi sperimentali. In particolare, nel test con iniezione di calore i variogrammi sperimentali delle temperature hanno valori sistematicamente inferiori, circostanza che assicura una migliore precisione nella stima dei parametri termici. Quindi eseguire il TRT in iniezione di calore risulta più preciso. Dall’analisi dei variogrammi sperimentali delle singole variabili quali temperatura del fluido in ingresso e uscita all’interno del geoscambiatore è stato confermato il fenomeno di smorzamento delle oscillazioni da parte del terreno. Dall’analisi delle singole variabili del test (temperature, potenza, portata) è stata confermata l’indipendenza temporale fra portate e temperature. Ciò è evidenziato dalle diverse strutture dei variogrammi diretti e dalle covarianze incrociate prossime a zero. Mediante correlogrami è stato dimostrato la possibilità di calcolare il tempo impiegato dal fluido termovettore per circolare all’interno della sonda. L’analisi geostatistica ha permesso quindi di studiare in dettaglio la sensitività dei risultati del TRT alle diverse condizioni al contorno, quelle legate al reservoir e quelle legate al funzionamento delle attrezzature

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Today’s evolving networks are experiencing a large number of different attacks ranging from system break-ins, infection from automatic attack tools such as worms, viruses, trojan horses and denial of service (DoS). One important aspect of such attacks is that they are often indiscriminate and target Internet addresses without regard to whether they are bona fide allocated or not. Due to the absence of any advertised host services the traffic observed on unused IP addresses is by definition unsolicited and likely to be either opportunistic or malicious. The analysis of large repositories of such traffic can be used to extract useful information about both ongoing and new attack patterns and unearth unusual attack behaviors. However, such an analysis is difficult due to the size and nature of the collected traffic on unused address spaces. In this dissertation, we present a network traffic analysis technique which uses traffic collected from unused address spaces and relies on the statistical properties of the collected traffic, in order to accurately and quickly detect new and ongoing network anomalies. Detection of network anomalies is based on the concept that an anomalous activity usually transforms the network parameters in such a way that their statistical properties no longer remain constant, resulting in abrupt changes. In this dissertation, we use sequential analysis techniques to identify changes in the behavior of network traffic targeting unused address spaces to unveil both ongoing and new attack patterns. Specifically, we have developed a dynamic sliding window based non-parametric cumulative sum change detection techniques for identification of changes in network traffic. Furthermore we have introduced dynamic thresholds to detect changes in network traffic behavior and also detect when a particular change has ended. Experimental results are presented that demonstrate the operational effectiveness and efficiency of the proposed approach, using both synthetically generated datasets and real network traces collected from a dedicated block of unused IP addresses.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

High-rate flooding attacks (aka Distributed Denial of Service or DDoS attacks) continue to constitute a pernicious threat within the Internet domain. In this work we demonstrate how using packet source IP addresses coupled with a change-point analysis of the rate of arrival of new IP addresses may be sufficient to detect the onset of a high-rate flooding attack. Importantly, minimizing the number of features to be examined, directly addresses the issue of scalability of the detection process to higher network speeds. Using a proof of concept implementation we have shown how pre-onset IP addresses can be efficiently represented using a bit vector and used to modify a “white list” filter in a firewall as part of the mitigation strategy.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Aims: This paper describes the development of a risk adjustment (RA) model predictive of individual lesion treatment failure in percutaneous coronary interventions (PCI) for use in a quality monitoring and improvement program. Methods and results: Prospectively collected data for 3972 consecutive revascularisation procedures (5601 lesions) performed between January 2003 and September 2011 were studied. Data on procedures to September 2009 (n = 3100) were used to identify factors predictive of lesion treatment failure. Factors identified included lesion risk class (p < 0.001), occlusion type (p < 0.001), patient age (p = 0.001), vessel system (p < 0.04), vessel diameter (p < 0.001), unstable angina (p = 0.003) and presence of major cardiac risk factors (p = 0.01). A Bayesian RA model was built using these factors with predictive performance of the model tested on the remaining procedures (area under the receiver operating curve: 0.765, Hosmer–Lemeshow p value: 0.11). Cumulative sum, exponentially weighted moving average and funnel plots were constructed using the RA model and subjectively evaluated. Conclusion: A RA model was developed and applied to SPC monitoring for lesion failure in a PCI database. If linked to appropriate quality improvement governance response protocols, SPC using this RA tool might improve quality control and risk management by identifying variation in performance based on a comparison of observed and expected outcomes.