931 resultados para Real-time PCR
Resumo:
Genetic Programming (GP) is a widely used methodology for solving various computational problems. GP's problem solving ability is usually hindered by its long execution times. In this thesis, GP is applied toward real-time computer vision. In particular, object classification and tracking using a parallel GP system is discussed. First, a study of suitable GP languages for object classification is presented. Two main GP approaches for visual pattern classification, namely the block-classifiers and the pixel-classifiers, were studied. Results showed that the pixel-classifiers generally performed better. Using these results, a suitable language was selected for the real-time implementation. Synthetic video data was used in the experiments. The goal of the experiments was to evolve a unique classifier for each texture pattern that existed in the video. The experiments revealed that the system was capable of correctly tracking the textures in the video. The performance of the system was on-par with real-time requirements.
Resumo:
By the end of 2004, the Canadian swine population had experienced a severe 2 increase in the incidence of Porcine circovirus-associated disease (PCVAD), a problem that was 3 associated with the emergence of a new Porcine circovirus-2 genotype (PCV-2b), previously 4 unrecovered in North America. Thus it became important to develop a diagnostic tool that could 5 differentiate between the old and new circulating genotypes (PCV-2a and -2b, respectively). 6 Consequently, a multiplex real-time quantitative polymerase chain reaction (mrtqPCR) assay that 7 could sensitively and specifically identify and differentiate PCV-2 genotypes was developed. A 8 retrospective epidemiological survey that used the mrtqPCR assay was performed to determine if 9 cofactors could affect the risk of PCVAD. From 121 PCV-2–positive cases gathered for this 10 study, 4.13%, 92.56% and 3.31% were positive for PCV-2a, PCV-2b, and both genotypes, 11 respectively. In a data analysis using univariate logistic regressions, PCVAD compatible 12 (PCVAD/c) score was significantly associated with the presence of Porcine reproductive and 13 respiratory syndrome virus (PRRSV), PRRSV viral load, PCV-2 viral load, and PCV-2 14 immunohistochemistry (IHC) results. Polytomous logistic regression analysis revealed that 15 PCVAD/c score was affected by PCV-2 viral load (P = 0.0161) and IHC (P = 0.0128), but not by 16 the PRRSV variables (P > 0.9); suggesting that mrtqPCR in tissue is a reliable alternative to IHC. 17 Logistic regression analyses revealed that PCV-2 increased the odds ratio of isolating 2 major 18 swine pathogens of the respiratory tract, Actinobacillus pleuropneumoniae and Streptococcus 19 suis serotypes 1/2, 1, 2, 3, 4, and 7, which are serotypes commonly associated with clinical 20 diseases.
Resumo:
Les champignons mycorhizien à arbuscules (CMA) sont des organismes pouvant établir des symbioses avec 80% des plantes terrestres. Les avantages d'une telle symbiose sont de plus en plus caractérisés et exploités en agriculture. Par contre, jusqu'à maintenant, il n'existe aucun outil permettant à la fois l'identification et la quantification de ces champignons dans le sol de façon fiable et rapide. Un tel outil permettrait, entre autres, de mieux comprendre les dynamiques des populations des endomycorhizes dans le sol. Pour les producteurs d'inoculum mycorhiziens, cela permettrait également d'établir un suivi de leurs produits en champs et d'avoir un contrôle de qualité de plus sur leurs inoculants. C'est ce que nous avons tenté de développer au sein du laboratoire du Dr. Hijri. Depuis environ une trentaine d'années, des outils d'identification et/ou de quantification ont été développés en utilisant les profiles d'acides gras, les isozymes, les anticorps et finalement l'ADN nucléaire. À ce jour, ces méthodes d’identification et de quantification sont soit coûteuses, soit imprécises. Qui plus est, aucune méthode ne permet à la fois la quantification et l’identification de souches particulières de CMA. L’ADN mitochondrial ne présente pas le même polymorphisme de séquence que celui qui rend l’ADN nucléaire impropre à la quantification. C'est pourquoi nous avons analysé les séquences d’ADN mitochondrial et sélectionné les régions caractéristiques de deux espèces de champignons mycorhiziens arbusculaires (CMA). C’est à partir de ces régions que nous avons développé des marqueurs moléculaires sous forme de sondes et d’amorces TaqMan permettant de quantifier le nombre de mitochondries de chacune de ces espèces dans un échantillon d’ADN. Nous avons ensuite tenté de déterminer une unité de quantification des CMA, soit un nombre de mitochondries par spore. C’est alors que nous avons réalisé que la méthode de préparation des échantillons de spores ainsi que la méthode d’extraction d’ADN avaient des effets significatifs sur l’unité de quantification de base. Nous avons donc optimisé ces protocoles, avant d’en e tester l’application sur des échantillons de sol et de racines ayant été inoculés avec chacune des deux espèces cibles. À ce stade, cet outil est toujours semi-quantificatif, mais il permet 9 l’identification précise de deux espèces de CMA compétentes dans des milieux saturés en phosphore inorganique. Ces résultats , en plus d’être prometteurs, ont permis d’augmenter les connaissances méthodologiques reliées à la quantification des CMA dans le sol, et suggèrent qu’à cause de leurs morphologies différentes, l’élaboration d’un protocole de quantification standardisé pour toutes les espèces de CMA demeure un objectif complexe, qui demande de nouvelles études in vivo.
Resumo:
The present research problem is to study the existing encryption methods and to develop a new technique which is performance wise superior to other existing techniques and at the same time can be very well incorporated in the communication channels of Fault Tolerant Hard Real time systems along with existing Error Checking / Error Correcting codes, so that the intention of eaves dropping can be defeated. There are many encryption methods available now. Each method has got it's own merits and demerits. Similarly, many crypt analysis techniques which adversaries use are also available.
Resumo:
Machine tool chatter is an unfavorable phenomenon during metal cutting, which results in heavy vibration of cutting tool. With increase in depth of cut, the cutting regime changes from chatter-free cutting to one with chatter. In this paper, we propose the use of permutation entropy (PE), a conceptually simple and computationally fast measurement to detect the onset of chatter from the time series using sound signal recorded with a unidirectional microphone. PE can efficiently distinguish the regular and complex nature of any signal and extract information about the dynamics of the process by indicating sudden change in its value. Under situations where the data sets are huge and there is no time for preprocessing and fine-tuning, PE can effectively detect dynamical changes of the system. This makes PE an ideal choice for online detection of chatter, which is not possible with other conventional nonlinear methods. In the present study, the variation of PE under two cutting conditions is analyzed. Abrupt variation in the value of PE with increase in depth of cut indicates the onset of chatter vibrations. The results are verified using frequency spectra of the signals and the nonlinear measure, normalized coarse-grained information rate (NCIR).
Resumo:
Department of Computer Applications, Cochin University of Science and Technology
Resumo:
In this paper the effectiveness of a novel method of computer assisted pedicle screw insertion was studied using testing of hypothesis procedure with a sample size of 48. Pattern recognition based on geometric features of markers on the drill has been performed on real time optical video obtained from orthogonally placed CCD cameras. The study reveals the exactness of the calculated position of the drill using navigation based on CT image of the vertebra and real time optical video of the drill. The significance value is 0.424 at 95% confidence level which indicates good precision with a standard mean error of only 0.00724. The virtual vision method is less hazardous to both patient and the surgeon
Resumo:
Femtosecond time-resolved techniques with KETOF (kinetic energy time-of-flight) detection in a molecular beam are developed for studies of the vectorial dynamics of transition states. Application to the dissociation reaction of IHgI is presented. For this system, the complex [I---Hg---I](++)* is unstable and, through the symmetric and asymmetric stretch motions, yields different product fragments: [I---Hg---I](++)* -> HgI(X^2/sigma^+) + I(^2P_3/2) [or I*(^2P_l/2)] (1a); [I---Hg---I](++)* -> Hg(^1S_0) + I(^2P_3/2) + I(^2P_3/2) [or I* (^2P_1/2)] (1 b). These two channels, (1a) and (1b), lead to different kinetic energy distributions in the products. It is shown that the motion of the wave packet in the transition-state region can be observed by MPI mass detection; the transient time ranges from 120 to 300 fs depending on the available energy. With polarized pulses, the vectorial properties (transition moments alignment relative to recoil direction) are studied for fragment separations on the femtosecond time scale. The results indicate the nature of the structure (symmetry properties) and the correlation to final products. For 311-nm excitation, no evidence of crossing between the I and I* potentials is found at the internuclear separations studied. (Results for 287-nm excitation are also presented.) Molecular dynamics simulations and studies by laser-induced fluorescence support these findings.
Resumo:
Real-time studies of the dynamics were performed on the reaction of HgI_2 in a molecular beam. Excitation was by either one or multi pump photons (311 nm), leading to two separate sets of dynamics, each of which could be investigated by a time-delayed probe laser (622 nm) that ionized the parent molecule and the fragments by REMPI processes. These dynamics were distinguished by combining the information from transients taken at each mass (HgI_2, HgI, I_2, Hg, and I) with the results of pump (and probe) power dependence studies on each mass. A method of plotting the slope of the intensity dependence against the pump-probe time delay proved essential. In the preceding publication, we detailed the dynamics of the reaction initiated by a one photon excitation to the A-continuum. Here, we present studies of higher-energy states. Multiphoton excitation accesses predissociative states of HgI_2, for which there are crossings into the symmetric and asymmetric stretch coordinates. The dynamics of these channels, which lead to atomic (I or Hg) and diatomic (HgI) fragments, are discussed and related to the nature of the intermediates along the reaction pathway.
Resumo:
Femtosecond reaction dynamics of OClO in a supersonic molecular beam are reported. The system is excited to the A^2A_2 state with a femtosecond pulse, covering a range of excitation in the symmetric stretch between v_1 = 17 to v_1 = 11 (308-352 nm). A time-delayed femtosecond probe pulse ionizes the OClO, and OClO^+ is detected. This ion has not been observed in previous experiments because of its ultrafast fragmentation. Transients are reported for the mass of the parent OClO as well as the mass of the ClO. Apparent biexponential decays are observed and related to the fragmentation dynamics: OClO+hv \rightarrow (OClO)^{(++)*} \rightarrow ClO+O \rightarrow Cl+O_2. Clusters of OClO with water (OClO)_n (H_2 0)_m with n from 1 to 3 and m from 0 to 3 are also observed. The dynamics of the fragmentation reveal the nuclear motions and the electronic coupling between surfaces. The time scale for bond breakage is in the range of 300-500 fs, depending on v_1; surface crossing to form new intermediates is a pathway for the two channels of fragmentation: ClO+O (primary) and Cl+O_2 (minor). Comparisons with results of ab initio calculations are made.
Resumo:
Die zunehmende Vernetzung der Informations- und Kommunikationssysteme führt zu einer weiteren Erhöhung der Komplexität und damit auch zu einer weiteren Zunahme von Sicherheitslücken. Klassische Schutzmechanismen wie Firewall-Systeme und Anti-Malware-Lösungen bieten schon lange keinen Schutz mehr vor Eindringversuchen in IT-Infrastrukturen. Als ein sehr wirkungsvolles Instrument zum Schutz gegenüber Cyber-Attacken haben sich hierbei die Intrusion Detection Systeme (IDS) etabliert. Solche Systeme sammeln und analysieren Informationen von Netzwerkkomponenten und Rechnern, um ungewöhnliches Verhalten und Sicherheitsverletzungen automatisiert festzustellen. Während signatur-basierte Ansätze nur bereits bekannte Angriffsmuster detektieren können, sind anomalie-basierte IDS auch in der Lage, neue bisher unbekannte Angriffe (Zero-Day-Attacks) frühzeitig zu erkennen. Das Kernproblem von Intrusion Detection Systeme besteht jedoch in der optimalen Verarbeitung der gewaltigen Netzdaten und der Entwicklung eines in Echtzeit arbeitenden adaptiven Erkennungsmodells. Um diese Herausforderungen lösen zu können, stellt diese Dissertation ein Framework bereit, das aus zwei Hauptteilen besteht. Der erste Teil, OptiFilter genannt, verwendet ein dynamisches "Queuing Concept", um die zahlreich anfallenden Netzdaten weiter zu verarbeiten, baut fortlaufend Netzverbindungen auf, und exportiert strukturierte Input-Daten für das IDS. Den zweiten Teil stellt ein adaptiver Klassifikator dar, der ein Klassifikator-Modell basierend auf "Enhanced Growing Hierarchical Self Organizing Map" (EGHSOM), ein Modell für Netzwerk Normalzustand (NNB) und ein "Update Model" umfasst. In dem OptiFilter werden Tcpdump und SNMP traps benutzt, um die Netzwerkpakete und Hostereignisse fortlaufend zu aggregieren. Diese aggregierten Netzwerkpackete und Hostereignisse werden weiter analysiert und in Verbindungsvektoren umgewandelt. Zur Verbesserung der Erkennungsrate des adaptiven Klassifikators wird das künstliche neuronale Netz GHSOM intensiv untersucht und wesentlich weiterentwickelt. In dieser Dissertation werden unterschiedliche Ansätze vorgeschlagen und diskutiert. So wird eine classification-confidence margin threshold definiert, um die unbekannten bösartigen Verbindungen aufzudecken, die Stabilität der Wachstumstopologie durch neuartige Ansätze für die Initialisierung der Gewichtvektoren und durch die Stärkung der Winner Neuronen erhöht, und ein selbst-adaptives Verfahren eingeführt, um das Modell ständig aktualisieren zu können. Darüber hinaus besteht die Hauptaufgabe des NNB-Modells in der weiteren Untersuchung der erkannten unbekannten Verbindungen von der EGHSOM und der Überprüfung, ob sie normal sind. Jedoch, ändern sich die Netzverkehrsdaten wegen des Concept drif Phänomens ständig, was in Echtzeit zur Erzeugung nicht stationärer Netzdaten führt. Dieses Phänomen wird von dem Update-Modell besser kontrolliert. Das EGHSOM-Modell kann die neuen Anomalien effektiv erkennen und das NNB-Model passt die Änderungen in Netzdaten optimal an. Bei den experimentellen Untersuchungen hat das Framework erfolgversprechende Ergebnisse gezeigt. Im ersten Experiment wurde das Framework in Offline-Betriebsmodus evaluiert. Der OptiFilter wurde mit offline-, synthetischen- und realistischen Daten ausgewertet. Der adaptive Klassifikator wurde mit dem 10-Fold Cross Validation Verfahren evaluiert, um dessen Genauigkeit abzuschätzen. Im zweiten Experiment wurde das Framework auf einer 1 bis 10 GB Netzwerkstrecke installiert und im Online-Betriebsmodus in Echtzeit ausgewertet. Der OptiFilter hat erfolgreich die gewaltige Menge von Netzdaten in die strukturierten Verbindungsvektoren umgewandelt und der adaptive Klassifikator hat sie präzise klassifiziert. Die Vergleichsstudie zwischen dem entwickelten Framework und anderen bekannten IDS-Ansätzen zeigt, dass der vorgeschlagene IDSFramework alle anderen Ansätze übertrifft. Dies lässt sich auf folgende Kernpunkte zurückführen: Bearbeitung der gesammelten Netzdaten, Erreichung der besten Performanz (wie die Gesamtgenauigkeit), Detektieren unbekannter Verbindungen und Entwicklung des in Echtzeit arbeitenden Erkennungsmodells von Eindringversuchen.
Resumo:
As exploration of our solar system and outerspace move into the future, spacecraft are being developed to venture on increasingly challenging missions with bold objectives. The spacecraft tasked with completing these missions are becoming progressively more complex. This increases the potential for mission failure due to hardware malfunctions and unexpected spacecraft behavior. A solution to this problem lies in the development of an advanced fault management system. Fault management enables spacecraft to respond to failures and take repair actions so that it may continue its mission. The two main approaches developed for spacecraft fault management have been rule-based and model-based systems. Rules map sensor information to system behaviors, thus achieving fast response times, and making the actions of the fault management system explicit. These rules are developed by having a human reason through the interactions between spacecraft components. This process is limited by the number of interactions a human can reason about correctly. In the model-based approach, the human provides component models, and the fault management system reasons automatically about system wide interactions and complex fault combinations. This approach improves correctness, and makes explicit the underlying system models, whereas these are implicit in the rule-based approach. We propose a fault detection engine, Compiled Mode Estimation (CME) that unifies the strengths of the rule-based and model-based approaches. CME uses a compiled model to determine spacecraft behavior more accurately. Reasoning related to fault detection is compiled in an off-line process into a set of concurrent, localized diagnostic rules. These are then combined on-line along with sensor information to reconstruct the diagnosis of the system. These rules enable a human to inspect the diagnostic consequences of CME. Additionally, CME is capable of reasoning through component interactions automatically and still provide fast and correct responses. The implementation of this engine has been tested against the NEAR spacecraft advanced rule-based system, resulting in detection of failures beyond that of the rules. This evolution in fault detection will enable future missions to explore the furthest reaches of the solar system without the burden of human intervention to repair failed components.
Resumo:
In this report, a face recognition system that is capable of detecting and recognizing frontal and rotated faces was developed. Two face recognition methods focusing on the aspect of pose invariance are presented and evaluated - the whole face approach and the component-based approach. The main challenge of this project is to develop a system that is able to identify faces under different viewing angles in realtime. The development of such a system will enhance the capability and robustness of current face recognition technology. The whole-face approach recognizes faces by classifying a single feature vector consisting of the gray values of the whole face image. The component-based approach first locates the facial components and extracts them. These components are normalized and combined into a single feature vector for classification. The Support Vector Machine (SVM) is used as the classifier for both approaches. Extensive tests with respect to the robustness against pose changes are performed on a database that includes faces rotated up to about 40 degrees in depth. The component-based approach clearly outperforms the whole-face approach on all tests. Although this approach isproven to be more reliable, it is still too slow for real-time applications. That is the reason why a real-time face recognition system using the whole-face approach is implemented to recognize people in color video sequences.
Resumo:
This paper describes a trainable system capable of tracking faces and facialsfeatures like eyes and nostrils and estimating basic mouth features such as sdegrees of openness and smile in real time. In developing this system, we have addressed the twin issues of image representation and algorithms for learning. We have used the invariance properties of image representations based on Haar wavelets to robustly capture various facial features. Similarly, unlike previous approaches this system is entirely trained using examples and does not rely on a priori (hand-crafted) models of facial features based on optical flow or facial musculature. The system works in several stages that begin with face detection, followed by localization of facial features and estimation of mouth parameters. Each of these stages is formulated as a problem in supervised learning from examples. We apply the new and robust technique of support vector machines (SVM) for classification in the stage of skin segmentation, face detection and eye detection. Estimation of mouth parameters is modeled as a regression from a sparse subset of coefficients (basis functions) of an overcomplete dictionary of Haar wavelets.
Resumo:
Testing constraints for real-time systems are usually verified through the satisfiability of propositional formulae. In this paper, we propose an alternative where the verification of timing constraints can be done by counting the number of truth assignments instead of boolean satisfiability. This number can also tell us how “far away” is a given specification from satisfying its safety assertion. Furthermore, specifications and safety assertions are often modified in an incremental fashion, where problematic bugs are fixed one at a time. To support this development, we propose an incremental algorithm for counting satisfiability. Our proposed incremental algorithm is optimal as no unnecessary nodes are created during each counting. This works for the class of path RTL. To illustrate this application, we show how incremental satisfiability counting can be applied to a well-known rail-road crossing example, particularly when its specification is still being refined.