845 resultados para combinatorial pattern matching
Resumo:
Document representations can rapidly become unwieldy if they try to encapsulate all possible document properties, ranging from abstract structure to detailed rendering and layout. We present a composite document approach wherein an XMLbased document representation is linked via a shadow tree of bi-directional pointers to a PDF representation of the same document. Using a two-window viewer any material selected in the PDF can be related back to the corresponding material in the XML, and vice versa. In this way the treatment of specialist material such as mathematics, music or chemistry (e.g. via read aloud or play aloud ) can be activated via standard tools working within the XML representation, rather than requiring that application-specific structures be embedded in the PDF itself. The problems of textual recognition and tree pattern matching between the two representations are discussed in detail. Comparisons are drawn between our use of a shadow tree of pointers to map between document representations and the use of a code-replacement shadow tree in technologies such as XBL.
Resumo:
SQL Injection Attack (SQLIA) remains a technique used by a computer network intruder to pilfer an organisation’s confidential data. This is done by an intruder re-crafting web form’s input and query strings used in web requests with malicious intent to compromise the security of an organisation’s confidential data stored at the back-end database. The database is the most valuable data source, and thus, intruders are unrelenting in constantly evolving new techniques to bypass the signature’s solutions currently provided in Web Application Firewalls (WAF) to mitigate SQLIA. There is therefore a need for an automated scalable methodology in the pre-processing of SQLIA features fit for a supervised learning model. However, obtaining a ready-made scalable dataset that is feature engineered with numerical attributes dataset items to train Artificial Neural Network (ANN) and Machine Leaning (ML) models is a known issue in applying artificial intelligence to effectively address ever evolving novel SQLIA signatures. This proposed approach applies numerical attributes encoding ontology to encode features (both legitimate web requests and SQLIA) to numerical data items as to extract scalable dataset for input to a supervised learning model in moving towards a ML SQLIA detection and prevention model. In numerical attributes encoding of features, the proposed model explores a hybrid of static and dynamic pattern matching by implementing a Non-Deterministic Finite Automaton (NFA). This combined with proxy and SQL parser Application Programming Interface (API) to intercept and parse web requests in transition to the back-end database. In developing a solution to address SQLIA, this model allows processed web requests at the proxy deemed to contain injected query string to be excluded from reaching the target back-end database. This paper is intended for evaluating the performance metrics of a dataset obtained by numerical encoding of features ontology in Microsoft Azure Machine Learning (MAML) studio using Two-Class Support Vector Machines (TCSVM) binary classifier. This methodology then forms the subject of the empirical evaluation.
Resumo:
Recent years have seen an astronomical rise in SQL Injection Attacks (SQLIAs) used to compromise the confidentiality, authentication and integrity of organisations’ databases. Intruders becoming smarter in obfuscating web requests to evade detection combined with increasing volumes of web traffic from the Internet of Things (IoT), cloud-hosted and on-premise business applications have made it evident that the existing approaches of mostly static signature lack the ability to cope with novel signatures. A SQLIA detection and prevention solution can be achieved through exploring an alternative bio-inspired supervised learning approach that uses input of labelled dataset of numerical attributes in classifying true positives and negatives. We present in this paper a Numerical Encoding to Tame SQLIA (NETSQLIA) that implements a proof of concept for scalable numerical encoding of features to a dataset attributes with labelled class obtained from deep web traffic analysis. In the numerical attributes encoding: the model leverages proxy in the interception and decryption of web traffic. The intercepted web requests are then assembled for front-end SQL parsing and pattern matching by applying traditional Non-Deterministic Finite Automaton (NFA). This paper is intended for a technique of numerical attributes extraction of any size primed as an input dataset to an Artificial Neural Network (ANN) and statistical Machine Learning (ML) algorithms implemented using Two-Class Averaged Perceptron (TCAP) and Two-Class Logistic Regression (TCLR) respectively. This methodology then forms the subject of the empirical evaluation of the suitability of this model in the accurate classification of both legitimate web requests and SQLIA payloads.
Resumo:
A visibility/invisibility paradox of trust operates in the development of distributed educational leadership for online communities. If trust is to be established, the team-based informal ethos of online collaborative networked communities requires a different kind of leadership from that observed in more formal face-to-face positional hierarchies. Such leadership is more flexible and sophisticated, being capable of encompassing both ambiguity and agile response to change. Online educational leaders need to be partially invisible, delegating discretionary powers, to facilitate the effective distribution of leadership tasks in a highly trusting team-based culture. Yet, simultaneously, online communities are facilitated by the visibility and subtle control effected by expert leaders. This paradox: that leaders need to be both highly visible and invisible when appropriate, was derived during research on 'Trust and Leadership' and tested in the analysis of online community case study discussions using a pattern-matching process to measure conversational interactions. This paper argues that both leader visibility and invisibility are important for effective trusting collaboration in online distributed leadership. Advanced leadership responses to complex situations in online communities foster positive group interaction, mutual trust and effective decision-making, facilitated through the active distribution of tasks.
Resumo:
We consider the statistical problem of catalogue matching from a machine learning perspective with the goal of producing probabilistic outputs, and using all available information. A framework is provided that unifies two existing approaches to producing probabilistic outputs in the literature, one based on combining distribution estimates and the other based on combining probabilistic classifiers. We apply both of these to the problem of matching the HI Parkes All Sky Survey radio catalogue with large positional uncertainties to the much denser SuperCOSMOS catalogue with much smaller positional uncertainties. We demonstrate the utility of probabilistic outputs by a controllable completeness and efficiency trade-off and by identifying objects that have high probability of being rare. Finally, possible biasing effects in the output of these classifiers are also highlighted and discussed.
Resumo:
We consider brightness/contrast-invariant and rotation-discriminating template matching that searches an image to analyze A for a query image Q We propose to use the complex coefficients of the discrete Fourier transform of the radial projections to compute new rotation-invariant local features. These coefficients can be efficiently obtained via FFT. We classify templates in ""stable"" and ""unstable"" ones and argue that any local feature-based template matching may fail to find unstable templates. We extract several stable sub-templates of Q and find them in A by comparing the features. The matchings of the sub-templates are combined using the Hough transform. As the features of A are computed only once, the algorithm can find quickly many different sub-templates in A, and it is Suitable for finding many query images in A, multi-scale searching and partial occlusion-robust template matching. (C) 2009 Elsevier Ltd. All rights reserved.
Resumo:
We develop a neoclassical trade model with heterogeneous factors of production. We consider a world with two factors, labor and .managers., each with a distribution of ability levels. Production combines a manager of some type with a group of workers. The output of a unit depends on the types of the two factors, with complementarity between them, while exhibiting diminishing returns to the number of workers. We examine the sorting of factors to sectors and the matching of factors within sectors, and we use the model to study the determinants of the trade pattern and the effects of trade on the wage and salary distributions. Finally, we extend the model to include search frictions and consider the distribution of employment rates.
Resumo:
INTRODUCTION. Patient-ventilator asynchrony is a frequent issue in non invasivemechanical ventilation (NIV) and leaks at the patient-mask interface play a major role in itspathogenesis. NIV algorithms alleviate the deleterious impact of leaks and improve patient-ventilator interaction. Neurally adusted ventilatory assist (NAVA), a neurally triggered modethat avoids interferences between leaks and the usual pneumatic trigger, could further improvepatient-ventilator interaction in NIV patients.OBJECTIVES. To evaluate the feasibility ofNAVAin patients receiving a prophylactic postextubationNIV and to compare the respective impact ofPSVandNAVAwith and withoutNIValgorithm on patient-ventilator interaction.METHODS. Prospective study conducted in 16 beds adult critical care unit (ICU) in a tertiaryuniversity hospital. Over a 2 months period, were included 17 adult medical ICU patientsextubated for less than 2 h and in whom a prophylactic post-extubation NIV was indicated.Patients were randomly mechanically ventilated for 10 min with: PSV without NIV algorithm(PSV-NIV-), PSV with NIV algorithm (PSV-NIV+),NAVAwithout NIV algorithm (NAVANIV-)and NAVA with NIV algorithm (NAVA-NIV+). Breathing pattern descriptors, diaphragmelectrical activity, leaks volume, inspiratory trigger delay (Tdinsp), inspiratory time inexcess (Tiexcess) and the five main asynchronies were quantified. Asynchrony index (AI) andasynchrony index influenced by leaks (AIleaks) were computed.RESULTS. Peak inspiratory pressure and diaphragm electrical activity were similar in thefour conditions. With both PSV and NAVA, NIV algorithm significantly reduced the level ofleak (p\0.01). Tdinsp was not affected by NIV algorithm but was shorter in NAVA than inPSV (p\0.01). Tiexcess was shorter in NAVA and PSV-NIV+ than in PSV-NIV- (p\0.05).The prevalence of double triggering was significantly lower in PSV-NIV+ than in NAVANIV+.As compared to PSV,NAVAsignificantly reduced the prevalence of premature cyclingand late cycling while NIV algorithm did not influenced premature cycling. AI was not affectedby NIV algorithm but was significantly lower in NAVA than in PSV (p\0.05). AIleaks wasquasi null with NAVA and significantly lower than in PSV (p\0.05).CONCLUSIONS. NAVA is feasible in patients receiving a post-extubation prophylacticNIV. NAVA and NIV improve patient-ventilator synchrony in different manners. NAVANIV+offers the best patient-ventilator interaction. Clinical studies are required to assess thepotential clinical benefit of NAVA in patients receiving NIV.
Resumo:
A multicomponent indicator displacement assay ( MIDA) based on an organometallic receptor and three dyes can be used for the identification and quantification of nucleotides in aqueous solution at neutral pH.
Resumo:
We study the preservation of the periodic orbits of an A-monotone tree map f:T→T in the class of all tree maps g:S→S having a cycle with the same pattern as A. We prove that there is a period-preserving injective map from the set of (almost all) periodic orbits of ƒ into the set of periodic orbits of each map in the class. Moreover, the relative positions of the corresponding orbits in the trees T and S (which need not be homeomorphic) are essentially preserved
Resumo:
The usage of digital content, such as video clips and images, has increased dramatically during the last decade. Local image features have been applied increasingly in various image and video retrieval applications. This thesis evaluates local features and applies them to image and video processing tasks. The results of the study show that 1) the performance of different local feature detector and descriptor methods vary significantly in object class matching, 2) local features can be applied in image alignment with superior results against the state-of-the-art, 3) the local feature based shot boundary detection method produces promising results, and 4) the local feature based hierarchical video summarization method shows promising new new research direction. In conclusion, this thesis presents the local features as a powerful tool in many applications and the imminent future work should concentrate on improving the quality of the local features.
Resumo:
Formalizing algorithm derivations is a necessary prerequisite for developing automated algorithm design systems. This report describes a derivation of an algorithm for incrementally matching conjunctive patterns against a growing database. This algorithm, which is modeled on the Rete matcher used in the OPS5 production system, forms a basis for efficiently implementing a rule system. The highlights of this derivation are: (1) a formal specification for the rule system matching problem, (2) derivation of an algorithm for this task using a lattice-theoretic model of conjunctive and disjunctive variable substitutions, and (3) optimization of this algorithm, using finite differencing, for incrementally processing new data.
Resumo:
2D electrophoresis is a well-known method for protein separation which is extremely useful in the field of proteomics. Each spot in the image represents a protein accumulation and the goal is to perform a differential analysis between pairs of images to study changes in protein content. It is thus necessary to register two images by finding spot correspondences. Although it may seem a simple task, generally, the manual processing of this kind of images is very cumbersome, especially when strong variations between corresponding sets of spots are expected (e.g. strong non-linear deformations and outliers). In order to solve this problem, this paper proposes a new quadratic assignment formulation together with a correspondence estimation algorithm based on graph matching which takes into account the structural information between the detected spots. Each image is represented by a graph and the task is to find a maximum common subgraph. Successful experimental results using real data are presented, including an extensive comparative performance evaluation with ground-truth data. (C) 2010 Elsevier B.V. All rights reserved.