935 resultados para face recognition,face detection,face verification,web application


Relevância:

100.00% 100.00%

Publicador:

Resumo:

The present study examines the effect of the goodness of view on the minimal exposure time required to recognize depth-rotated objects. In a previous study, Verfaillie and Boutsen (1995) derived scales of goodness of view, using a new corpus of images of depth-rotated objects. In the present experiment, a subset of this corpus (five views of 56 objects) is used to determine the recognition exposure time for each view, by increasing exposure time across successive presentations until the object is recognized. The results indicate that, for two thirds of the objects, good views are recognized more frequently and have lower recognition exposure times than bad views.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Models are central tools for modern scientists and decision makers, and there are many existing frameworks to support their creation, execution and composition. Many frameworks are based on proprietary interfaces, and do not lend themselves to the integration of models from diverse disciplines. Web based systems, or systems based on web services, such as Taverna and Kepler, allow composition of models based on standard web service technologies. At the same time the Open Geospatial Consortium has been developing their own service stack, which includes the Web Processing Service, designed to facilitate the executing of geospatial processing - including complex environmental models. The current Open Geospatial Consortium service stack employs Extensible Markup Language as a default data exchange standard, and widely-used encodings such as JavaScript Object Notation can often only be used when incorporated with Extensible Markup Language. Similarly, no successful engagement of the Web Processing Service standard with the well-supported technologies of Simple Object Access Protocol and Web Services Description Language has been seen. In this paper we propose a pure Simple Object Access Protocol/Web Services Description Language processing service which addresses some of the issues with the Web Processing Service specication and brings us closer to achieving a degree of interoperability between geospatial models, and thus realising the vision of a useful 'model web'.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The Semantic Web relies on carefully structured, well defined, data to allow machines to communicate and understand one another. In many domains (e.g. geospatial) the data being described contains some uncertainty, often due to incomplete knowledge; meaningful processing of this data requires these uncertainties to be carefully analysed and integrated into the process chain. Currently, within the SemanticWeb there is no standard mechanism for interoperable description and exchange of uncertain information, which renders the automated processing of such information implausible, particularly where error must be considered and captured as it propagates through a processing sequence. In particular we adopt a Bayesian perspective and focus on the case where the inputs / outputs are naturally treated as random variables. This paper discusses a solution to the problem in the form of the Uncertainty Markup Language (UncertML). UncertML is a conceptual model, realised as an XML schema, that allows uncertainty to be quantified in a variety of ways i.e. realisations, statistics and probability distributions. UncertML is based upon a soft-typed XML schema design that provides a generic framework from which any statistic or distribution may be created. Making extensive use of Geography Markup Language (GML) dictionaries, UncertML provides a collection of definitions for common uncertainty types. Containing both written descriptions and mathematical functions, encoded as MathML, the definitions within these dictionaries provide a robust mechanism for defining any statistic or distribution and can be easily extended. Universal Resource Identifiers (URIs) are used to introduce semantics to the soft-typed elements by linking to these dictionary definitions. The INTAMAP (INTeroperability and Automated MAPping) project provides a use case for UncertML. This paper demonstrates how observation errors can be quantified using UncertML and wrapped within an Observations & Measurements (O&M) Observation. The interpolation service uses the information within these observations to influence the prediction outcome. The output uncertainties may be encoded in a variety of UncertML types, e.g. a series of marginal Gaussian distributions, a set of statistics, such as the first three marginal moments, or a set of realisations from a Monte Carlo treatment. Quantifying and propagating uncertainty in this way allows such interpolation results to be consumed by other services. This could form part of a risk management chain or a decision support system, and ultimately paves the way for complex data processing chains in the Semantic Web.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Our research explores the possibility of categorizing webpages and webpage genre by structure or layout. Based on our results, we believe that webpage structure could play an important role, along with textual and visual keywords, in webpage categorization and searching.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

In this paper the key features of a two-layered model for describing the semantic of dynamical web resources are introduced. In the current Semantic Web proposal [Berners-Lee et al., 2001] web resources are classified into static ontologies which describes the semantic network of their inter-relationships [Kalianpur, 2001][Handschuh & Staab, 2002] and complex constraints described by logical quantified formula [Boley et al., 2001][McGuinnes & van Harmelen, 2004][McGuinnes et al., 2004], the basic idea is that software agents can use techniques of automatic reasoning in order to relate resources and to support sophisticated web application. On the other hand, web resources are also characterized by their dynamical aspects, which are not adequately addressed by current web models. Resources on the web are dynamical since, in the minimal case, they can appear or disappear from the web and their content is upgraded. In addition, resources can traverse different states, which characterized the resource life-cycle, each resource state corresponding to different possible uses of the resource. Finally most resources are timed, i.e. they information they provide make sense only if contextualised with respect to time, and their validity and accuracy is greatly bounded by time. Temporal projection and deduction based on dynamical and time constraints of the resources can be made and exploited by software agents [Hendler, 2001] in order to make previsions about the availability and the state of a resource, for deciding when consulting the resource itself or in order to deliberately induce a resource state change for reaching some agent goal, such as in the automated planning framework [Fikes & Nilsson, 1971][Bacchus & Kabanza,1998].

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Most research in the area of emotion detection in written text focused on detecting explicit expressions of emotions in text. In this paper, we present a rule-based pipeline approach for detecting implicit emotions in written text without emotion-bearing words based on the OCC Model. We have evaluated our approach on three different datasets with five emotion categories. Our results show that the proposed approach outperforms the lexicon matching method consistently across all the three datasets by a large margin of 17–30% in F-measure and gives competitive performance compared to a supervised classifier. In particular, when dealing with formal text which follows grammatical rules strictly, our approach gives an average F-measure of 82.7% on “Happy”, “Angry-Disgust” and “Sad”, even outperforming the supervised baseline by nearly 17% in F-measure. Our preliminary results show the feasibility of the approach for the task of implicit emotion detection in written text.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Negli ultimi vent'anni con lo sviluppo di Internet, il modo di comunicare tra le persone �è totalmente cambiato. Grazie a Internet si sono ridotte le distanze e soprattutto tramite i siti web le aziende hanno una propria vetrina sul mondo sempre accessibile. Tutto ci�ò ha portato a nuovi comportamenti da parte dei consumatori che divengono sempre pi�u esigenti nella vastità di informazioni presenti sul Web. Perciò è necessario che le web companies riescano a produrre website efficienti e usabili per favorire l'interazione con l'utente. Inoltre il web ha avuto una rapida espansione per quanto concerne le metodologie di sviluppo e analisi del comportamento del consumatore. Si cercano sempre nuovi spunti per poter acquisire quello che �è il percorso di un utente affinché porti a termine una determinata azione nel proprio dominio. Per questo motivo, oltre agli strumenti gi�à consolidati come il riempimento di questionari o il tracking per mezzo di piattaforme come Google Analytics, si �è pensato di andare oltre e cercare di analizzare ancora pi�u a fondo il "consumAttore". Grazie ad un eye-tracker �è possibile riconoscere quelli che sono i modelli cognitivi che riguardano il percorso di ricerca, valutazione e acquisto di un prodotto o una call to action, e come i contenuti di una web application influenzano l'attenzione e la user experience. Pertanto l'obiettivo che si pone questo studio �è quello di poter misurare l'engagement della navigazione utente di una web application e, nel caso fosse necessario, ottimizzare i contenuti al suo interno. Per il rilevamento delle informazioni necessarie durante l'esperimento, mi sono servito di uno strumento a supporto delle decisioni, ovvero un eye-tracker e della successiva somministrazione di questionari.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Nell'elaborato viene introdotto l'ambito della Computer Vision e come l'algoritmo SIFT si inserisce nel suo panorama. Viene inoltre descritto SIFT stesso, le varie fasi di cui si compone e un'applicazione al problema dell'object recognition. Infine viene presentata un'implementazione di SIFT in linguaggio Python creata per ottenere un'applicazione didattica interattiva e vengono mostrati esempi di questa applicazione.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Dimensional and form inspections are key to the manufacturing and assembly of products. Product verification can involve a number of different measuring instruments operated using their dedicated software. Typically, each of these instruments with their associated software is more suitable for the verification of a pre-specified quality characteristic of the product than others. The number of different systems and software applications to perform a complete measurement of products and assemblies within a manufacturing organisation is therefore expected to be large. This number becomes even larger as advances in measurement technologies are made. The idea of a universal software application for any instrument still appears to be only a theoretical possibility. A need for information integration is apparent. In this paper, a design of an information system to consistently manage (store, search, retrieve, search, secure) measurement results from various instruments and software applications is introduced. Two of the main ideas underlying the proposed system include abstracting structures and formats of measurement files from the data so that complexity and compatibility between different approaches to measurement data modelling is avoided. Secondly, the information within a file is enriched with meta-information to facilitate its consistent storage and retrieval. To demonstrate the designed information system, a web application is implemented. © Springer-Verlag Berlin Heidelberg 2010.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Background Delirium is highly prevalent, especially in older patients. It independently leads to adverse outcomes, but remains under-detected, particularly hypoactive forms. Although early identification and intervention is important, delirium prevention is key to improving outcomes. The delirium prodrome concept has been mooted for decades, but remains poorly characterised. Greater understanding of this prodrome would promote prompt identification of delirium-prone patients, and facilitate improved strategies for delirium prevention and management. Methods Medical inpatients of ≥70 years were screened for prevalent delirium using the Revised Delirium Rating Scale (DRS--‐R98). Those without prevalent delirium were assessed daily for delirium development, prodromal features and motor subtype. Survival analysis models identified which prodromal features predicted the emergence of incident delirium in the cohort in the first week of admission. The Delirium Motor Subtype Scale-4 was used to ascertain motor subtype. Results Of 555 patients approached, 191 patients were included in the prospective study. The median age was 80 (IQR 10) and 101 (52.9%) were male. Sixty-one patients developed incident delirium within a week of admission. Several prodromal features predicted delirium emergence in the cohort. Firstly, using a novel Prodromal Checklist based on the existing literature, and controlling for confounders, seven predictive behavioural features were identified in the prodromal period (for example, increasing confusion; and being easily distractible). Additionally, using serial cognitive tests and the DRS-R98 daily, multiple cognitive and other core delirium features were detected in the prodrome (for example inattention; and sleep-wake cycle disturbance). Examining longitudinal motor subtypes in delirium cases, subtypes were found to be predominantly stable over time, the most prevalent being hypoactive subtype (62.3%). Discussion This thesis explored multiple aspects of delirium in older medical inpatients, with particular focus on the characterisation of the delirium prodrome. These findings should help to inform future delirium educational programmes, and detection and prevention strategies.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Intrusion Detection Systems (IDSs) provide an important layer of security for computer systems and networks, and are becoming more and more necessary as reliance on Internet services increases and systems with sensitive data are more commonly open to Internet access. An IDS’s responsibility is to detect suspicious or unacceptable system and network activity and to alert a systems administrator to this activity. The majority of IDSs use a set of signatures that define what suspicious traffic is, and Snort is one popular and actively developing open-source IDS that uses such a set of signatures known as Snort rules. Our aim is to identify a way in which Snort could be developed further by generalising rules to identify novel attacks. In particular, we attempted to relax and vary the conditions and parameters of current Snort rules, using a similar approach to classic rule learning operators such as generalisation and specialisation. We demonstrate the effectiveness of our approach through experiments with standard datasets and show that we are able to detect previously undetected variants of various attacks. We conclude by discussing the general effectiveness and appropriateness of generalisation in Snort based IDS rule processing. Keywords: anomaly detection, intrusion detection, Snort, Snort rules

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Intrusion Detection Systems (IDSs) provide an important layer of security for computer systems and networks, and are becoming more and more necessary as reliance on Internet services increases and systems with sensitive data are more commonly open to Internet access. An IDS’s responsibility is to detect suspicious or unacceptable system and network activity and to alert a systems administrator to this activity. The majority of IDSs use a set of signatures that define what suspicious traffic is, and Snort is one popular and actively developing open-source IDS that uses such a set of signatures known as Snort rules. Our aim is to identify a way in which Snort could be developed further by generalising rules to identify novel attacks. In particular, we attempted to relax and vary the conditions and parameters of current Snort rules, using a similar approach to classic rule learning operators such as generalisation and specialisation. We demonstrate the effectiveness of our approach through experiments with standard datasets and show that we are able to detect previously undetected variants of various attacks. We conclude by discussing the general effectiveness and appropriateness of generalisation in Snort based IDS rule processing. Keywords: anomaly detection, intrusion detection, Snort, Snort rules

Relevância:

100.00% 100.00%

Publicador:

Resumo:

El trabajo de graduación describe el desarrollo de una Aplicación web desarrollada para la Asociación cooperativa de producción pesquera y servicios múltiples “Bocana Lempa - Jaltepec de R.L”, en el Cantón las Mesas del municipio de Tecoluca del departamento de San Vicente, con dicha aplicación se automatizaron los procesos de las áreas de comercialización y administración de la cooperativa, los cuales son, Activo fijo, Recursos Humanos, Facturación, Inventario y el desarrollo de una página web a la que accesan los clientes para la realización de pedidos. Con el desarrollo de dicha aplicación se ha obtenido una mayor agilización en los procesos, los cuales conllevan a adquirir mayores logros entre los cuales podemos mencionar: información detallada de los movimientos de inventario, presentación de informes, mejoras en el control de activo fijo, salarios, vacaciones, permisos de los empleados, registro de pedidos realizados desde la página web, impresión de facturas y la reducción del tiempo en el desarrollo de los procesos. SUMMARY The graduation work describes the development of a web application developed for the cooperative association of fish production and multiple services "Bocana Lempa - Jaltepec RL" in the Canton Tables Township Tecoluca the department of San Vicente, with such application they automated processes as the marketing and administration of the cooperative, which are, fixed Assets, Human Resources, Billing, Inventory and development of a website that access to customers for ordering. With the development of this application has gained greater streamlining processes, which lead to acquire greater achievements among which include: detailed information on inventory movements, reporting, improved control of fixed assets, wages holidays, permits employees record orders made from the website, print invoices and time reduction in development processes. Keywords : Online ordering , fixed assets, payroll employees , billing, control fishing. The graduation work describes the development of a web application developed for the cooperative association of fish production and multiple services "Bocana Lempa - Jaltepec RL" in the Canton Tables Township Tecoluca the department of San Vicente, with such application they automated processes as the marketing and administration of the cooperative, which are, fixed Assets, Human Resources, Billing, Inventory and development of a website that access to customers for ordering. With the development of this application has gained greater streamlining processes, which lead to acquire greater achievements among which include: detailed information on inventory movements, reporting, improved control of fixed assets, wages holidays, permits employees record orders made from the website, print invoices and time reduction in development processes.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

El principal objetivo de este trabajo fin de grado es la implementación de una aplicación web que permita realizar actividades de enseñanza/aprendizaje en las ramas de cálculo y álgebra de las matemáticas, enfocada principalmente en los contenidos impartidos en las asignaturas de matemáticas en Bachillerato, concretamente de la unidad de realización de derivadas. Se han desarrollado dos modelos: Uno para el profesor, que permite la generación de los ejercicios, así como la consulta de las posibles soluciones. Esta aplicación ofrece al docente la posibilidad de mediante una serie de ajustes, generar de forma dinámica las actividades deseadas. Otro para el alumno, que permite la realización de las actividades y la inclusión de respuestas paso a paso, además de la visualización de las correcciones a sus procedimientos. Para el desarrollo de estas aplicaciones se ha usado el lenguaje Java con ayuda de JSON para el intercambio de datos. También se ha utilizado el motor de respuestas Wolfram Alpha para realizar las correcciones paso a paso de las respuestas de los alumnos.