97 resultados para Language-based Editor
em Universidad Politécnica de Madrid
Resumo:
El objetivo de este proyecto ha sido el de realizar un análisis del importante desarrollo que han sufrido las telecomunicaciones, haciendo un especial hincapié en la telefonía móvil y el impacto y repercusión que ha causado actualmente en nuestra sociedad. Para ello se hará un repaso evolutivo de las tecnologías de la información y las telecomunicaciones, y se establecerá una relación entre la gran difusión de éstas y su efecto sobre los usos, y cambios percibidos por los consumidores del nuevo siglo. Ciertamente la historia de la tecnología, nos enseña que la gente y las organizaciones acaban utilizándola para unos propósitos muy diferentes de aquellos que inicialmente fueron concebidas. Además cuanto más interactiva sea una tecnología, tanto más probable será que los usuarios se conviertan en productores o modificadores de la misma. Por tanto, la sociedad necesita resolver las incógnitas que pueda suscitar el rápido y continúo cambio de las comunicaciones. Este proyecto trata de ayudar a responder alguna de las cuestiones que actualmente se están planteando. ¿Son los teléfonos móviles una expresión de identidad, artilugios de moda, herramientas de la vida cotidiana, o todo lo anterior? ¿Existen nuevos modelos de comportamiento y conducta social? ¿La comunicación móvil está favoreciendo la aparición de una nueva cultura joven con un lenguaje propio basado en la comunicación textual y multimodal? ¿Tienen los teléfonos móviles efectos nocivos en la salud? La respuesta a estas preguntas afecta a nuestras vidas y también condiciona las políticas públicas y las estrategias de negocio, por eso requiere adquirir un conocimiento cimentado en la información, y la recopilación de datos de diversas fuentes, tanto de estadísticas provenientes de diferentes estudios e investigaciones, como de empresas consultoras, siempre basada en una perspectiva global. En conjunto, se espera dentro de los límites del conocimiento actual, contribuir a establecer las bases para el análisis y valoración de la relación existente entre comunicación, tecnología y sociedad en todo el mundo. Abstract The purpose of this project has been to analyse the significant development undergone by telecommunications, putting a special emphasis on mobile phones and the impact it has caused in society. We will go over the evolution of IT technologies and telecommunications as well as establish a relationship between its spread and effect of its uses and changes understood by the new century consumers. Technology history shows us that people and organizations use it for very different purposes from those originally thought. Furthermore, the more interactive technologies are, the more users will modify or produce it. Therefore, society needs to solve the mysteries of the quick and continuous change of communications. This project tries to help and answer some of the questions considered these days. Are mobile phones an expression of identity, fashionable devices, tools for everyday life or all at once? Are there any new models of performance and social behaviour? Is mobile communication favouring the existence of a new young culture with a typical language based on textual and multimodal communication? Are mobile phones bad for our health? The answer to these questions affects us all and conditions public politics and business strategies so it is required to get firm knowledge based on information. It is also important to compile data from various sources, from statistics of research and studies, based on a global perspective. As a whole, we hope to contribute to establish the bases for the future analysis and assessment of a fundamental trend that is redefining the relationship between communication, technology and society worldwide by transforming the wireless networks that make our lives.
Resumo:
Andorra-I is the first implementation of a language based on the Andorra Principie, which states that determinate goals can (and shonld) be run before other goals, and even in a parallel fashion. This principie has materialized in a framework called the Basic Andorra model, which allows or-parallelism as well as (dependent) and-parallelism for determinate goals. In this report we show that it is possible to further extend this model in order to allow general independent and-parallelism for nondeterminate goals, withont greatly modifying the underlying implementation machinery. A simple an easy way to realize such an extensión is to make each (nondeterminate) independent goal determinate, by using a special "bagof" constract. We also show that this can be achieved antomatically by compile-time translation from original Prolog programs. A transformation that fulfüls this objective and which can be easily antomated is presented in this report.
Resumo:
El presente proyecto final de carrera titulado “Modelado de alto nivel con SystemC” tiene como objetivo principal el modelado de algunos módulos de un codificador de vídeo MPEG-2 utilizando el lenguaje de descripción de sistemas igitales SystemC con un nivel de abstracción TLM o Transaction Level Modeling. SystemC es un lenguaje de descripción de sistemas digitales basado en C++. En él hay un conjunto de rutinas y librerías que implementan tipos de datos, estructuras y procesos especiales para el modelado de sistemas digitales. Su descripción se puede consultar en [GLMS02] El nivel de abstracción TLM se caracteriza por separar la comunicación entre los módulos de su funcionalidad. Este nivel de abstracción hace un mayor énfasis en la funcionalidad de la comunicación entre los módulos (de donde a donde van datos) que la implementación exacta de la misma. En los documentos [RSPF] y [HG] se describen el TLM y un ejemplo de implementación. La arquitectura del modelo se basa en el codificador MVIP-2 descrito en [Gar04], de dicho modelo, los módulos implementados son: · IVIDEOH: módulo que realiza un filtrado del vídeo de entrada en la dimensión horizontal y guarda en memoria el video filtrado. · IVIDEOV: módulo que lee de la memoria el vídeo filtrado por IVIDEOH, realiza el filtrado en la dimensión horizontal y escribe el video filtrado en memoria. · DCT: módulo que lee el video filtrado por IVIDEOV, hace la transformada discreta del coseno y guarda el vídeo transformado en la memoria. · QUANT: módulo que lee el video transformado por DCT, lo cuantifica y guarda el resultado en la memoria. · IQUANT: módulo que lee el video cuantificado por QUANT, realiza la cuantificación inversa y guarda el resultado en memoria. · IDCT: módulo que lee el video procesado por IQUANT, realiza la transformada inversa del coseno y guarda el resultado en memoria. · IMEM: módulo que hace de interfaz entre los módulos anteriores y la memoria. Gestiona las peticiones simultáneas de acceso a la memoria y asegura el acceso exclusivo a la memoria en cada instante de tiempo. Todos estos módulos aparecen en gris en la siguiente figura en la que se muestra la arquitectura del modelo: Figura 1. Arquitectura del modelo (VER PDF DEL PFC) En figura también aparecen unos módulos en blanco, dichos módulos son de pruebas y se han añadido para realizar simulaciones y probar los módulos del modelo: · CAMARA: módulo que simula una cámara en blanco y negro, lee la luminancia de un fichero de vídeo y lo envía al modelo a través de una FIFO. · FIFO: hace de interfaz entre la cámara y el modelo, guarda los datos que envía la cámara hasta que IVIDEOH los lee. · CONTROL: módulo que se encarga de controlar los módulos que procesan el vídeo, estos le indican cuando terminan de procesar un frame de vídeo y este módulo se encarga de iniciar los módulos que sean necesarios para seguir con la codificación. Este módulo se encarga del correcto secuenciamiento de los módulos procesadores de vídeo. · RAM: módulo que simula una memoria RAM, incluye un retardo programable en el acceso. Para las pruebas también se han generado ficheros de vídeo con el resultado de cada módulo procesador de vídeo, ficheros con mensajes y un fichero de trazas en el que se muestra el secuenciamiento de los procesadores. Como resultado del trabajo en el presente PFC se puede concluir que SystemC permite el modelado de sistemas digitales con bastante sencillez (hace falta conocimientos previos de C++ y programación orientada objetos) y permite la realización de modelos con un nivel de abstracción mayor a RTL, el habitual en Verilog y VHDL, en el caso del presente PFC, el TLM. ABSTRACT This final career project titled “High level modeling with SystemC” have as main objective the modeling of some of the modules of an MPEG-2 video coder using the SystemC digital systems description language at the TLM or Transaction Level Modeling abstraction level. SystemC is a digital systems description language based in C++. It contains routines and libraries that define special data types, structures and process to model digital systems. There is a complete description of the SystemC language in the document [GLMS02]. The main characteristic of TLM abstraction level is that it separates the communication among modules of their functionality. This abstraction level puts a higher emphasis in the functionality of the communication (from where to where the data go) than the exact implementation of it. The TLM and an example are described in the documents [RSPF] and [HG]. The architecture of the model is based in the MVIP-2 video coder (described in the document [Gar04]) The modeled modules are: · IVIDEOH: module that filter the video input in the horizontal dimension. It saves the filtered video in the memory. · IVIDEOV: module that read the IVIDEOH filtered video, filter it in the vertical dimension and save the filtered video in the memory. · DCT: module that read the IVIDEOV filtered video, do the discrete cosine transform and save the transformed video in the memory. · QUANT: module that read the DCT transformed video, quantify it and save the quantified video in the memory. · IQUANT: module that read the QUANT processed video, do the inverse quantification and save the result in the memory. · IDCT: module that read the IQUANT processed video, do the inverse cosine transform and save the result in the memory. · IMEM: this module is the interface between the modules described previously and the memory. It manage the simultaneous accesses to the memory and ensure an unique access at each instant of time All this modules are included in grey in the following figure (SEE PDF OF PFC). This figure shows the architecture of the model: Figure 1. Architecture of the model This figure also includes other modules in white, these modules have been added to the model in order to simulate and prove the modules of the model: · CAMARA: simulates a black and white video camera, it reads the luminance of a video file and sends it to the model through a FIFO. · FIFO: is the interface between the camera and the model, it saves the video data sent by the camera until the IVIDEOH module reads it. · CONTROL: controls the modules that process the video. These modules indicate the CONTROL module when they have finished the processing of a video frame. The CONTROL module, then, init the necessary modules to continue with the video coding. This module is responsible of the right sequence of the video processing modules. · RAM: it simulates a RAM memory; it also simulates a programmable delay in the access to the memory. It has been generated video files, text files and a trace file to check the correct function of the model. The trace file shows the sequence of the video processing modules. As a result of the present final career project, it can be deduced that it is quite easy to model digital systems with SystemC (it is only needed previous knowledge of C++ and object oriented programming) and it also allow the modeling with a level of abstraction higher than the RTL used in Verilog and VHDL, in the case of the present final career project, the TLM.
Resumo:
La Mezquita-Catedral de Córdoba es un edificio vivo. Un edificio que ha sido transformado sucesivamente por hombres de razas, culturas y religiones distintas durante sus más de 1.200 años de vida y que, a pesar de ello, no ha dejado de estar en uso ni uno solo de esos días de esa larga vida. De esta forma, el edificio se muestra ante el visitante como un complejo objeto arquitectónico, resultado de una continua transformación. La capacidad de la transformación de los edificios es algo inherente a su propia condición arquitectónica, no es un hecho exclusivo de la Mezquita-Catedral. Sin embargo, en este edificio esa transformación se produce con una gran intensidad y sin pérdida de su autenticidad. Tradicionalmente, los edificios se han adaptado a los nuevos requerimientos de cada época en un proceso que ha buscado en el propio edificio las leyes o principios que habían de regir la intervención. De esta forma, tanto las sucesivas ampliaciones de la Mezquita de Abd al-Rahman I como las siguientes intervenciones cristianas debieron asumir lo preexistente como material de trabajo. Así, los arquitectos del califa al-Hakam II dialogaron con sus antecesores complejizando el espacio que recibieron, así como los Hernán Ruiz consiguieron un nuevo organismo resultante de la introducción de su arquitectura luminosa en la trama hispanomusulmana. El siglo XIX confirmó el deseo por descubrir las huellas de un pasado esplendoroso que la intervención barroca había silenciado bajo un tratamiento homogéneo del espacio. La recuperación de esas huellas supuso, hace exactamente dos siglos, el inicio de la última gran etapa en la transformación del edificio, la de la restauración. La fábrica es considerada como objeto a conservar y los esfuerzos desde ese momento se centraron en la recuperación de la arquitectura omeya latente. De este modo, la práctica de la restauración como disciplina se encontró absolutamente influenciada por la Arqueología como única fuente de conocimiento. Las intervenciones buscaban lo original como modo de recuperar espacial y formalmente aquel pasado, concentrándose en los lugares del edificio considerados como esenciales. La declaración del edificio como monumento nacional en 1882 propició que el Estado se hiciera cargo de su mantenimiento y conservación, sustituyendo en esa tarea a los Obispos y Cabildos del siglo XIX, que tuvieron un entendimiento muy avanzado para su época. La llegada del arquitecto Velázquez Bosco en las últimas décadas del siglo XIX supuso un cambio trascendental en la historia del edificio, puesto que recibió un edificio con importantes deterioros y consiguió poner las bases del edificio que hoy contemplamos. El empeño por la recuperación material y espacial devolvió a la Mezquita-Catedral buena parte de su imagen original, reproduciendo con exactitud los modelos hallados en las exploraciones arqueológicas. La llegada de Antonio Flórez tras la muerte de Velázquez Bosco supuso la traslación al edificio del debate disciplinar que se desarrolló en las dos primeras décadas del siglo XX. Flórez procuró un nuevo entendimiento de la intervención, considerando la conservación como actuación prioritaria. En 1926 el Estado reformó la manera en que se atendía al patrimonio con la creación de un sistema de zonas y unos arquitectos a cargo de ellas. La existencia de un nuevo marco legislativo apuntaló esa nueva visión conservativa, avalada por la Carta de Atenas de 1931. Este modelo restauración científica huía de la intervención en estilo y valoraba la necesidad de intervenir de la manera más escueta posible y con un lenguaje diferenciado, basándose en los datos que ofrecía la Arqueología. Por tanto, se continuaba con la valoración del edificio como documento histórico, buscando en este caso una imagen diferenciada de la intervención frente a la actitud mimética de Velázquez. Resulta destacable la manera en la que el historiador Manuel Gómez-Moreno influyó en varias generaciones de arquitectos, arqueólogos e historiadores, tanto en el entendimiento científico de la restauración como en la propia estructura administrativa. La labor desarrollada en el edificio por José Mª Rodríguez Cano primero y Félix Hernández a continuación estuvo influida de manera teórica por el método de Gómez-Moreno, aunque en muchos aspectos su labor no representó una gran diferencia con lo hecho por Velázquez Bosco. La búsqueda de lo original volvió a ser recurrente, pero la carga económica del mantenimiento de un edificio tan extenso conllevó la no realización de muchos de los proyectos más ambiciosos. Esta obsesiva búsqueda de la imagen original del edificio tuvo su última y anacrónica etapa con la intervención de la Dirección General de Arquitectura en los 70. Sin embargo, el agotamiento del modelo científico ya había propiciado un nuevo escenario a nivel europeo, que cristalizó en la Carta de Venecia de 1964 y en una nueva definición del objeto a preservar, más allá del valor como documento histórico. Esta nueva posición teórica tuvo su traslación al modelo restaurador español en el último cuarto de siglo XX, coincidiendo con la Transición. El arquitecto Dionisio Hernández Gil defendió una interpretación distinta a la de los arqueólogos y de los historiadores, que había prevalecido durante todo el siglo. En opinión de Hernández Gil, los problemas de intervención debían enfocarse fundamentalmente como problemas de Arquitectura, abandonando la idea de que solamente podían ser resueltos por especialistas. Esta convicción teórica fue defendida desde la nueva Administración y deparó la utilización de unos criterios de intervención particularizados, provenientes del análisis multifocal de cada situación y no sólo desde el valor de los edificios como documentos históricos. Y este cambio tuvo su traslación a la Mezquita-Catedral con la práctica de Gabriel Ruiz Cabrero y Gabriel Rebollo. En consecuencia con esa nueva perspectiva, aceptaron el edificio que recibieron, sustituyendo la búsqueda de aquella página original por la aceptación de cada una de las páginas de su historia y el respeto a las técnicas constructivas del pasado. La búsqueda de soluciones específicas desde el propio objeto arquitectónico significó la renovada atención a la potente estructura formal-constructiva como origen de toda reflexión. Considerar la Mezquita-Catedral en primer lugar como Arquitectura implicaba la atención a todo tipo de factores además de los históricos, como medio para preservar su autenticidad. Esta tesis pretende demostrar que la práctica de la restauración realizada en la Mezquita-Catedral a lo largo del siglo XX ha evolucionado desde la búsqueda de lo original hasta la búsqueda de lo auténtico, como reflejo de una visión basada en lo arqueológico frente a una renovada visión arquitectónica más completa, que incluye a la anterior. La consideración de la intervención en este edificio como otra página más de su historia y no como la última, significa la reedición de un mecanismo recurrente en la vida del edificio y un nuevo impulso en ese proceso de continua transformación. ABSTRACT The Mosque-Cathedral of Cordoba is a living building. A building transformed by men of different races, cultures and religions during more than 1.200 years old and that, nevertheless, it has continued to be in use all days in that long life. Thus, the building shows to the visitor as a complex architectural object, the result of continuous transformation. This transformation capacity of the buildings is inherent in their own architectural condition, it’s not an exclusive fact of the Mosque-Cathedral. However, in this building that transformation happens with a great intensity, without losing their authenticity. Traditionally, buildings have been adapted to the new requirements of times in a process that looked for laws or principles in order to guide the intervention. Thus, both the successive enlargements of the Mosque of Abd al-Rahman and Christian interventions must assume the preexistence as a working material. So, the architects of the caliph al-Hakam II spoke to their predecessors, complexing the receiving space, as well as Hernan Ruiz got a new organism as result the introduction of his luminous architecture into hispanic-muslim weft. The nineteenth century confirmed the desire to discover the traces of a glorious past that Baroque intervention had silenced, under a uniform space treatment. Exactly two centuries ago, the recovery of these traces meant the start of the last major phase in the transformation of the building: the restoration. The building was considered subject to conserve and since then, efforts focused on the recovery of latent Umayyad architecture. Thus, the practice of restoration as a discipline was absolutely influenced by Archaeology as the only source of knowledge. Interventions were seeking the original as the way to recover that past in a space and formal way, concentrating on essential sites of the building. The statement as a national monument in 1882 prompted the State take charge of its maintenance and preservation, replacing to the nineteenth century Bishops and Cabildos, which had a very advanced understanding for that time. The arrival of the architect Velazquez Bosco in the last decades of the nineteenth century involved a momentous change in the history of the building, since he received a building with significant damage and he achieved the foundations of the building that we can see today. Efforts to a material and space recover returned the Mosque-Cathedral to its original image, accurately reproducing the models found in archaeological explorations. The arrival of Antonio Florez after Velazquez’s death involved the translation of discipline debate, which was developed in the first two decades of the twentieth century. Florez tried a new understanding of the intervention, considering conservation as a priority action. In 1926, the State reformed the way in which heritage was attended, creating a zones system with a few architects in charge of them. The existence of a new legislative framework, underpinned this new conservative vision, supported by the Athens Charter of 1931. This scientific restoration model fleeing from intervention in style and it appreciated the need to intervene in the most concise way, with a distinct language based on the data offered by Archaeology. Therefore, it continued with the appraisement of the building as a historical document, seeking in this case a differentiated image of intervention, against Velazquez mimetic attitude. It is remarkable the way in which the historian Manuel Gomez-Moreno influenced several generations of architects, archaeologists and historians, both in the scientific understanding of the restoration and the administrative structure. The work of Jose Maria Rodriguez Cano first and then Felix Hernandez was theoretically influenced by the Gomez-Moreno’s method, although in many respects their work did not represent a great difference to Velazquez Bosco. The search of the original returned to recur, but the economic charge of maintaining such a large building led to the non-realization of many of the most ambitious projects. This obsessive search for the original image of the building had its last and anachronistic stage with the intervention of the Department of Architecture at 70’s. However, the exhaustion of the scientific model had already led to a new scenario at European level, which crystallized in the Venice Charter of 1964 and a new definition of the object to be preserved beyond the value as a historical document. This new theoretical position had its translation to Spanish restaurateur model in the last quarter of the twentieth century, coinciding with the Transition. The architect Dionisio Hernandez Gil defended a different interpretation from archaeologists and historians, that had prevailed throughout the century. According to Hernandez Gil, the problems of intervention should focus primarily as architectural issues, abandoning the idea that they could only be determined by specialist. This theoretical conviction was defended from the new administration and led to the use of particularized criteria, from a multifocal analysis of each situation. And this change had its translation to the Mosque with the practice of Gabriel Ruiz Cabrero and Gabriel Rebollo. Consistent with this new perspective, they accepted the receiving building, replacing the search on original page for acceptance of all historical pages and respecting the constructive techniques of the past. The search for specific solutions from the architectural object meant the renewed attention to the powerful formal-constructive structure as the origin of all thought. Consider the Mosque-Cathedral as Architecture, involved the attention to all kinds of factors in addition to the historical, as a means to preserve its authenticity. This thesis aims to demonstrate that the practice of restoration in the Mosque-Cathedral throughout the twentieth century has evolved from the search of the original to the search for the authentic, reflecting a vision based on the archaeological against a renewed more complete architectural vision, including the above. Consideration of intervention in this building as another page in its history and not the last one, means the reissue of an own mechanism and a new impetus in that continuous transformation process.
Resumo:
This article describes a knowledge-based method for generating multimedia descriptions that summarize the behavior of dynamic systems. We designed this method for users who monitor the behavior of a dynamic system with the help of sensor networks and make decisions according to prefixed management goals. Our method generates presentations using different modes such as text in natural language, 2D graphics and 3D animations. The method uses a qualitative representation of the dynamic system based on hierarchies of components and causal influences. The method includes an abstraction generator that uses the system representation to find and aggregate relevant data at an appropriate level of abstraction. In addition, the method includes a hierarchical planner to generate a presentation using a model with dis- course patterns. Our method provides an efficient and flexible solution to generate concise and adapted multimedia presentations that summarize thousands of time series. It is general to be adapted to differ- ent dynamic systems with acceptable knowledge acquisition effort by reusing and adapting intuitive rep- resentations. We validated our method and evaluated its practical utility by developing several models for an application that worked in continuous real time operation for more than 1 year, summarizing sen- sor data of a national hydrologic information system in Spain.
Resumo:
The term "Logic Programming" refers to a variety of computer languages and execution models which are based on the traditional concept of Symbolic Logic. The expressive power of these languages offers promise to be of great assistance in facing the programming challenges of present and future symbolic processing applications in Artificial Intelligence, Knowledge-based systems, and many other areas of computing. The sequential execution speed of logic programs has been greatly improved since the advent of the first interpreters. However, higher inference speeds are still required in order to meet the demands of applications such as those contemplated for next generation computer systems. The execution of logic programs in parallel is currently considered a promising strategy for attaining such inference speeds. Logic Programming in turn appears as a suitable programming paradigm for parallel architectures because of the many opportunities for parallel execution present in the implementation of logic programs. This dissertation presents an efficient parallel execution model for logic programs. The model is described from the source language level down to an "Abstract Machine" level suitable for direct implementation on existing parallel systems or for the design of special purpose parallel architectures. Few assumptions are made at the source language level and therefore the techniques developed and the general Abstract Machine design are applicable to a variety of logic (and also functional) languages. These techniques offer efficient solutions to several areas of parallel Logic Programming implementation previously considered problematic or a source of considerable overhead, such as the detection and handling of variable binding conflicts in AND-Parallelism, the specification of control and management of the execution tree, the treatment of distributed backtracking, and goal scheduling and memory management issues, etc. A parallel Abstract Machine design is offered, specifying data areas, operation, and a suitable instruction set. This design is based on extending to a parallel environment the techniques introduced by the Warren Abstract Machine, which have already made very fast and space efficient sequential systems a reality. Therefore, the model herein presented is capable of retaining sequential execution speed similar to that of high performance sequential systems, while extracting additional gains in speed by efficiently implementing parallel execution. These claims are supported by simulations of the Abstract Machine on sample programs.
Resumo:
We present two approaches to cluster dialogue-based information obtained by the speech understanding module and the dialogue manager of a spoken dialogue system. The purpose is to estimate a language model related to each cluster, and use them to dynamically modify the model of the speech recognizer at each dialogue turn. In the first approach we build the cluster tree using local decisions based on a Maximum Normalized Mutual Information criterion. In the second one we take global decisions, based on the optimization of the global perplexity of the combination of the cluster-related LMs. Our experiments show a relative reduction of the word error rate of 15.17%, which helps to improve the performance of the understanding and the dialogue manager modules.
Resumo:
This paper introduces a semantic language developed with the objective to be used in a semantic analyzer based on linguistic and world knowledge. Linguistic knowledge is provided by a Combinatorial Dictionary and several sets of rules. Extra-linguistic information is stored in an Ontology. The meaning of the text is represented by means of a series of RDF-type triples of the form predicate (subject, object). Semantic analyzer is one of the options of the multifunctional ETAP-3 linguistic processor. The analyzer can be used for Information Extraction and Question Answering. We describe semantic representation of expressions that provide an assessment of the number of objects involved and/or give a quantitative evaluation of different types of attributes. We focus on the following aspects: 1) parametric and non-parametric attributes; 2) gradable and non-gradable attributes; 3) ontological representation of different classes of attributes; 4) absolute and relative quantitative assessment; 5) punctual and interval quantitative assessment; 6) intervals with precise and fuzzy boundaries
Resumo:
We present two approaches to cluster dialogue-based information obtained by the speech understanding module and the dialogue manager of a spoken dialogue system. The purpose is to estimate a language model related to each cluster, and use them to dynamically modify the model of the speech recognizer at each dialogue turn. In the first approach we build the cluster tree using local decisions based on a Maximum Normalized Mutual Information criterion. In the second one we take global decisions, based on the optimization of the global perplexity of the combination of the cluster-related LMs. Our experiments show a relative reduction of the word error rate of 15.17%, which helps to improve the performance of the understanding and the dialogue manager modules.
Resumo:
In this paper, we describe new results and improvements to a lan-guage identification (LID) system based on PPRLM previously introduced in [1] and [2]. In this case, we use as parallel phone recognizers the ones provided by the Brno University of Technology for Czech, Hungarian, and Russian lan-guages, and instead of using traditional n-gram language models we use a lan-guage model that is created using a ranking with the most frequent and discrim-inative n-grams. In this language model approach, the distance between the ranking for the input sentence and the ranking for each language is computed, based on the difference in relative positions for each n-gram. This approach is able to model reliably longer span information than in traditional language models obtaining more reliable estimations. We also describe the modifications that we have being introducing along the time to the original ranking technique, e.g., different discriminative formulas to establish the ranking, variations of the template size, the suppression of repeated consecutive phones, and a new clus-tering technique for the ranking scores. Results show that this technique pro-vides a 12.9% relative improvement over PPRLM. Finally, we also describe re-sults where the traditional PPRLM and our ranking technique are combined.
Resumo:
We present an approach to adapt dynamically the language models (LMs) used by a speech recognizer that is part of a spoken dialogue system. We have developed a grammar generation strategy that automatically adapts the LMs using the semantic information that the user provides (represented as dialogue concepts), together with the information regarding the intentions of the speaker (inferred by the dialogue manager, and represented as dialogue goals). We carry out the adaptation as a linear interpolation between a background LM, and one or more of the LMs associated to the dialogue elements (concepts or goals) addressed by the user. The interpolation weights between those models are automatically estimated on each dialogue turn, using measures such as the posterior probabilities of concepts and goals, estimated as part of the inference procedure to determine the actions to be carried out. We propose two approaches to handle the LMs related to concepts and goals. Whereas in the first one we estimate a LM for each one of them, in the second one we apply several clustering strategies to group together those elements that share some common properties, and estimate a LM for each cluster. Our evaluation shows how the system can estimate a dynamic model adapted to each dialogue turn, which helps to improve the performance of the speech recognition (up to a 14.82% of relative improvement), which leads to an improvement in both the language understanding and the dialogue management tasks.
Resumo:
This paper presents a description of our system for the Albayzin 2012 LRE competition. One of the main characteristics of this evaluation was the reduced number of available files for training the system, especially for the empty condition where no training data set was provided but only a development set. In addition, the whole database was created from online videos and around one third of the training data was labeled as noisy files. Our primary system was the fusion of three different i-vector based systems: one acoustic system based on MFCCs, a phonotactic system using trigrams of phone-posteriorgram counts, and another acoustic system based on RPLPs that improved robustness against noise. A contrastive system that included new features based on the glottal source was also presented. Official and postevaluation results for all the conditions using the proposed metrics for the evaluation and the Cavg metric are presented in the paper.
Resumo:
The aim of the paper is to discuss the use of knowledge models to formulate general applications. First, the paper presents the recent evolution of the software field where increasing attention is paid to conceptual modeling. Then, the current state of knowledge modeling techniques is described where increased reliability is available through the modern knowledge acquisition techniques and supporting tools. The KSM (Knowledge Structure Manager) tool is described next. First, the concept of knowledge area is introduced as a building block where methods to perform a collection of tasks are included together with the bodies of knowledge providing the basic methods to perform the basic tasks. Then, the CONCEL language to define vocabularies of domains and the LINK language for methods formulation are introduced. Finally, the object oriented implementation of a knowledge area is described and a general methodology for application design and maintenance supported by KSM is proposed. To illustrate the concepts and methods, an example of system for intelligent traffic management in a road network is described. This example is followed by a proposal of generalization for reuse of the resulting architecture. Finally, some concluding comments are proposed about the feasibility of using the knowledge modeling tools and methods for general application design.
Resumo:
Background: Early and effective identification of developmental disorders during childhood remains a critical task for the international community. The second highest prevalence of common developmental disorders in children are language delays, which are frequently the first symptoms of a possible disorder. Objective: This paper evaluates a Web-based Clinical Decision Support System (CDSS) whose aim is to enhance the screening of language disorders at a nursery school. The common lack of early diagnosis of language disorders led us to deploy an easy-to-use CDSS in order to evaluate its accuracy in early detection of language pathologies. This CDSS can be used by pediatricians to support the screening of language disorders in primary care. Methods: This paper details the evaluation results of the ?Gades? CDSS at a nursery school with 146 children, 12 educators, and 1 language therapist. The methodology embraces two consecutive phases. The first stage involves the observation of each child?s language abilities, carried out by the educators, to facilitate the evaluation of language acquisition level performed by a language therapist. Next, the same language therapist evaluates the reliability of the observed results. Results: The Gades CDSS was integrated to provide the language therapist with the required clinical information. The validation process showed a global 83.6% (122/146) success rate in language evaluation and a 7% (7/94) rate of non-accepted system decisions within the range of children from 0 to 3 years old. The system helped language therapists to identify new children with potential disorders who required further evaluation. This process will revalidate the CDSS output and allow the enhancement of early detection of language disorders in children. The system does need minor refinement, since the therapists disagreed with some questions from the CDSS knowledge base (KB) and suggested adding a few questions about speech production and pragmatic abilities. The refinement of the KB will address these issues and include the requested improvements, with the support of the experts who took part in the original KB development. Conclusions: This research demonstrated the benefit of a Web-based CDSS to monitor children?s neurodevelopment via the early detection of language delays at a nursery school. Current next steps focus on the design of a model that includes pseudo auto-learning capacity, supervised by experts.
Resumo:
We present a methodology for legacy language resource adaptation that generates domain-specific sentiment lexicons organized around domain entities described with lexical information and sentiment words described in the context of these entities. We explain the steps of the methodology and we give a working example of our initial results. The resulting lexicons are modelled as Linked Data resources by use of established formats for Linguistic Linked Data (lemon, NIF) and for linked sentiment expressions (Marl), thereby contributing and linking to existing Language Resources in the Linguistic Linked Open Data cloud.