914 resultados para search and matching


Relevância:

80.00% 80.00%

Publicador:

Resumo:

BACKGROUND With the advent of new and more efficient anti-androgen drugs targeting androgen receptor (AR) in breast cancer (BC) is becoming an increasingly important area of investigation. This would potentially be most useful in triple negative BC (TNBC), where better therapies are still needed. The assessment of AR status is generally performed on the primary tumor even if the tumor has already metastasized. Very little is known regarding discrepancies of AR status during tumor progression. To determine the prevalence of AR positivity, with emphasis on TNBCs, and to investigate AR status during tumor progression, we evaluated a large series of primary BCs and matching metastases and recurrences. METHODS AR status was performed on 356 primary BCs, 135 matching metastases, and 12 recurrences using a next-generation Tissue Microarray (ngTMA). A commercially available AR antibody was used to determine AR-status by immunohistochemistry. AR positivity was defined as any nuclear staining in tumor cells ≥1 %. AR expression was correlated with pathological tumor features of the primary tumor. Additionally, the concordance rate of AR expression between the different tumor sites was determined. RESULTS AR status was positive in: 87 % (307/353) of primary tumors, 86.1 % (105/122) of metastases, and in 66.7 % (8/12) of recurrences. TNBC tested positive in 11.4 %, (4/35) of BCs. A discrepant result was seen in 4.3 % (5/117) of primary BC and matching lymph node (LN) metastases. Three AR negative primary BCs were positive in the matching LN metastasis, representing 17.6 % of all negative BCs with lymph node metastases (3/17). Two AR positive primary BCs were negative in the matching LN metastasis, representing 2.0 % of all AR positive BCs with LN metastases (2/100). No discrepancies were seen between primary BC and distant metastases or recurrence (n = 17). CONCLUSIONS Most primary (87 %) and metastasized (86.1 %) BCs are AR positive including a significant fraction of TNBCs (11.4 %). Further, AR status is highly conserved during tumor progression and a change only occurs in a small fraction (4.1 %). Our study supports the notion that targeting AR could be effective for many BC patients and that re-testing of AR status in formerly negative or mixed type BC's is recommended.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

This paper presents a new tool for large-area photo-mosaicking (LAPM tool). This tool was developed specifically for the purpose of underwater mosaicking, and it is aimed at providing end-user scientists with an easy and robust way to construct large photo-mosaics from any set of images. It is notably capable of constructing mosaics with an unlimited number of images on any modern computer (minimum 1.30 GHz, 2 GB RAM). The mosaicking process can rely on both feature matching and navigation data. This is complemented by an intuitive graphical user interface, which gives the user the ability to select feature matches between any pair of overlapping images. Finally, mosaic files are given geographic attributes that permit direct import into ArcGIS. So far, the LAPM tool has been successfully used to construct geo-referenced photo-mosaics with photo and video material from several scientific cruises. The largest photo-mosaic contained more than 5000 images for a total area of about 105,000 m**2. This is the first article to present and to provide a finished and functional program to construct large geo-referenced photo-mosaics of the seafloor using feature detection and matching techniques. It also presents concrete examples of photo-mosaics produced with the LAPM tool.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Since the early days of logic programming, researchers in the field realized the potential for exploitation of parallelism present in the execution of logic programs. Their high-level nature, the presence of nondeterminism, and their referential transparency, among other characteristics, make logic programs interesting candidates for obtaining speedups through parallel execution. At the same time, the fact that the typical applications of logic programming frequently involve irregular computations, make heavy use of dynamic data structures with logical variables, and involve search and speculation, makes the techniques used in the corresponding parallelizing compilers and run-time systems potentially interesting even outside the field. The objective of this article is to provide a comprehensive survey of the issues arising in parallel execution of logic programming languages along with the most relevant approaches explored to date in the field. Focus is mostly given to the challenges emerging from the parallel execution of Prolog programs. The article describes the major techniques used for shared memory implementation of Or-parallelism, And-parallelism, and combinations of the two. We also explore some related issues, such as memory management, compile-time analysis, and execution visualization.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Interlinking text documents with Linked Open Data enables the Web of Data to be used as background knowledge within document-oriented applications such as search and faceted browsing. As a step towards interconnecting the Web of Documents with the Web of Data, we developed DBpedia Spotlight, a system for automatically annotating text documents with DBpedia URIs. DBpedia Spotlight allows users to congure the annotations to their specic needs through the DBpedia Ontology and quality measures such as prominence, topical pertinence, contextual ambiguity and disambiguation condence. We compare our approach with the state of the art in disambiguation, and evaluate our results in light of three baselines and six publicly available annotation systems, demonstrating the competitiveness of our system. DBpedia Spotlight is shared as open source and deployed as a Web Service freely available for public use.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Cultural content on the Web is available in various domains (cultural objects, datasets, geospatial data, moving images, scholarly texts and visual resources), concerns various topics, is written in different languages, targeted to both laymen and experts, and provided by different communities (libraries, archives museums and information industry) and individuals (Figure 1). The integration of information technologies and cultural heritage content on the Web is expected to have an impact on everyday life from the point of view of institutions, communities and individuals. In particular, collaborative environment scan recreate 3D navigable worlds that can offer new insights into our cultural heritage (Chan 2007). However, the main barrier is to find and relate cultural heritage information by end-users of cultural contents, as well as by organisations and communities managing and producing them. In this paper, we explore several visualisation techniques for supporting cultural interfaces, where the role of metadata is essential for supporting the search and communication among end-users (Figure 2). A conceptual framework was developed to integrate the data, purpose, technology, impact, and form components of a collaborative environment, Our preliminary results show that collaborative environments can help with cultural heritage information sharing and communication tasks because of the way in which they provide a visual context to end-users. They can be regarded as distributed virtual reality systems that offer graphically realised, potentially infinite, digital information landscapes. Moreover, collaborative environments also provide a new way of interaction between an end-user and a cultural heritage data set. Finally, the visualisation of metadata of a dataset plays an important role in helping end-users in their search for heritage contents on the Web.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

A compact planar array with parasitic elements is studied to be used in MIMO systems. Classical compact arrays suffer from high coupling which makes correlation and matching efficiency to be worse. A proper matching network improves these lacks although its bandwidth is low and may increase the antenna size. The proposed antenna makes use of parasitic elements to improve both correlation and efficiency. A specific software based on MoM has been developed to analyze radiating structures with several feed points. The array is optimized through a Genetic Algorithm to determine parasitic elements position in order to fulfill different figures of merit. The proposed design provides the required correlation and matching efficiency to have a good performance over a significant bandwidth.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Interlinking text documents with Linked Open Data enables the Web of Data to be used as background knowledge within document-oriented applications such as search and faceted browsing. As a step towards interconnecting the Web of Documents with the Web of Data, we developed DBpedia Spotlight, a system for automatically annotating text documents with DBpedia URIs. DBpedia Spotlight allows users to configure the annotations to their specific needs through the DBpedia Ontology and quality measures such as prominence, topical pertinence, contextual ambiguity and disambiguation confidence. We compare our approach with the state of the art in disambiguation, and evaluate our results in light of three baselines and six publicly available annotation systems, demonstrating the competitiveness of our system. DBpedia Spotlight is shared as open source and deployed as a Web Service freely available for public use.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

This article presents the model of a multi-agent system (SMAF), which objectives are the input of fuzzy incidents as the human experts express them with different severities degrees and the further search and suggestion of solutions. The solutions will be later confirm or not by the users. This model was designed, implemented and tested in the telecommunications field, with heterogeneous agents in a cooperative model. In the design, different abstract levels where considered, according to the agents? objectives, their ways to carry it out and the environment in which they act. Each agent is modeled with different spectrum of the knowledge base

Relevância:

80.00% 80.00%

Publicador:

Resumo:

This article presents the model and implementation of a multiagent fuzzy system (MAFS), to automate the search of solutions of incidents in telecommunications, expressed by the users in an imprecise way and, later, registered in a a knowledge base keeping their original vaguenesses and the relationships between the incidents considered as ancestors and descendants. The process of the fuzzy incidents, no matter their causes, is based on the application of a formula which transforms the intervals of the fuzzy incidents to a computational language and in the interaction between the different kinds of software agents and the humans. To search and suggest solutions of the incident originally stated, a search algorithm is used and illustrated with an example. The preliminary results obtained show the users' satisfaction, in a great percentage of the presented cases. The system is adaptive and allows to record new solutions for future users.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Background: Cognitive skills training for minimally invasive surgery has traditionally relied upon diverse tools, such as seminars or lectures. Web technologies for e-learning have been adopted to provide ubiquitous training and serve as structured repositories for the vast amount of laparoscopic video sources available. However, these technologies fail to offer such features as formative and summative evaluation, guided learning, or collaborative interaction between users. Methodology: The "TELMA" environment is presented as a new technology-enhanced learning platform that increases the user's experience using a four-pillared architecture: (1) an authoring tool for the creation of didactic contents; (2) a learning content and knowledge management system that incorporates a modular and scalable system to capture, catalogue, search, and retrieve multimedia content; (3) an evaluation module that provides learning feedback to users; and (4) a professional network for collaborative learning between users. Face validation of the environment and the authoring tool are presented. Results: Face validation of TELMA reveals the positive perception of surgeons regarding the implementation of TELMA and their willingness to use it as a cognitive skills training tool. Preliminary validation data also reflect the importance of providing an easy-to-use, functional authoring tool to create didactic content. Conclusion: The TELMA environment is currently installed and used at the Jesús Usón Minimally Invasive Surgery Centre and several other Spanish hospitals. Face validation results ascertain the acceptance and usefulness of this new minimally invasive surgery training environment.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Problem-based learning has been applied over the last three decades to a diverse range of learning environments. In this educational approach, different problems are posed to the learners so that they can develop different solutions while learning about the problem domain. When applied to conceptual modelling, and particularly to Qualitative Reasoning, the solutions to problems are models that represent the behaviour of a dynamic system. The learner?s task then is to bridge the gap between their initial model, as their first attempt to represent the system, and the target models that provide solutions to that problem. We propose the use of semantic technologies and resources to help in bridging that gap by providing links to terminology and formal definitions, and matching techniques to allow learners to benefit from existing models.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Desde hace ya muchos años, uno de los servicios de telecomunicaciones más demandado por los españoles ha sido la televisión de pago, complementando y ampliando la oferta de contenidos audiovisuales que habitualmente son ofrecidos de manera gratuita por la televisión analógica y recientemente por la televisión digital terrestre o TDT. Estos servicios de video, han sido tradicionalmente ofrecidos por operadores satélites, operadores de cable u otros operadores de telecomunicaciones con los que a través de una conexión de datos (ADSL, VDSL o fibra óptica), ofrecían sus contenidos a través de IP. La propia evolución y mejora de la tecnología utilizada para la emisión de contenidos sobre IP, ha permitido que a día de hoy, la televisión se conciba como un servicio Over The Top (OTT) ajeno al medio de transmisión, permitiendo a cualquier agente, distribuir sus contenidos audiovisuales de manera sencilla y a todos sus clientes en todas las partes del mundo; siendo solamente necesario disponer de una conexión a internet. De esta manera, el proyecto desarrollado va a girar en torno a la herramienta StormTest de la empresa S3Group, comprada por CENTUM Solutions (empresa especializada en ofrecer servicio de ingeniería para sistema de comunicaciones, control e inteligencia de señal) con el objetivo de satisfacer las necesidades de sus clientes y con la que en definitiva se ha contado para la realización de este proyecto. El principal objetivo de este proyecto es la definición e implementación de un banco de pruebas que permita optimizar los procesos de validación técnica, mejorando los tiempos de ejecución y concentrando la actividad de los ingenieros en tareas de mayor valor. Para la realización de este proyecto, se han fijado diversos objetivos necesarios para el desarrollo de este tipo de actividades. Los principales son los siguientes:  Análisis de la problemática actual: donde en los procesos de aceptación técnica se dedica muchas horas de trabajo para la realización de pruebas repetitivas y de poco valor las cuales se pueden automatizar por herramientas existentes en el mercado.  Búsqueda y selección de una herramienta que satisfaga las necesidades de pruebas.  Instalación en los laboratorios.  Configuración y adaptación de la herramienta a las necesidades y proyectos específicos. Con todo ello, este proyecto cubrirá los siguientes logros:  Reducir los tiempos de ejecución de las campañas de pruebas, gracias a la automatización de gran parte ellas.  Realizar medidas de calidad subjetiva y objetiva complejas, imposibles de ejecutar a través de las personas. Mejorar y automatizar los sistemas de reporte de resultados. Abstract: Many years ago, one of the telecommunications services most demanded in Spain has been pay television, complementing and extending the offer of audiovisual content which are usually offered for free by analog tv and recently by digital terrestrial televisión or TDT. These video services, have been traditionally offered by satellite operators, cable operators or other telecommunications operators that through a data connection (ADSL,VDSL or fiber optic), offered its content over IP. The evolution and improvement of the technology used for broadcasting over IP, has allowed that to date, television is conceived as a service Over The Top (OTT), not dependent on the transmission medium, allowing any agent to distribute audiovisual content in a very simple way and to all its customers in all parts of the world; being only necessary to have an decent internet connection. In this way, the project will have relation with S3Group’s StormTest tool, bought by CENTUM Solutions (company specialized in engineering services for communications, control and signal intelligence system) with the aim of satisfying the needs of its customers and which ultimately has counted for the realization of this project. The main objective of this project is the definition and implementation of a test bench that allows to optimize the processes of technical validation, improving execution times and concentrating the activities of engineers on higher value tasks. For the realization of this project, it has been defined several objectives necessary for the development of this type of activity. The most important tones are listed below:  Analysis of the current situation: where in technical acceptance processes it is dedicated many hours of work for the completion of repetitive testing and without value which can be automated by tools available on the market  Search and selection of a tool that meets the needs of testing.  Installation on the laboratories.  Configuration and customization of the tool to specific projects. With all this, this project will cover the following achievements: Reduce the execution time of the testing campaigns, thanks to the automation of many of them.  Measurements of subjective and objective quality tests, impossible to run with engineers (due to subjective perception). Improve and automate reporting of results systems

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Cultural heritage is a complex and diverse concept, which brings together a wide domain of information. Resources linked to a cultural heritage site may consist of physical artefacts, books, works of art, pictures, historical maps, aerial photographs, archaeological surveys and 3D models. Moreover, all these resources are listed and described by a set of a variety of metadata specifications that allow their online search and consultation on the most basic characteristics of them. Some examples include Norma ISO 19115, Dublin Core, AAT, CDWA, CCO, DACS, MARC, MoReq, MODS, MuseumDat, TGN, SPECTRUM, VRA Core and Z39.50. Gateways are in place to fit in these metadata standards into those used in a SDI (ISO 19115 or INSPIRE), but substantial work still remains to be done for the complete incorporation of cultural heritage information. Therefore, the aim of this paper is to demonstrate how the complexity of cultural heritage resources can be dealt with by a visual exploration of their metadata within a 3D collaborative environment. The 3D collaborative environments are promising tools that represent the new frontier of our capacity of learning, understanding, communicating and transmitting culture.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

RESUMEN Las aplicaciones de los Sistemas de Información Geográfica (SIG) a la Arqueología, u otra disciplina humanística no son una novedad. La evolución de los mismos hacia sistemas distribuidos e interoperables, y estructuras donde las políticas de uso, compartido y coordinado de los datos sí lo son, estando todos estos aspectos contemplados en la Infraestructura de Datos Espaciales. INSPIRE es el máximo exponente europeo en cuestiones de iniciativa y marco legal en estos aspectos. La metodología arqueológica recopila y genera gran cantidad de datos, y entre los atributos o características intrínsecas están la posición y el tiempo, aspectos que tradicionalmente explotan los SIG. Los datos se catalogan, organizan, mantienen, comparten y publican, y los potenciales consumidores comienzan a tenerlos disponibles. Toda esta información almacenada de forma tradicional en fichas y posteriormente en bases de datos relacionadas alfanuméricas pueden ser considerados «metadatos» en muchos casos por contener información útil para más usuarios en los procesos de descubrimiento, y explotación de los datos. Además estos datos también suelen ir acompañados de información sobre ellos mismos, que describe su especificaciones, calidad, etc. Cotidianamente usamos los metadatos: ficha bibliográfica del libro o especificaciones de un ordenador. Pudiéndose definir como: «información descriptiva sobre el contexto, calidad, condición y características de un recurso, dato u objeto que tiene la finalidad de facilitar su recuperación, identificación,evaluación, preservación y/o interoperabilidad». En España existe una iniciativa para estandarizar la descripción de los metadatos de los conjuntos de datos geoespaciales: Núcleo Español de Metadatos (NEM), los mismos contienen elementos para la descripción de las particularidades de los datos geográficos, que incluye todos los registros obligatorios de la Norma ISO19115 y del estudio de metadatos Dublin Core, tradicionalmente usado en contextos de Biblioteconomía. Conscientes de la necesidad de los metadatos, para optimizar la búsqueda y recuperación de los datos, se pretende formalizar la documentación de los datos arqueológicos a partir de la utilización del NEM, consiguiendo así la interoperabilidad de la información arqueológica. SUMMARY The application of Geographical Information Systems (GIS) to Archaeology and other social sciences is not new. Their evolution towards inter-operating, distributed systems, and structures in which policies for shared and coordinated data use are, and all these aspects are included in the Spatial Data Infrastructure (SDI). INSPIRE is the main European exponent in matters related to initiative and legal frame. Archaeological methodology gathers and creates a great amount of data, and position and time, aspects traditionally exploited by GIS, are among the attributes or intrinsic characteristics. Data are catalogued, organised, maintained, shared and published, and potential consumers begin to have them at their disposal. All this information, traditionally stored as cards and later in relational alphanumeric databases may be considered «metadata» in many cases, as they contain information that is useful for more users in the processes of discovery and exploitation of data. Moreover, this data are often accompanied by information about themselves, describing its especifications, quality, etc. We use metadata very often: in a book’s bibliographical card, or in the description of the characteristics of a computer. They may be defined as «descriptive information regarding the context, quality, condition and characteristics of a resource, data or object with the purpose of facilitating is recuperation, identification, evaluation, preservation and / interoperability.» There is an initiative in Spain to standardise the description of metadata in sets of geo-spatial data: the Núcleo Español de Metadatos (Spanish Metadata Nucleus), which contains elements for the description of the particular characteristics of geographical data, includes all the obligatory registers from the ISO Norm 19115 and from the metadata study Dublin Core, traditionally used in library management. Being aware of the need of metadata, to optimise the search and retrieval of data, the objective is to formalise the documentation of archaeological data from the Núcleo Español de Metadatos (Spanish Metadata Nucleus), thus obtaining the interoperability of the archaeological information.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

El propósito de este proyecto de �fin de carrera es la caracterización e instrumentación de un sensor de ultrasonidos modelado por el tutor de este proyecto: Don César Briso Rodrí��guez. Una vez realizado el modelado de dicho sensor, simulando tanto sus caracter�í�sticas f�í�sicas, como sus caracterí��sticas eléctricas, se procede a la intrumentación y uso del mismo. La parte de intrumentaci�ón incluye tanto la electrónica que ser��á necesaria para la excitación del piezoeléctrico, en el modo de emisi�ón, como para la recepción de los pulsos el�éctricos generados por el sensor, como respuesta a los ecos recibidos, y su adecuación a niveles de señal correctos para la adquisici�ón, en el modo de escucha. Tras la adecuaci�ón de las señales para la adquisici�ón, éstas ser�án digitalizadas, tratadas y representadas por pantalla en un PC, a trav�es de una tarjeta de adquisición de datos por puerto USB encargada del muestreo de las señales de respuesta ya tratadas y su posterior enví��o al software de control y representaci�ón desarrollado en este proyecto. El entorno de usuario, el software de control de la tarjeta de adquisición y el software de tratamiento y representaci�ón se ha desarrollado con Visual Basic 2008 y las utilidades gr�áfi�cas de las librer��ías OpenGL. ABSTRACT The purpose of this project is to limit the characterization and implementation of an ultrasonic sensor modeled by Mr. C�ésar Briso Rodr��íguez. Once the sensor modeling by simulating physical characteristics and electrical characteristics, we proceed to the instrumentation and use. This section includes electronic instrumentation that would be necessary for the piezoelectric excitation in the emission mode and for receiving electrical pulses generated by the sensor in response to the received echoes, and matching signal levels right to acquire, in the reception mode. After the adjustment of the signals for the acquisition, these signals will be digitalized, processed and represented on the screen on a PC through a data acquisition card by USB port. Acquisition card is able to sample the response signals and transmit the samples to representation and control software developed in this project. The user interface, the acquisition card control software and processing and representation software has been developed with Visual Basic 2008 and OpenGL graphical libraries.