15 resultados para Information access
em Universidad Politécnica de Madrid
Resumo:
The main goal of the bilingual and monolingual participation of the MIRACLE team in CLEF 2004 was to test the effect of combination approaches on information retrieval. The starting point was a set of basic components: stemming, transformation, filtering, generation of n-grams, weighting and relevance feedback. Some of these basic components were used in different combinations and order of application for document indexing and for query processing. A second order combination was also tested, mainly by averaging or selective combination of the documents retrieved by different approaches for a particular query.
Resumo:
This paper describes the first set of experiments defined by the MIRACLE (Multilingual Information RetrievAl for the CLEf campaign) research group for some of the cross language tasks defined by CLEF. These experiments combine different basic techniques, linguistic-oriented and statistic-oriented, to be applied to the indexing and retrieval processes.
Resumo:
ImageCLEF is a pilot experiment run at CLEF 2003 for cross language image retrieval using textual captions related to image contents. In this paper, we describe the participation of the MIRACLE research team (Multilingual Information RetrievAl at CLEF), detailing the different experiments and discussing their preliminary results.
Resumo:
The Web has witnessed an enormous growth in the amount of semantic information published in recent years. This growth has been stimulated to a large extent by the emergence of Linked Data. Although this brings us a big step closer to the vision of a Semantic Web, it also raises new issues such as the need for dealing with information expressed in different natural languages. Indeed, although the Web of Data can contain any kind of information in any language, it still lacks explicit mechanisms to automatically reconcile such information when it is expressed in different languages. This leads to situations in which data expressed in a certain language is not easily accessible to speakers of other languages. The Web of Data shows the potential for being extended to a truly multilingual web as vocabularies and data can be published in a language-independent fashion, while associated language-dependent (linguistic) information supporting the access across languages can be stored separately. In this sense, the multilingual Web of Data can be realized in our view as a layer of services and resources on top of the existing Linked Data infrastructure adding i) linguistic information for data and vocabularies in different languages, ii) mappings between data with labels in different languages, and iii) services to dynamically access and traverse Linked Data across different languages. In this article we present this vision of a multilingual Web of Data. We discuss challenges that need to be addressed to make this vision come true and discuss the role that techniques such as ontology localization, ontology mapping, and cross-lingual ontology-based information access and presentation will play in achieving this. Further, we propose an initial architecture and describe a roadmap that can provide a basis for the implementation of this vision.
Resumo:
With the advancement of Information and Communication Technology ICT which favors increasingly fast, easy, and accessible communication for all and which can reach large groups of people, there have been changes, in recent years in our society that have modified the way we interact, communicate and transmit information. Access to this, it is possible, not only through computers situated in a fixed location, but new mobile devices make it available, wherever the user happens to be located. Now, information "travels" with the user. These forms of communication, transmission and access to information, have also affected the way to conceive and manage business. To these new forms of business that the Internet has brought, is now added the concept of companies in the Cloud Computing ClC. The ClC technology is based on the supply and consumption of services on demand and pay per use, and it gives a 180 degree turn to the business management concept. Small and large businesses may use the latest developments in ICT, to manage their organizations without the need for expensive investments in them. This will enable enterprises to focus more specifically within the scope of their business, leaving the ICT control to the experts. We believe that education can also and should benefit from these new philosophies. ?Due to the global economic crisis in general and each country in particular, economic cutbacks have come to most universities. These are seen in the need to raise tuition rates, which makes increasingly fewer students have the opportunity to pursue higher education?. In this paper we propose using ClC technologies in universities and we make a dissertation on the advantages that it can provide to both: universities and students. For the universities, we expose two focuses, one: ?to reorganize university ICT structures with the ClC philosophy? and the other one, ?to extend the offer of the university education with education on demand?. Regarding the former we propose to use public or private Clouds, to reuse resources across the education community, to save costs on infrastructure investment, in upgrades and in maintenance of ICT, and paying only for what you use and with the ability to scale according to needs. Regarding the latter, we propose an educational model in the ClC, to increase the current university offerings, using educational units in the form of low-cost services and where students pay only for the units consumed on demand. For the students, they could study at any university in the world (virtually), from anywhere, without travel costs: money and time, and what is most important paying only for what they consume. We think that this proposal of education on demand may represent a great change in the current educational model, because strict registration deadlines disappear, and also the problem of economically disadvantaged students, who will not have to raise large amounts of money for an annual tuition. Also it will decrease the problem of loss of the money invested in an enrollment when the student dropout. In summary we think that this proposal is interesting for both, universities and students, we aim for "Higher education from anywhere, with access from any mobile device, at any time, without requiring large investments for students, and with reuse and optimization of resources by universities. Cost by consumption and consumption by service?. We argue for a Universal University "wisdom and knowledge accessible to all?
Resumo:
This paper describes our participation at the RepLab 2014 reputation dimensions scenario. Our idea was to evaluate the best combination strategy of a machine learning classifier with a rule-based algorithm based on logical expressions of terms. Results show that our baseline experiment using just Naive Bayes Multinomial with a term vector model representation of the tweet text is ranked second among runs from all participants in terms of accuracy.
Resumo:
This paper describes our participation at PAN 2014 author profiling task. Our idea was to define, develop and evaluate a simple machine learning classifier able to guess the gender and the age of a given user based on his/her texts, which could become part of the solution portfolio of the company. We were interested in finding not the best possible classifier that achieves the highest accuracy, but to find the optimum balance between performance and throughput using the most simple strategy and less dependent of external systems. Results show that our software using Naive Bayes Multinomial with a term vector model representation of the text is ranked quite well among the rest of participants in terms of accuracy.
Resumo:
Una de las características de la cartografía y SIG Participativos (SIGP) es incluir en sus métodos a la sociedad civil para aportar contenidos cualitativos a la información de sus territorios. Sin embargo no sólo se trata de datos, sino de los efectos que pueden tener estas prácticas sobre el territorio y su sociedad. El acceso a esa información se ve reducida en contraste con el incremento de información difundida a través de servicios de visualización, geoinformación y cartografía on-line. Todo esto hace que sea necesario el análisis del alcance real de las metodologías participativas en el uso de Información Geográfica (IG) y la comparación desde distintos contextos geográficos. También es importante conocer los beneficios e inconvenientes del acceso a la información para el planeamiento; desde la visibilidad de muchos pueblos desapercibidos en zonas rurales y periféricas, hasta la influencia en programas de gobierno sobre la gestión del territorio pasando por el conocimiento local espacial. El análisis se centró en los niveles de participación de la sociedad civil y sus grados de accesibilidad a la información (acceso y uso), dentro del estudio de los SIGP, Participatory Mapping, además se estudió de los TIG (Tecnologías de Información Geográfica), cartografías on-line (geoweb) y plataformas de geovisualización espacial, como recursos de Neocartografía. En este sentido, se realizó un trabajo de campo de cartografía participativa en Bolivia, se evaluaron distintos proyectos SIGP en países del norte y sur (comparativa de contextos en países en desarrollo) y se analizaron los resultados del cruce de las distintas variables.(validación, accesibilidad, verificación de datos, valor en la planificación e identidad) La tesis considera que ambos factores (niveles de participación y grado de accesibilidad) afectan a la (i) validación, verificación y calidad de los datos, la (ii) valor analítico en la planificación, y al (iii) modelo de identidad de un lugar, y que al ser tratados de forma integral, constituyen el valor añadido que los SIGP pueden aportar para lograr una planificación efectiva. Asimismo se comprueba, que la dimensión participativa en los SIGP varía según el contexto, la centralización de sus actores e intereses sectoriales. La información resultante de las prácticas SIGP tiende a estar restringida por la falta de legislaciones y por la ausencia de formatos estándar, que limitan la difusión e intercambio de la información. Todo esto repercute en la efectividad de una planificación estratégica y en la viabilidad de la implementación de cualquier proyecto sobre el territorio, y en consecuencia sobre los niveles de desarrollo de un país. Se confirma la hipótesis de que todos los elementos citados en los SIGP y mapeo participativo actuarán como herramientas válidas para el fortalecimiento y la eficacia en la planificación sólo si están interconectadas y vinculadas entre sí. Se plantea una propuesta metodológica ante las formas convencionales de planificación (nueva ruta del planeamiento; que incluye el intercambio de recursos y determinación participativa local antes de establecer la implementación), con ello, se logra incorporar los beneficios de las metodologías participativas en el manejo de la IG y los SIG (Sistemas de Información Geográfica) como instrumentos estratégicos para el desarrollo de la identidad local y la optimización en los procesos de planeamiento y estudios del territorio. Por último, se fomenta que en futuras líneas de trabajo los mapas de los SIGP y la cartografía participativa puedan llegar a ser instrumentos visuales representativos que transfieran valores identitarios del territorio y de su sociedad, y de esta manera, ayudar a alcanzar un mayor conocimiento, reconocimiento y valoración de los territorios para sus habitantes y sus planificadores. ABSTRACT A feature of participatory mapping and PGIS is to include the participation of the civil society, to provide qualitative information of their territories. However, focus is not only data, but also the effects that such practices themselves may have on the territory and their society. Access to this information is reduced in contrast to the increase of information disseminated through visualization services, geoinformation, and online cartography. Thus, the analysis of the real scope of participatory methodologies in the use of Geographic Information (GI) is necessary, including the comparison of different geographical contexts. It is also important to know the benefits and disadvantages of access to information needed for planning in different contexts, ranging from unnoticed rural areas and suburbs to influence on government programs on land management through local spatial knowledge. The analysis focused on the participation levels of civil society and the degrees of accessibility of the information (access and use) within the study of Participatory GIS (PGIS). In addition, this work studies GIT (Geographic Information Technologies), online cartographies (Geoweb) and platforms of spatial geovisualization, as resources of Neocartography. A participatory cartographic fieldwork was carried out in Bolivia. Several PGIS projects were evaluated in Northern and Southern countries (comparatively with the context of developing countries), and the results were analyzed for each these different variables. (validation, accessibility, verification,value, identity). The thesis considers that both factors (participation levels and degree of accessibility) affect the (i) validation, verification and quality of the data, (ii) analytical value for planning, and (iii) the identity of a place. The integrated management of all the above cited criteria constitutes an added value that PGISs can contribute to reach an effective planning. Also, it confirms the participatory dimension of PGISs varies according to the context, the centralization of its actors, and to sectorial interests. The resulting information from PGIS practices tends to be restricted by the lack of legislation and by the absence of standard formats, which limits in turn the diffusion and exchange of the information. All of this has repercussions in the effectiveness of a strategic planning and in the viability of the implementation of projects about the territory, and consequentially in the land development levels. The hypothesis is confirmed since all the described elements in PGISs and participatory mapping will act as valid tools in strengthening and improving the effectivity in planning only if they are interconnected and linked amongst themselves. This work, therefore, suggests a methodological proposal when faced with the conventional ways of planning: a new planning route which includes the resources exchange and local participatory determination before any plan is established -. With this, the benefits of participatory methodologies in the management of GI and GIS (Geographic Information Systems) is incorporated as a strategic instrument for development of local identity and optimization in planning processes and territory studies. Finally, the study outlines future work on PGIS maps and Participatory Mapping, such that these could eventually evolve into visual representative instruments that transfer identity values of the territory and its society. In this way, they would contribute to attain a better knowledge, recognition, and appraisement of the territories for their inhabitants and planners.
Resumo:
Globalization has intensified competition, as evidenced by the growing number of international classification systems (rankings) and the attention paid to them. Doctoral education has an international character in itself. It should promote opportunities for graduate students lo participate in these international studies. The quality and competitiveness are two of the most important issues for universities. To promote the interest of graduates to continue their education after the graduate level, it would be necessary to improve the published information of ihe doctoral programs. It should increase the visibility and provide high-quality, easily accessible and comparable information which includes all the relevant aspects of these programs. The authors analysed the website contents of doctoral programs, it was observed a lack of quality of them and very poor information about the contents, so that it was decided that any of them could constitute a model for creating new websites. The recommendations on the format and contents in the web were made by a discussion group. They recommended an attractive design; a page with easy access to contents and easy to find on Ihe net and with the information in more than one language. It should include complete program and academic staff information. It should also be included the study's results which should be easily accessible and includes quantitative data, such as number of students who completed scholars, publications, research projects, average duration of the studies, etc. It will facilitate the choice of program
Resumo:
Aeronautical charts underlie the representation of aeronautic geographic information that supports pilots in flight. Nevertheless, charts become complex due to the high density of data and the different kinds that support each phase of flight. These features make difficult using them on board. After conducting a study that aims to understand and to evaluate pilot’s needs related to Geographic Information, it is proposed a solution to implement a platform based on geographic information standards (OGC, ISO) and supported by a distributed Web architecture. This platform facilitates the use, retrieval, updating of information and its exchange among different institutions through private and public users. As a first element to ensure interoperability and the harmonisation of information, we propose an aeronautical metadata profile that sets guidelines and elements for its description. This profile meets the standards set by ICAO, Eurocontrol and ISO. The platform offers three levels of access to data through different types of devices and user profiles. This paper suggests an alternative and reliable way for distributing aeronautical geoinformation, focusing on specific functions or displaying and querying.
Resumo:
Sensor networks are increasingly becoming one of the main sources of Big Data on the Web. However, the observations that they produce are made available with heterogeneous schemas, vocabularies and data formats, making it difficult to share and reuse these data for other purposes than those for which they were originally set up. In this thesis we address these challenges, considering how we can transform streaming raw data to rich ontology-based information that is accessible through continuous queries for streaming data. Our main contribution is an ontology-based approach for providing data access and query capabilities to streaming data sources, allowing users to express their needs at a conceptual level, independent of implementation and language-specific details. We introduce novel query rewriting and data translation techniques that rely on mapping definitions relating streaming data models to ontological concepts. Specific contributions include: • The syntax and semantics of the SPARQLStream query language for ontologybased data access, and a query rewriting approach for transforming SPARQLStream queries into streaming algebra expressions. • The design of an ontology-based streaming data access engine that can internally reuse an existing data stream engine, complex event processor or sensor middleware, using R2RML mappings for defining relationships between streaming data models and ontology concepts. Concerning the sensor metadata of such streaming data sources, we have investigated how we can use raw measurements to characterize streaming data, producing enriched data descriptions in terms of ontological models. Our specific contributions are: • A representation of sensor data time series that captures gradient information that is useful to characterize types of sensor data. • A method for classifying sensor data time series and determining the type of data, using data mining techniques, and a method for extracting semantic sensor metadata features from the time series.
Resumo:
Presentación realizada en el PhD Seminar del ITS 2011 en Budapest. ICTs (Information and Communication Technologies) currently account for 2% of total carbon emissions. However, although modern standards require strict measures to reduce energy consumption across all industrial and services sectors, the ICT sector also faces an increase in services and bandwidth demand. The deployment of Next Generation Networks (NGN) will be the answer to this new demand; more specifically, Next Generation Access Networks (NGANs) will provide higher bandwidth access to users. Several policy and cost analyses are being carried out to understand the risks and opportunities of new deployments, but the question of what role energy consumption plays in NGANs seems off the table. Thus, this paper proposes a model to analyse the energy consumption of the main fibre-based NGAN architectures: Fibre To The House (FTTH), in both Passive Optical Network (PON) and Point-to-Point (PtP) variations, and FTTx/VDSL. The aim of this analysis is to provide deeper insight on the impact of new deployments on the energy consumption of the ICT sector and the effects of energy consumption on the life-cycle cost of NGANs. The paper also presents an energy consumption comparison of the presented architectures, particularised to the specific geographic and demographic distribution of users of Spain but easily extendable to other countries.
Resumo:
Nanotechnology represents an area of particular promise and significant opportunity across multiple scientific disciplines. Ongoing nanotechnology research ranges from the characterization of nanoparticles and nanomaterials to the analysis and processing of experimental data seeking correlations between nanoparticles and their functionalities and side effects. Due to their special properties, nanoparticles are suitable for cellular-level diagnostics and therapy, offering numerous applications in medicine, e.g. development of biomedical devices, tissue repair, drug delivery systems and biosensors. In nanomedicine, recent studies are producing large amounts of structural and property data, highlighting the role for computational approaches in information management. While in vitro and in vivo assays are expensive, the cost of computing is falling. Furthermore, improvements in the accuracy of computational methods (e.g. data mining, knowledge discovery, modeling and simulation) have enabled effective tools to automate the extraction, management and storage of these vast data volumes. Since this information is widely distributed, one major issue is how to locate and access data where it resides (which also poses data-sharing limitations). The novel discipline of nanoinformatics addresses the information challenges related to nanotechnology research. In this paper, we summarize the needs and challenges in the field and present an overview of extant initiatives and efforts.
Resumo:
The contribution to global energy consumption of the information and communications technology (ICT) sector has increased considerably in the last decade, along with its growing relevance to the overall economy. This trend will continue due to the seemingly ever greater use of these technologies, with broadband data traffic generated by the usage of telecommunication networks as a primary component. In fact, in response to user demand, the telecommunications industry is initiating the deployment of next generation networks (NGNs). However, energy consumption is mostly absent from the debate on these deployments, in spite of the potential impact on both expenses and sustainability. In addition, consumers are unaware of the energy impact of their choices in ultra-broadband services. This paper focuses on forecasting energy consumption in the access part of NGNs by modelling the combined effect of the deployment of two different ultra-broadband technologies (FTTH-GPON and LTE), the evolution of traffic per user, and the energy consumption in each of the networks and user devices. Conclusions are presented on the levels of energy consumption, their cost and the impact of different network design parameters. The effect of technological developments, techno-economic and policy decisions on energy consumption is highlighted. On the consumer side, practical figures and comparisons across technologies are provided. Although the paper focuses on Spain, the analysis can be extended to similar countries.
Resumo:
Secret-key agreement, a well-known problem in cryptography, allows two parties holding correlated sequences to agree on a secret key communicating over a public channel. It is usually divided into three different procedures: advantage distillation, information reconciliation and privacy amplification. The efficiency of each one of these procedures is needed if a positive key rate is to be attained from the legitimate parties? correlated sequences. Quantum key distribution (QKD) allows the two parties to obtain correlated sequences, provided that they have access to an authenticated channel. The new generation of QKD devices is able to work at higher speeds and in noisier or more absorbing environments. This exposes the weaknesses of current information reconciliation protocols, a key component to their performance. Here we present a new protocol based in low-density parity-check (LDPC) codes that presents the advantages of low interactivity, rate adaptability and high efficiency,characteristics that make it highly suitable for next generation QKD devices.