962 resultados para Web, Search Engine, Overlap
Resumo:
* This work was financially supported by RFBF-04-01-00858.
Resumo:
The paper presents results from the development of a methodology and corresponding software tools for building an academic repository. The repository was filled up with gaming material. The repository architecture and key features of the search engine are discussed. The emphasis falls on solutions of the large set of problems concerning the development of proper mechanisms for semantics-based search in a digital repository.
Resumo:
Purpose: A case study is presented concerning a gamified awards system designed to encourage software users to explore a suite of tools, and to share their expertise level in profile pages. Majestic is a high-tech business based in the West Midlands (UK) w hich offers a Link Intelligence database using a Software as a Service (SaaS) business model. Customers leverage the database for tasks including Search Engine Optimisation (SEO) by using a suite of web-based tools. Getting to know all the tools and how they can be deployed to good effect represents a considerable learning challenge, and Majestic were aware that. Design/methodology/approach: We present the development of Majestic Awards as a case study highlighting the most important design decisions. Then we reflect on the development process as an example of innovation adoption, thereby identifying resources and cu ltura l factors which were critical in ensuring the success of the project. Findings: The gamified awards system makes learning the tools an enjoyable, explorative experience. Success factors included identifying a clear business goal, the process/ project f it, senior management buy in, and identifying the knowledge and resources to resolve t echnical issues. Originality/value: Prior to gamification of the system, only the most expert users regu larly utilized all the tools. The user base is now more knowl edgable about the system and some users choose to use the system to publicize their expertise.
Resumo:
The design of interfaces to facilitate user search has become critical for search engines, ecommercesites, and intranets. This study investigated the use of targeted instructional hints to improve search by measuring the quantitative effects of users' performance and satisfaction. The effects of syntactic, semantic and exemplar search hints on user behavior were evaluated in an empirical investigation using naturalistic scenarios. Combining the three search hint components, each with two levels of intensity, in a factorial design generated eight search engine interfaces. Eighty participants participated in the study and each completed six realistic search tasks. Results revealed that the inclusion of search hints improved user effectiveness, efficiency and confidence when using the search interfaces, but with complex interactions that require specific guidelines for search interface designers. These design guidelines will allow search designers to create more effective interfaces for a variety of searchapplications.
Resumo:
With the dramatic growth of text information, there is an increasing need for powerful text mining systems that can automatically discover useful knowledge from text. Text is generally associated with all kinds of contextual information. Those contexts can be explicit, such as the time and the location where a blog article is written, and the author(s) of a biomedical publication, or implicit, such as the positive or negative sentiment that an author had when she wrote a product review; there may also be complex context such as the social network of the authors. Many applications require analysis of topic patterns over different contexts. For instance, analysis of search logs in the context of the user can reveal how we can improve the quality of a search engine by optimizing the search results according to particular users; analysis of customer reviews in the context of positive and negative sentiments can help the user summarize public opinions about a product; analysis of blogs or scientific publications in the context of a social network can facilitate discovery of more meaningful topical communities. Since context information significantly affects the choices of topics and language made by authors, in general, it is very important to incorporate it into analyzing and mining text data. In general, modeling the context in text, discovering contextual patterns of language units and topics from text, a general task which we refer to as Contextual Text Mining, has widespread applications in text mining. In this thesis, we provide a novel and systematic study of contextual text mining, which is a new paradigm of text mining treating context information as the ``first-class citizen.'' We formally define the problem of contextual text mining and its basic tasks, and propose a general framework for contextual text mining based on generative modeling of text. This conceptual framework provides general guidance on text mining problems with context information and can be instantiated into many real tasks, including the general problem of contextual topic analysis. We formally present a functional framework for contextual topic analysis, with a general contextual topic model and its various versions, which can effectively solve the text mining problems in a lot of real world applications. We further introduce general components of contextual topic analysis, by adding priors to contextual topic models to incorporate prior knowledge, regularizing contextual topic models with dependency structure of context, and postprocessing contextual patterns to extract refined patterns. The refinements on the general contextual topic model naturally lead to a variety of probabilistic models which incorporate different types of context and various assumptions and constraints. These special versions of the contextual topic model are proved effective in a variety of real applications involving topics and explicit contexts, implicit contexts, and complex contexts. We then introduce a postprocessing procedure for contextual patterns, by generating meaningful labels for multinomial context models. This method provides a general way to interpret text mining results for real users. By applying contextual text mining in the ``context'' of other text information management tasks, including ad hoc text retrieval and web search, we further prove the effectiveness of contextual text mining techniques in a quantitative way with large scale datasets. The framework of contextual text mining not only unifies many explorations of text analysis with context information, but also opens up many new possibilities for future research directions in text mining.
Resumo:
International audience
Resumo:
Repeat photography is an efficient, effective and useful method to identify trends of changes in the landscapes. It was used to illustrate long-term changes occurring in the landscapes. In the Northeast of Portugal, landscapes changes is currently driven mostly by agriculture abandonment and agriculture and energy policy. However, there is a need to monitoring changes in the region using a multitemporal and multiscale approach. This project aimed to establish an online repository of oblique digital photography from the region to be used to register the condition of the landscape as recorded in historical and contemporary photography over time as well as to support qualitative and quantitative assessment of change in the landscape using repeat photography techniques and methods. It involved the development of a relational database and a series of web-based services using PHP: Hypertext Preprocessor language, and the development of an interface, with Joomla, of pictures uploading and downloading by users. The repository will make possible to upload, store, search by location, theme, or date, display, and download pictures for Northeastern Portugal. The website service is devoted to help researchers to obtain quickly the photographs needed to apply RP through a developed search engine. It can be accessed at: http://esa.ipb.pt/digitalandscape/.
Resumo:
Los trastornos del comportamiento alimentario (TCA) son las patologías psicológicas que más se han incrementado en los últimos años. Uno de los factores que determina la elevada prevalencia de TCA en nuestra sociedad es el gran desconocimiento sobre alimentación. Este desconocimiento puede deberse a la consulta de recursos online sin validez científica. El objetivo de este trabajo ha sido analizar la calidad científica y el posicionamiento de los sitios web en español con información sobre nutrición, TCA y obesidad. Material y métodos: Se realizó una búsqueda de páginas web en el navegador Google Chrome con las palabras clave: dieta, anorexia, bulimia, nutrición y obesidad, seleccionándose los 20 primeros resultados de cada búsqueda según los índices de posicionamiento ofrecidos por SEOquake (Page Rank, Alexa Rank y SEMrush Rank). Las variables de análisis fueron: información relacionada con dietas y hábitos alimentarios, información sobre alimentación saludable, información sobre TCA y sus criterios diagnósticos e información de carácter formativo acerca de temas profesionales de salud general. Sólo el 50% de las web encontradas cumplían los criterios de inclusión en el estudio. La mayoría no seguían las pautas establecidas por e-Europa sobre calidad. La mediana de Page Rank fue de 2, excepto en aquellas asociadas a instituciones sanitarias de prestigio. Dada la escasez de webs sanitarias con rigor científico, es imprescindible la revisión de las existentes y la creación de nuevos espacios on-line cuya supervisión sea realizada por profesionales especialistas en salud y nutrición.
Resumo:
Die Arbeit geht dem Status quo der unternehmensweiten Suche in österreichischen Großunternehmen nach und beleuchtet Faktoren, die darauf Einfluss haben. Aus der Analyse des Ist-Zustands wird der Bedarf an Enterprise-Search-Software abgeleitet und es werden Rahmenbedingungen für deren erfolgreiche Einführung skizziert. Die Untersuchung stützt sich auf eine im Jahr 2009 durchgeführte Onlinebefragung von 469 österreichischen Großunternehmen (Rücklauf 22 %) und daran anschließende Leitfadeninterviews mit zwölf Teilnehmern der Onlinebefragung. Der theoretische Teil verortet die Arbeit im Kontext des Informations- und Wissensmanagements. Der Fokus liegt auf dem Ansatz der Enterprise Search, ihrer Abgrenzung gegenüber der Suche im Internet und ihrem Leistungsspektrum. Im empirischen Teil wird zunächst aufgezeigt, wie die Unternehmen ihre Informationen organisieren und welche Probleme dabei auftreten. Es folgt eine Analyse des Status quo der Informationssuche im Unternehmen. Abschließend werden Bekanntheit und Einsatz von Enterprise-Search-Software in der Zielgruppe untersucht sowie für die Einführung dieser Software nötige Rahmenbedingungen benannt. Defizite machen die Befragten insbesondere im Hinblick auf die übergreifende Suche im Unternehmen und die Suche nach Kompetenzträgern aus. Hier werden Lücken im Wissensmanagement offenbar. 29 % der Respondenten der Onlinebefragung geben zudem an, dass es in ihren Unternehmen gelegentlich bis häufig zu Fehlentscheidungen infolge defizitärer Informationslagen kommt. Enterprise-Search-Software kommt in 17 % der Unternehmen, die sich an der Onlinebefragung beteiligten, zum Einsatz. Die durch Enterprise-Search-Software bewirkten Veränderungen werden grundsätzlich positiv beurteilt. Alles in allem zeigen die Ergebnisse, dass Enterprise-Search-Strategien nur Erfolg haben können, wenn man sie in umfassende Maßnahmen des Informations- und Wissensmanagements einbettet.
Resumo:
Most of the existing open-source search engines, utilize keyword or tf-idf based techniques to find relevant documents and web pages relative to an input query. Although these methods, with the help of a page rank or knowledge graphs, proved to be effective in some cases, they often fail to retrieve relevant instances for more complicated queries that would require a semantic understanding to be exploited. In this Thesis, a self-supervised information retrieval system based on transformers is employed to build a semantic search engine over the library of Gruppo Maggioli company. Semantic search or search with meaning can refer to an understanding of the query, instead of simply finding words matches and, in general, it represents knowledge in a way suitable for retrieval. We chose to investigate a new self-supervised strategy to handle the training of unlabeled data based on the creation of pairs of ’artificial’ queries and the respective positive passages. We claim that by removing the reliance on labeled data, we may use the large volume of unlabeled material on the web without being limited to languages or domains where labeled data is abundant.
Resumo:
This thesis describes a project of terminology and localization focused on local and traditional cuisine from the province of Modena: the final products of this project are a specialized termbase and the localized version of the website of Trattoria Ermes, a small Modenese restaurant offering traditional dishes. It is a known fact the Internet has drastically altered the way companies and businesses communicate with their audience. Considering that food tourism is an invaluable sector of Italy’s economy and a great aid to safeguarding its culinary traditions, business owners can benefit from localizing their websites, allowing them to reach wider international audiences. The project is divided into two main sections: the first focuses on the terminological systematization of specialized terminology collected from Sandro Bellei’s cooking book and two web-derived monolingual corpora, while the second section offers insight into the analysis of the localization and optimization process of Trattoria Ermes website. In particular, the thesis approaches localization from the point of view of web marketing, with a theoretical and practical section dedicated to the Search Engine Optimization (SEO) processes employed by web marketing teams to ensure the visibility and popularity of the website
Resumo:
La ricerca di documenti rilevanti è un task fondamentale, può avvenire tramite ambienti chiusi, come le biblioteche digitali o tramite ambienti aperti, come il World Wide Web. Quello che analizzeremo in questo progetto di tesi riguarderà le interfacce per mostrare i risultati di ricerca su una collezione di documenti. L'obiettivo, tuttavia, non è l'analisi dei motori di ricerca, ma analizzare i diversi meccanismi che permettono di visualizzare i risultati. Vedremo, inoltre, le diverse visualizzazioni rilevanti nella ricerca di informazioni sul web, in particolare parleremo di visualizzazioni nello spazio, messe a confronto con la classica visualizzazione testuale. Analizzeremo anche la classificazione in una collezione di documenti, oltre che la personalizzazione, ovvero la configurazione della visualizzazione a vantaggio dell'utente. Una volta trovati i documenti rilevanti, analizzeremo i frammenti di testo come, gli snippet, i riassunti descrittivi e gli abstract, mettendo in luce il modo in cui essi aiutato l'utente a migliorare l'accesso a determinati tipi di risultati. Infine, andremo ad analizzare le visualizzazioni di frammenti rilevanti all'interno di un testo. In particolare presenteremo le tecniche di navigazione e la ricerca di determinate parole all'interno di documenti, vale a dire le panoramiche di documenti e del metodo preferito dall'utente per cercare una parola.
Resumo:
Chemical cross-linking has emerged as a powerful approach for the structural characterization of proteins and protein complexes. However, the correct identification of covalently linked (cross-linked or XL) peptides analyzed by tandem mass spectrometry is still an open challenge. Here we present SIM-XL, a software tool that can analyze data generated through commonly used cross-linkers (e.g., BS3/DSS). Our software introduces a new paradigm for search-space reduction, which ultimately accounts for its increase in speed and sensitivity. Moreover, our search engine is the first to capitalize on reporter ions for selecting tandem mass spectra derived from cross-linked peptides. It also makes available a 2D interaction map and a spectrum-annotation tool unmatched by any of its kind. We show SIM-XL to be more sensitive and faster than a competing tool when analyzing a data set obtained from the human HSP90. The software is freely available for academic use at http://patternlabforproteomics.org/sim-xl. A video demonstrating the tool is available at http://patternlabforproteomics.org/sim-xl/video. SIM-XL is the first tool to support XL data in the mzIdentML format; all data are thus available from the ProteomeXchange consortium (identifier PXD001677).
Resumo:
The study reported here is a classical bottom-up proteomic approach where proteins from wasp venom were extracted and separated by 2-DE; the individual protein spots were proteolytically digested and subsequently identified by using tandem mass spectrometry and database query with the protein search engine MASCOT. Eighty-four venom proteins belonging to 12 different molecular functions were identified. These proteins were classified into three groups; the first is constituted of typical venom proteins: antigens-5, hyaluronidases, phospholipases, heat shock proteins, metalloproteinases, metalloproteinase-desintegrin like proteins, serine proteinases, proteinase inhibitors, vascular endothelial growth factor-related protein, arginine kinases, Sol i-II and -II like proteins, alpha-glucosidase, and superoxide dismutases. The second contained proteins structurally related to the muscles that involves the venom reservoir. The third group, associated with the housekeeping of cells from venom glands, was composed of enzymes, membrane proteins of different types, and transcriptional factors. The composition of P. paulista venom permits us to hypothesize about a general envenoming mechanism based on five actions: (i) diffusion of venom through the tissues and to the blood, (ii) tissue, (iii) hemolysis, (iv) inflammation, and (v) allergy-played by antigen-5, PLA1, hyaluronidase, HSP 60, HSP 90, and arginine kinases.
Resumo:
OBJECTIVE To review studies on the readability of package leaflets of medicinal products for human use.METHODS We conducted a systematic literature review between 2008 and 2013 using the keywords “Readability and Package Leaflet” and “Readability and Package Insert” in the academic search engine Biblioteca do Conhecimento Online,comprising different bibliographic resources/databases. The preferred reporting items for systematic reviews and meta-analyses criteria were applied to prepare the draft of the report. Quantitative and qualitative original studies were included. Opinion or review studies not written in English, Portuguese, Italian, French, or Spanish were excluded.RESULTS We identified 202 studies, of which 180 were excluded and 22 were enrolled [two enrolling healthcare professionals, 10 enrolling other type of participants (including patients), three focused on adverse reactions, and 7 descriptive studies]. The package leaflets presented various readability problems, such as complex and difficult to understand texts, small font size, or few illustrations. The main methods to assess the readability of the package leaflet were usability tests or legibility formulae. Limitations with these methods included reduced number of participants; lack of readability formulas specifically validated for specific languages (e.g., Portuguese); and absence of an assessment on patients literacy, health knowledge, cognitive skills, levels of satisfaction, and opinions.CONCLUSIONS Overall, the package leaflets presented various readability problems. In this review, some methodological limitations were identified, including the participation of a limited number of patients and healthcare professionals, the absence of prior assessments of participant literacy, humor or sense of satisfaction, or the predominance of studies not based on role-plays about the use of medicines. These limitations should be avoided in future studies and be considered when interpreting the results.