590 resultados para Automatized Indexing
Resumo:
Welding is one of the most employed process for joining steel pipes. Although, manual welding is still the most used one, mechanized version and even automatized one have increased its demand. Thus, this work deals with girth welding of API 5L X65 pipes with 8” of nominal diameter and 8.0 mm thickness, beveled with V-30º narrow gap. Torch is moved by a bug carrier (mechanized welding) and further the parameters are controlled as a function of angular position (automatized welding). Welding parameters are presented for filling the joint with two-passes (root and filling/capping passes). Parameters for the root pass were extracted from previous author´s work with weldments carried out in plates, but validated in this work for pipe welding. GMAW processes were assessed with short-circuit metal transfer in both conventional and derivative modes using different technologies (RMD, STT and CMT). After the parameter determination, mechanical testing was performed for welding qualification (uniaxial tension, face and root bending, nick break, Charpy V-notch impact, microhardness and macrograph). The initially obtained results for RMD and CMT were acceptable for all testing and, in a second moment, also for the STT. However, weld beads carried out by using the conventional process failed and revealed the existence of lack of fusion, which required further parametrization. Thus, a Parameter-Variation System for Girth Welding (SVP) was designed and built to allow varying the welding parameters as a function of angular position by using an inclinometer. The parameters were set for each of the three angular positions (flat, vertical downhill and overhead). By using such equipment and approach, the conventional process with parameter variation allowed reducing the welding time for joint accomplishment of the order of 38% for the root pass and 30% for the filling/capping pass.
Resumo:
Postprint
Representing clinical documents to support automatic retrieval of evidence from the Cochrane Library
Resumo:
The overall aim of our research is to develop a clinical information retrieval system that retrieves systematic reviews and underlying clinical studies from the Cochrane Library to support physician decision making. We believe that in order to accomplish this goal we need to develop a mechanism for effectively representing documents that will be retrieved by the application. Therefore, as a first step in developing the retrieval application we have developed a methodology that semi-automatically generates high quality indices and applies them as descriptors to documents from The Cochrane Library. In this paper we present a description and implementation of the automatic indexing methodology and an evaluation that demonstrates that enhanced document representation results in the retrieval of relevant documents for clinical queries. We argue that the evaluation of information retrieval applications should also include an evaluation of the quality of the representation of documents that may be retrieved. ©2010 IEEE.
Resumo:
Scientific reading research has produced substantial evidence linking specific reading components to a range of constructs including phonological awareness (PA), morphological awareness, orthographic processing (OP), rapid automatized naming, working memory and vocabulary. There is a paucity of research on Arabic, although 420 million people around the world (Gordon, 2005) speak Arabic. As a Semitic language, Arabic differs in many ways from Indo-European languages. Over the past three decades, literacy research has begun to elucidate the importance of morphological awareness (MA) in reading. Morphology is a salient aspect of Arabic word structure. This study was designed to (a) examine the dimensions underlying MA in Arabic; (b) determine how well MA predicts reading; (c) investigate the role of the standard predictors in different reading outcomes; and (d) investigate the construct of reading in Arabic. This study was undertaken in two phases. In Phase I, 10 MA measures and two reading measures were developed, and tested in a sample of 102 Grade 3 Arabic-speaking children. Factor analysis of the 10 MA tasks yielded one predominant factor supporting the construct validity of MA in Arabic. Hierarchical regression analyses, controlling for age and gender, indicated that the MA factor solution accounted for 41– 43% of the variance in reading. In Phase II, the widely studied predictor measures were developed for PA and OP in addition to one additional measure of MA (root awareness), and three reading measures In Phase II, all measures were administered to another sample of 201 Grade 3 Arabic-speaking children. The construct of reading in Arabic was examined using factor analysis. The joint and unique effects of all standard predictors were examined using different sets of hierarchical regression analyses. Results of Phase II showed that: (a) all five reading measures loaded on one factor; (b) MA consistently accounted for unique variance in reading, particularly in comprehension, above and beyond the standard predictors; and (c) the standard predictors had differential contributions. These findings underscore the contribution of MA to all components of Arabic reading. The need for more emphasis on including morphology in Arabic reading instruction and assessment is discussed.
Resumo:
This research includes a review of the log management of the company Telia. The research has also included a comparison of the two log management sys- tems Splunk and ELK. The review of the company’s log management shows that log messages are being stored in files on a hard drive that can be accessed through the network. The log messages are system-specific. ELK is able to fetch log messages of different formats simultaneously, but this feature is not possible in Splunk where the process of uploading log messages has to be re- peated for log messages that have different formats. Both systems store log messages through a file system on a hard drive, where the systems are installed. In networks that involve multiple servers, ELK is distributing the log messages between the servers. Thus, the workload to perform searches and storing large amounts of data is reduced. Using Splunk in networks can also reduce the workload. This is done by using forwarders that send the log messages to one or multiple central servers which stores the messages. Searches of log messages in Splunk are performed by using a graphical interface. Searches in ELK is done by using a REST-API which can be used by external systems as well, to retrieve search results. Splunk also has a REST-API that can be used by external sys- tems to receive search results. The research revealed that ELK had a lower search time than Splunk. However, no method was found that could be used to measure the indexing time of ELK, which meant that no comparison could be made with respect to the indexing time for Splunk. For future work there should be an investigation whether there is any possibility to measure the indexing time of ELK. Another recommendation is to include more log management sys- tem in the research to improve the results that may be suitable candidates for the company Telia. An improvement suggestion as well, is to do performance tests in a network with multiple servers and thereby draw conclusions how the performance is in practice.
Resumo:
Journal impact factors have become an important criterion to judge the quality of scientific publications over the years, influencing the evaluation of institutions and individual researchers worldwide. However, they are also subject to a number of criticisms. Here we point out that the calculation of a journal’s impact factor is mainly based on the date of publication of its articles in print form, despite the fact that most journals now make their articles available online before that date. We analyze 61 neuroscience journals and show that delays between online and print publication of articles increased steadily over the last decade. Importantly, such a practice varies widely among journals, as some of them have no delays, while for others this period is longer than a year. Using a modified impact factor based on online rather than print publication dates, we demonstrate that online-to-print delays can artificially raise a journal’s impact factor, and that this inflation is greater for longer publication lags. We also show that correcting the effect of publication delay on impact factors changes journal rankings based on this metric. We thus suggest that indexing of articles in citation databases and calculation of citation metrics should be based on the date of an article’s online appearance, rather than on that of its publication in print.
Resumo:
La présente thèse examine les associations entre les dimensions du TDAH et les habiletés en lecture sur les plans phénotypique, génétique et cognitif. En premier lieu, les associations entre les dimensions du TDAH (inattention et hyperactivité/impulsivité) et les habiletés en lecture (décodage et compréhension en lecture) chez des enfants au début du primaire (6-8 ans) ont été examinées. Les résultats révèlent des associations similaires. Toutefois, seules celles entre l’inattention et les habiletés en lecture demeurent après que l’hyperactivité/impulsivité, les symptômes de trouble du comportement et les habiletés non verbales aient été contrôlés. De plus, les associations entre l’inattention et les habiletés en lecture s’expliquent en grande partie par des facteurs génétiques. En second lieu, les associations entre les dimensions du TDAH et les habiletés en lecture (lecture de mots et exactitude/vitesse lors de la lecture d’un texte) ont été étudiées à 14-15 ans. Seule l’inattention demeure associée aux habiletés en lecture après que l’hyperactivité/impulsivité, les habiletés verbales et les habiletés non verbales aient été contrôlées. L’inattention et les habiletés en lecture sont aussi corrélées sur le plan génétique, mais ces corrélations deviennent non significatives lorsque les habiletés verbales sont contrôlées. En dernier lieu, des habiletés cognitives ont été étudiées comme mécanismes sous-jacents potentiels de l’association entre l’inattention et les habiletés en lecture (décodage et compréhension en lecture) à l’enfance. Il apparait que la conscience phonologique, la vitesse de dénomination de chiffres, le traitement temporel bimodal et le vocabulaire sont des médiateurs de l’association entre l’inattention et le décodage alors que la conscience phonologique, la vitesse de dénomination de chiffres et de couleurs et le vocabulaire sont des médiateurs de l’association entre l’inattention et la compréhension en lecture. De plus, des facteurs génétiques communs ont été observés entre certains médiateurs (conscience phonologique, vitesse de dénomination des chiffres et traitement temporel bimodal), l’inattention et le décodage. Somme toute, la présente thèse montre que des facteurs génétiques expliquent en partie ces associations à l’enfance et l’adolescence. Des médiateurs cognitifs sous-tendent ces associations, possiblement par des processus génétiques et environnementaux qui devront être précisés dans le futur.
Resumo:
The Iowa Department of Transportation began preparation for the acquisition of an electronic document management system in 1996. The first phase was development of a strategic plan. The plan provided guidelines for defining the acquisition and implementation of a document management system to automate document handling and distribution. Phase 2 involved developing draft standards (document, indexing and technology) for planning and implementation of a document management system. These standards were to identify existing industry standards and determine which standards would best support the specific requirements of the Iowa Department of Transportation. During development of these standards, the decision was made to enlarge the scope of this effort from a document management system to a records management system (RMS). Phase .3 identified business processes that were to be further developed as pilot projects of a much larger agency-wide records management system.
Resumo:
Each two years the amount of available information in the world double. This is the Information Age, where the success depends on what one knows, not on what one has. A new economy appears with the capacity to generate, to store, to process and to apply effectively the knowledge, based on information, determining the companies productivity and competitiveness.The objective of this work is to understand the information management model of a technological research institute - CTGÁS (Gas Technology Center). The research has been done focused on the 5 main processes and the 15 support processes of the organization value chain , aiming to understand the information management in the organization based on Davenport´s Information Management model (1998). Therefore, it was necessary to identify how the necessary information for the organizational processes accomplishment are determined, obtained, distributed and used by the organization. The research can be classified as descriptive, regarding to its aims, and as a case study, related to the research ways. Interviews with the managers of the organization value chain processes have been carried through, with the objective to identify how they perceive the Information Management process that circulates in the organizational processes. Complementarily, a documentary research has been carried through, associated to the direct observation and procedures and actions follow up, involving the Information Management. The data treatment and analysis have been done from the authors theoretical support and from the managers interviews analysis, documents and processes observed by the researcher in the organization. It was noticed that the organization has raised its level of information needs that are not difficult to be determined and are satisfactorily obtained and distributed, although the majority of them are not structuralized, automatized or even classified regarding to its confidence. These peaces of information have good quality and are important, however they reflect a medium dependence on external and informal information, besides being used only in its great majority for people to know what and how to do something
Resumo:
Multimedia objects, especially images and figures, are essential for the visualization and interpretation of research findings. The distribution and reuse of these scientific objects is significantly improved under open access conditions, for instance in Wikipedia articles, in research literature, as well as in education and knowledge dissemination, where licensing of images often represents a serious barrier. Whereas scientific publications are retrievable through library portals or other online search services due to standardized indices there is no targeted retrieval and access to the accompanying images and figures yet. Consequently there is a great demand to develop standardized indexing methods for these multimedia open access objects in order to improve the accessibility to this material. With our proposal, we hope to serve a broad audience which looks up a scientific or technical term in a web search portal first. Until now, this audience has little chance to find an openly accessible and reusable image narrowly matching their search term on first try - frustratingly so, even if there is in fact such an image included in some open access article.
Resumo:
Estudo de caso único exploratório e descritivo voltado a analisar a indexação em uma das bibliotecas universitárias do SIB/FURG. Os objetivos específicos, delimitados a partir do contexto já citado, foram: a) Identificar e analisar, através de mapeamento cognitivo, os procedimentos metodológicos empregados na indexação nas atividades de análise, síntese e representação da informação; b) Identificar os conceitos/noções com maior importância na percepção da indexadora quanto ao processo de indexação e as relações entre tais conceitos de forma a construir o mapa cognitivo do processo a partir da percepção da indexadora; e c) Descrever e analisar a indexação de livros na unidade em estudo sob aspecto da análise, síntese e representação destes através da aplicação do Protocolo Verbal. As técnicas utilizadas para a coleta de informação no estudo de caso único foram a Self-Q e o Protocolo Verbal, ambas centradas na abordagem qualitativa. Conclui-se, a partir da construção do mapa cognitivo da indexadora, que as noções/conceitos que sustentam sua prática voltam-se, em sua maioria, a aspectos de caráter procedimental. Percebeu-se também que a prática de indexação ocorre desconectada dos princípios de especificidade e exaustividade. Sobre a indexação de livros conclui-se que, na unidade sob estudo, as operações de análise são desenvolvidas de modo empírico através da leitura e interpretação de partes do documento indexado. Identificou-se que o enfoque da prática não recai apenas no documento mas também, no usuário. A análise e síntese ocorrem de forma integrada, sendo que, em alguns momentos, a síntese é desenvolvida a partir do conhecimento dos descritores do tesauro. A delimitação dos conceitos, por sua vez, foi influenciada, por vezes, pelo(a): uso de termos já empregados na unidade em que atua/sistema, presença do descritor no sumário, conhecimento das demandas dos usuários, área de domínio em que indexa e percepção enquanto profissional. Percebeu-se que não existem níveis definidos quanto a exaustividade e especificidade na indexação. Na representação dos conceitos foram identificadas dificuldades ocasionadas pela ausência de relacionamentos entre termos e/ou ausência de termos voltados a área indexada no tesauro empregado. Conclui-se que faz-se necessário o desenvolvimento de uma política de indexação formalizada para basilar a prática desenvolvida no SIB/FURG.
Resumo:
A partir da web 2.0, com a maior participação do usuário, destaca-se a mudança nas mídias tradicionais, como rádio e televisão, que ganharam caráter social, permitindo a participação do público com algo mais além de só ouvir e ver, mas contribuir das mais diversas maneiras, dente elas está a folksonomia. Este trabalho teve como objetivo analisar as tags no site Last.fm por meio do modelo proposto por Sen et al. (2006) que as identifica em três categorias: fatuais, subjetivas e pessoais. Acredita-se que este estudo é importante para compreender como se dá a representação da informação em linguagem livre pelo usuário, assim como considera-se relevante a folksonomia como prática social, democrática e inclusiva que aproxima o usuário da informação. Como hipótese, considerou-se o predomínio de tags subjetivas. A pesquisa teve caráter descritivo-exploratória, e uma abordagem qualiquantitativa. A partir de uma pesquisa do IBOPE (2013), foram escolhidas 12 artistas e 12 músicas como amostra da pesquisa. Analisou-se 1109 tags, com os resultados: 759 (68%) fatuais; 232 (21%) subjetivas; 85 (8%) pessoais e 33 (3%) com tipologia não-identificada, desta maneira a hipótese de predomínio de tags subjetivas foi refutada. Como um desdobramento do modelo de Sen et al. (2006), distribuiu-se as tags em categorias descritivas, adaptadas do modelo de Laplante (2015), que mostraram que os gêneros/estilos musicais são predominantes da descrição dos itens. Conclui-se que mesmo com o predomínio de tags fatuais e de gêneros/estilos musicais, nem sempre a representação contida nessas etiquetas é fidedigna à realidade descritiva do item, contudo a folksonomia tem a contribuir com novos pontos de vista e de pensar uma informação.
Resumo:
Journal impact factors have become an important criterion to judge the quality of scientific publications over the years, influencing the evaluation of institutions and individual researchers worldwide. However, they are also subject to a number of criticisms. Here we point out that the calculation of a journal’s impact factor is mainly based on the date of publication of its articles in print form, despite the fact that most journals now make their articles available online before that date. We analyze 61 neuroscience journals and show that delays between online and print publication of articles increased steadily over the last decade. Importantly, such a practice varies widely among journals, as some of them have no delays, while for others this period is longer than a year. Using a modified impact factor based on online rather than print publication dates, we demonstrate that online-to-print delays can artificially raise a journal’s impact factor, and that this inflation is greater for longer publication lags. We also show that correcting the effect of publication delay on impact factors changes journal rankings based on this metric. We thus suggest that indexing of articles in citation databases and calculation of citation metrics should be based on the date of an article’s online appearance, rather than on that of its publication in print.
Resumo:
Edge-labeled graphs have proliferated rapidly over the last decade due to the increased popularity of social networks and the Semantic Web. In social networks, relationships between people are represented by edges and each edge is labeled with a semantic annotation. Hence, a huge single graph can express many different relationships between entities. The Semantic Web represents each single fragment of knowledge as a triple (subject, predicate, object), which is conceptually identical to an edge from subject to object labeled with predicates. A set of triples constitutes an edge-labeled graph on which knowledge inference is performed. Subgraph matching has been extensively used as a query language for patterns in the context of edge-labeled graphs. For example, in social networks, users can specify a subgraph matching query to find all people that have certain neighborhood relationships. Heavily used fragments of the SPARQL query language for the Semantic Web and graph queries of other graph DBMS can also be viewed as subgraph matching over large graphs. Though subgraph matching has been extensively studied as a query paradigm in the Semantic Web and in social networks, a user can get a large number of answers in response to a query. These answers can be shown to the user in accordance with an importance ranking. In this thesis proposal, we present four different scoring models along with scalable algorithms to find the top-k answers via a suite of intelligent pruning techniques. The suggested models consist of a practically important subset of the SPARQL query language augmented with some additional useful features. The first model called Substitution Importance Query (SIQ) identifies the top-k answers whose scores are calculated from matched vertices' properties in each answer in accordance with a user-specified notion of importance. The second model called Vertex Importance Query (VIQ) identifies important vertices in accordance with a user-defined scoring method that builds on top of various subgraphs articulated by the user. Approximate Importance Query (AIQ), our third model, allows partial and inexact matchings and returns top-k of them with a user-specified approximation terms and scoring functions. In the fourth model called Probabilistic Importance Query (PIQ), a query consists of several sub-blocks: one mandatory block that must be mapped and other blocks that can be opportunistically mapped. The probability is calculated from various aspects of answers such as the number of mapped blocks, vertices' properties in each block and so on and the most top-k probable answers are returned. An important distinguishing feature of our work is that we allow the user a huge amount of freedom in specifying: (i) what pattern and approximation he considers important, (ii) how to score answers - irrespective of whether they are vertices or substitution, and (iii) how to combine and aggregate scores generated by multiple patterns and/or multiple substitutions. Because so much power is given to the user, indexing is more challenging than in situations where additional restrictions are imposed on the queries the user can ask. The proposed algorithms for the first model can also be used for answering SPARQL queries with ORDER BY and LIMIT, and the method for the second model also works for SPARQL queries with GROUP BY, ORDER BY and LIMIT. We test our algorithms on multiple real-world graph databases, showing that our algorithms are far more efficient than popular triple stores.
Resumo:
Trabalho Final de Mestrado para obtenção do Grau de Mestre em Engenharia de Redes de Comunicação e Multimédia