973 resultados para Unstructured content search
Resumo:
A process has been elaborated for one-step low lignin content sugarcane bagasse hemicellulose extraction using alkaline solution of hydrogen peroxide. To maximize the hemicellulose yields several extraction conditions were examined applying the 2(4) factorial design: H(2)O(2) concentration from 2 to 6% (w/v), reaction time from 4 to 16 h, temperature from 20 to 60 degrees C, and magnesium sulfate absence or presence (0.5%, w/v). This approach allowed selection of conditions for the extraction of low and high lignin content hemicellulose. At midpoint the yield of hemicellulose was 94.5% with more than 88% of lignin removed. Lignin removal is suppressed at low extraction temperatures and in the absence of magnesium sulfate. Hemicellulose in 86% yield with low lignin content (5.9%) was obtained with 6% H(2)O(2) treatment for 4 h and 20 degrees C. This hemicellulose is much lighter in color than samples obtained at the midpoint condition and was found suitable for subsequent enzymatic hydrolysis. (C) 2009 Elsevier B.V. All rights reserved.
Resumo:
This paper discusses a document discovery tool based on Conceptual Clustering by Formal Concept Analysis. The program allows users to navigate e-mail using a visual lattice metaphor rather than a tree. It implements a virtual. le structure over e-mail where files and entire directories can appear in multiple positions. The content and shape of the lattice formed by the conceptual ontology can assist in e-mail discovery. The system described provides more flexibility in retrieving stored e-mails than what is normally available in e-mail clients. The paper discusses how conceptual ontologies can leverage traditional document retrieval systems and aid knowledge discovery in document collections.
Resumo:
Purpose: To describe and compare the content of instruments that assess environmental factors using the International Classification of Functioning, Disability and Health (ICF). Methods: A systematic search of PubMed, CINAHL and PEDro databases was conducted using a pre-determined search strategy. The identified instruments were screened independently by two investigators, and meaningful concepts were linked to the most precise ICF category according to published linking rules. Results: Six instruments were included, containing 526 meaningful concepts. Instruments had between 20% and 98% of items linked to categories in Chapter 1. The highest percentage of items from one instrument linked to categories in Chapters 2–5 varied between 9% and 50%. The presence or absence of environmental factors in a specific context is assessed in 3 instruments, while the other 3 assess the intensity of the impact of environmental factors. Discussion: Instruments differ in their content, type of assessment, and have several items linked to the same ICF category. Most instruments primarily assess products and technology (Chapter 1), highlighting the need to deepen the discussion on the theory that supports the measurement of environmental factors. This discussion should be thorough and lead to the development of methodologies and new tools that capture the underlying concepts of the ICF.
Resumo:
This papers aims at providing a combined strategy for solving systems of equalities and inequalities. The combined strategy uses two types of steps: a global search step and a local search step. The global step relies on a tabu search heuristic and the local step uses a deterministic search known as Hooke and Jeeves. The choice of step, at each iteration, is based on the level of reduction of the l2-norm of the error function observed in the equivalent system of equations, compared with the previous iteration.
Resumo:
Dissertação para obtenção do Grau de Mestre em Engenharia Informática
Resumo:
This research intends to examine if there were significant differences on the brand engagement and on the electronic word of mouth (e-WOM)1 referral intention through Facebook between Generation X and Generation Y (also called millennials). Also, this study intends to examine if there are differences in the motivations that drive these generations to interact with brands through Facebook. Results indicated that Generation Y members consumed more content on Facebook brands’ pages than Generation X. Also, they were more likely to have an e-WOM referral intention as well as being more driven by brand affiliation and opportunity seeking. Finally, currently employed individuals were found to contribute with more content than students. This study fills the gap in the literature by addressing how marketing professionals should market their brand and interact and engage with their customers, based on customers’ generational cohort.
Resumo:
A generic search for anomalous production of events with at least three charged leptons is presented. The data sample consists of pp collisions at s√=8 TeV collected in 2012 by the ATLAS experiment at the CERN Large Hadron Collider, and corresponds to an integrated luminosity of 20.3 fb−1. Events are required to have at least three selected lepton candidates, at least two of which must be electrons or muons, while the third may be a hadronically decaying tau. Selected events are categorized based on their lepton flavour content and signal regions are constructed using several kinematic variables of interest. No significant deviations from Standard Model predictions are observed. Model-independent upper limits on contributions from beyond the Standard Model phenomena are provided for each signal region, along with prescription to re-interpret the limits for any model. Constraints are also placed on models predicting doubly charged Higgs bosons and excited leptons. For doubly charged Higgs bosons decaying to eτ or μτ, lower limits on the mass are set at 400 GeV at 95% confidence level. For excited leptons, constraints are provided as functions of both the mass of the excited state and the compositeness scale Λ, with the strongest mass constraints arising in regions where the mass equals Λ. In such scenarios, lower mass limits are set at 3.0 TeV for excited electrons and muons, 2.5 TeV for excited taus, and 1.6 TeV for every excited-neutrino flavour.
Resumo:
Internet is increasingly used as a source of information on health issues and is probably a major source of patients' empowerment. This process is however limited by the frequently poor quality of web-based health information designed for consumers. A better diffusion of information about criteria defining the quality of the content of websites, and about useful methods designed for searching such needed information, could be particularly useful to patients and their relatives. A brief, six-items DISCERN version, characterized by a high specificity for detecting websites with good or very good content quality was recently developed. This tool could facilitate the identification of high-quality information on the web by patients and may improve the empowerment process initiated by the development of the health-related web.
Resumo:
We have identified new malaria vaccine candidates through the combination of bioinformatics prediction of stable protein domains in the Plasmodium falciparum genome, chemical synthesis of polypeptides, in vitro biological functional assays, and association of an antigen-specific antibody response with protection against clinical malaria. Within the predicted open reading frame of P. falciparum hypothetical protein PFF0165c, several segments with low hydrophobic amino acid content, which are likely to be intrinsically unstructured, were identified. The synthetic peptide corresponding to one such segment (P27A) was well recognized by sera and peripheral blood mononuclear cells of adults living in different regions where malaria is endemic. High antibody titers were induced in different strains of mice and in rabbits immunized with the polypeptide formulated with different adjuvants. These antibodies recognized native epitopes in P. falciparum-infected erythrocytes, formed distinct bands in Western blots, and were inhibitory in an in vitro antibody-dependent cellular inhibition parasite-growth assay. The immunological properties of P27A, together with its low polymorphism and association with clinical protection from malaria in humans, warrant its further development as a malaria vaccine candidate.
Resumo:
Current-day web search engines (e.g., Google) do not crawl and index a significant portion of theWeb and, hence, web users relying on search engines only are unable to discover and access a large amount of information from the non-indexable part of the Web. Specifically, dynamic pages generated based on parameters provided by a user via web search forms (or search interfaces) are not indexed by search engines and cannot be found in searchers’ results. Such search interfaces provide web users with an online access to myriads of databases on the Web. In order to obtain some information from a web database of interest, a user issues his/her query by specifying query terms in a search form and receives the query results, a set of dynamic pages that embed required information from a database. At the same time, issuing a query via an arbitrary search interface is an extremely complex task for any kind of automatic agents including web crawlers, which, at least up to the present day, do not even attempt to pass through web forms on a large scale. In this thesis, our primary and key object of study is a huge portion of the Web (hereafter referred as the deep Web) hidden behind web search interfaces. We concentrate on three classes of problems around the deep Web: characterization of deep Web, finding and classifying deep web resources, and querying web databases. <b>Characterizing deep Web: </b>Though the term deep Web was coined in 2000, which is sufficiently long ago for any web-related concept/technology, we still do not know many important characteristics of the deep Web. Another matter of concern is that surveys of the deep Web existing so far are predominantly based on study of deep web sites in English. One can then expect that findings from these surveys may be biased, especially owing to a steady increase in non-English web content. In this way, surveying of national segments of the deep Web is of interest not only to national communities but to the whole web community as well. In this thesis, we propose two new methods for estimating the main parameters of deep Web. We use the suggested methods to estimate the scale of one specific national segment of the Web and report our findings. We also build and make publicly available a dataset describing more than 200 web databases from the national segment of the Web. <b>Finding deep web resources: </b> The deep Web has been growing at a very fast pace. It has been estimated that there are hundred thousands of deep web sites. Due to the huge volume of information in the deep Web, there has been a significant interest to approaches that allow users and computer applications to leverage this information. Most approaches assumed that search interfaces to web databases of interest are already discovered and known to query systems. However, such assumptions do not hold true mostly because of the large scale of the deep Web – indeed, for any given domain of interest there are too many web databases with relevant content. Thus, the ability to locate search interfaces to web databases becomes a key requirement for any application accessing the deep Web. In this thesis, we describe the architecture of the I-Crawler, a system for finding and classifying search interfaces. Specifically, the I-Crawler is intentionally designed to be used in deepWeb characterization studies and for constructing directories of deep web resources. Unlike almost all other approaches to the deep Web existing so far, the I-Crawler is able to recognize and analyze JavaScript-rich and non-HTML searchable forms. <b>Querying web databases:</b> Retrieving information by filling out web search forms is a typical task for a web user. This is all the more so as interfaces of conventional search engines are also web forms. At present, a user needs to manually provide input values to search interfaces and then extract required data from the pages with results. The manual filling out forms is not feasible and cumbersome in cases of complex queries but such kind of queries are essential for many web searches especially in the area of e-commerce. In this way, the automation of querying and retrieving data behind search interfaces is desirable and essential for such tasks as building domain-independent deep web crawlers and automated web agents, searching for domain-specific information (vertical search engines), and for extraction and integration of information from various deep web resources. We present a data model for representing search interfaces and discuss techniques for extracting field labels, client-side scripts and structured data from HTML pages. We also describe a representation of result pages and discuss how to extract and store results of form queries. Besides, we present a user-friendly and expressive form query language that allows one to retrieve information behind search interfaces and extract useful data from the result pages based on specified conditions. We implement a prototype system for querying web databases and describe its architecture and components design.
Resumo:
Search engine optimization & marketing is a set of processes widely used on websites to improve search engine rankings which generate quality web traffic and increase ROI. Content is the most important part of any website. CMS web development is now become very essential for most of organizations and online businesses to develop their online system and websites. Every online business using a CMS wants to get users (customers) to make profit and ROI. This thesis comprises a brief study of existing SEO methods, tools and techniques and how they can be implemented to optimize a content base website. In results, the study provides recommendations about how to use SEO methods; tools and techniques to optimize CMS based websites on major search engines. This study compares popular CMS systems like Drupal, WordPress and Joomla SEO features and how implementing SEO can be improved on these CMS systems. Having knowledge of search engine indexing and search engine working is essential for a successful SEO campaign. This work is a complete guideline for web developers or SEO experts who want to optimize a CMS based website on all major search engines.
Resumo:
This study is dedicated to search engine marketing (SEM). It aims for developing a business model of SEM firms and to provide explicit research of trustworthy practices of virtual marketing companies. Optimization is a general term that represents a variety of techniques and methods of the web pages promotion. The research addresses optimization as a business activity, and it explains its role for the online marketing. Additionally, it highlights issues of unethical techniques utilization by marketers which created relatively negative attitude to them on the Internet environment. Literature insight combines in the one place both technical and economical scientific findings in order to highlight technological and business attributes incorporated in SEM activities. Empirical data regarding search marketers was collected via e-mail questionnaires. 4 representatives of SEM companies were engaged in this study to accomplish the business model design. Additionally, the fifth respondent was a representative of the search engine portal, who provided insight on relations between search engines and marketers. Obtained information of the respondents was processed qualitatively. Movement of commercial organizations to the online market increases demand on promotional programs. SEM is the largest part of online marketing, and it is a prerogative of search engines portals. However, skilled users, or marketers, are able to implement long-term marketing programs by utilizing web page optimization techniques, key word consultancy or content optimization to increase web site visibility to search engines and, therefore, user’s attention to the customer pages. SEM firms are related to small knowledge-intensive businesses. On the basis of data analysis the business model was constructed. The SEM model includes generalized constructs, although they represent a wider amount of operational aspects. Constructing blocks of the model includes fundamental parts of SEM commercial activity: value creation, customer, infrastructure and financial segments. Also, approaches were provided on company’s differentiation and competitive advantages evaluation. It is assumed that search marketers should apply further attempts to differentiate own business out of the large number of similar service providing companies. Findings indicate that SEM companies are interested in the increasing their trustworthiness and the reputation building. Future of the search marketing is directly depending on search engines development.
Resumo:
Poster at Open Repositories 2014, Helsinki, Finland, June 9-13, 2014
Resumo:
Presentation at Open Repositories 2014, Helsinki, Finland, June 9-13, 2014
Resumo:
The objective of this study was to analyze retinol equivalent and iron content in different food composition tables and nutritional evaluation software programs. A literature search was conduct to identify tables and software available in Brazil containing information about retinol equivalent and iron content that are currently used by nutritionists. Ten tables and five software programs were selected for this study. The methodology used to present the retinol equivalent and iron content was evaluated and no pattern to obtain such content was found in the tables and software programs analyzed. Only one of the tables had enough information for the calculation of retinol equivalents; this table is recommended to all Latin America As for the iron content, three of the tables analyzed stand out and therefore should be used; two of them are based on national foods and the other is recommended for use in all Latin America countries. None of the software programs evaluated use the conversion factors suggested by IVACG to assess the vitamin A content in foods. Special attention should be given to the content of iron provided in the software programs since they use tables as international sources and fortified foods.