959 resultados para Information by segment
Resumo:
Current-day web search engines (e.g., Google) do not crawl and index a significant portion of theWeb and, hence, web users relying on search engines only are unable to discover and access a large amount of information from the non-indexable part of the Web. Specifically, dynamic pages generated based on parameters provided by a user via web search forms (or search interfaces) are not indexed by search engines and cannot be found in searchers’ results. Such search interfaces provide web users with an online access to myriads of databases on the Web. In order to obtain some information from a web database of interest, a user issues his/her query by specifying query terms in a search form and receives the query results, a set of dynamic pages that embed required information from a database. At the same time, issuing a query via an arbitrary search interface is an extremely complex task for any kind of automatic agents including web crawlers, which, at least up to the present day, do not even attempt to pass through web forms on a large scale. In this thesis, our primary and key object of study is a huge portion of the Web (hereafter referred as the deep Web) hidden behind web search interfaces. We concentrate on three classes of problems around the deep Web: characterization of deep Web, finding and classifying deep web resources, and querying web databases. Characterizing deep Web: Though the term deep Web was coined in 2000, which is sufficiently long ago for any web-related concept/technology, we still do not know many important characteristics of the deep Web. Another matter of concern is that surveys of the deep Web existing so far are predominantly based on study of deep web sites in English. One can then expect that findings from these surveys may be biased, especially owing to a steady increase in non-English web content. In this way, surveying of national segments of the deep Web is of interest not only to national communities but to the whole web community as well. In this thesis, we propose two new methods for estimating the main parameters of deep Web. We use the suggested methods to estimate the scale of one specific national segment of the Web and report our findings. We also build and make publicly available a dataset describing more than 200 web databases from the national segment of the Web. Finding deep web resources: The deep Web has been growing at a very fast pace. It has been estimated that there are hundred thousands of deep web sites. Due to the huge volume of information in the deep Web, there has been a significant interest to approaches that allow users and computer applications to leverage this information. Most approaches assumed that search interfaces to web databases of interest are already discovered and known to query systems. However, such assumptions do not hold true mostly because of the large scale of the deep Web – indeed, for any given domain of interest there are too many web databases with relevant content. Thus, the ability to locate search interfaces to web databases becomes a key requirement for any application accessing the deep Web. In this thesis, we describe the architecture of the I-Crawler, a system for finding and classifying search interfaces. Specifically, the I-Crawler is intentionally designed to be used in deepWeb characterization studies and for constructing directories of deep web resources. Unlike almost all other approaches to the deep Web existing so far, the I-Crawler is able to recognize and analyze JavaScript-rich and non-HTML searchable forms. Querying web databases: Retrieving information by filling out web search forms is a typical task for a web user. This is all the more so as interfaces of conventional search engines are also web forms. At present, a user needs to manually provide input values to search interfaces and then extract required data from the pages with results. The manual filling out forms is not feasible and cumbersome in cases of complex queries but such kind of queries are essential for many web searches especially in the area of e-commerce. In this way, the automation of querying and retrieving data behind search interfaces is desirable and essential for such tasks as building domain-independent deep web crawlers and automated web agents, searching for domain-specific information (vertical search engines), and for extraction and integration of information from various deep web resources. We present a data model for representing search interfaces and discuss techniques for extracting field labels, client-side scripts and structured data from HTML pages. We also describe a representation of result pages and discuss how to extract and store results of form queries. Besides, we present a user-friendly and expressive form query language that allows one to retrieve information behind search interfaces and extract useful data from the result pages based on specified conditions. We implement a prototype system for querying web databases and describe its architecture and components design.
Resumo:
Designing user interfaces for novel software systems can be challenging since the usability preferences of the users are not well known. This thesis presents a usability study conducted for the development of a user interface for game developers to enter game specific information. By conducting usability testing, the usability preferences of game developers were explored and the design was shaped according to their needs. An assessment of the overall usability of the final design is provided together with the main findings that include the usability preferences and design recommendations. The results showed that the most valuable usability preferences are quickness, error tolerance and the ability to constantly inspect the entered information.
Resumo:
This article reviews the origins of the Documentation, Information and Research Branch (the 'Documentation Center') of Canada's Immigration and Refugee Board (IRB), established in 1988 as a part of a major revision of the procedure for determination of refugee status. The Documentation Center conducts research to produce documents describing conditions in refugee-producing countries, and also disseminates information from outside. The information is available to decision-makers, IRB staff, counsel and claimants. Given the importance of decisions on refugee status, the article looks at the credibility and the authoritativeness of the information, by analyzing the structure of information used. It recalls the different types of information 'package' produced, such as a country profiles and the Question and Answer Series, the Weekly Madia Review, the 'Perspectives' series, Responses to Information Requests and Country files, and considers the trend towards standardization across the country. The research process is reviewed, as are the hiring criteria for researchers, the composition of the 'collection', how acquisitions are made, and the development of databases, particularly on country of origin (human rights material) and legal information, which are accessible on-line. The author examines how documentary information can be used by decision-makers to draw conclusions as to whether the claim has a credible basis or the claimant has a well-founded fear of persecution. Relevant caselaw is available to assess and weigh the claim. The experience of Amnesty International in similar work is cited for comparative purposes. A number of 'safeguards' are mentioned, which contribute to the goal of impartiality in research, or which otherwise enhance the credibility of the information, and the author suggests that guidelines might be drafted to explain and assist in the realization of these aims. Greater resources might also enable the Center to undertake the task of 'certifying' the authoritativeness of sources. The author concludes that, as a new institution in Canadian administrative law, the Documentation Center opens interesting avenues for the future. Beacause it ensures an acceptable degree of impartiality of its research and the documents it produces, it may be a useful model for others tribunals adjudicating in fields where evidence is either difficult to gather, or is otherwise complex.
Resumo:
Le présent mémoire cherche à comprendre et à cerner le lien entre la stratégie de recherche d’information par le journaliste sur le web et les exigences de sa profession. Il vise à appréhender les précautions que prend le journaliste lors de sa recherche d’information sur le web en rapport avec les contraintes que lui imposent les règles de sa profession pour assurer la qualité des sources d’informations qu’il exploite. Nous avons examiné cette problématique en choisissant comme cadre d’étude Radio-Canada où nous avons rencontré quelques journalistes. Ceux-ci ont été suivis en situation de recherche d’information puis questionnés sur leurs expériences de recherche. L’arrivée d’internet et la révolution technologique qui en a découlé ont profondément bouleversé les pratiques journalistiques. La recherche d’information représente ainsi une zone importante de cette mutation des pratiques. Cette transformation amène surtout à s’interroger sur la façon dont la nouvelle façon de rechercher les sources d’information influence le travail du journaliste, et surtout les balises que se donne celui-ci pour résister aux pièges découlant de sa nouvelle méthode de travail.
Resumo:
Cette thèse est une collection de trois articles en économie de l'information. Le premier chapitre sert d'introduction et les Chapitres 2 à 4 constituent le coeur de l'ouvrage. Le Chapitre 2 porte sur l’acquisition d’information sur l’Internet par le biais d'avis de consommateurs. En particulier, je détermine si les avis laissés par les acheteurs peuvent tout de même transmettre de l’information à d’autres consommateurs, lorsqu’il est connu que les vendeurs peuvent publier de faux avis à propos de leurs produits. Afin de comprendre si cette manipulation des avis est problématique, je démontre que la plateforme sur laquelle les avis sont publiés (e.g. TripAdvisor, Yelp) est un tiers important à considérer, autant que les vendeurs tentant de falsifier les avis. En effet, le design adopté par la plateforme a un effet indirect sur le niveau de manipulation des vendeurs. En particulier, je démontre que la plateforme, en cachant une partie du contenu qu'elle détient sur les avis, peut parfois améliorer la qualité de l'information obtenue par les consommateurs. Finalement, le design qui est choisi par la plateforme peut être lié à la façon dont elle génère ses revenus. Je montre qu'une plateforme générant des revenus par le biais de commissions sur les ventes peut être plus tolérante à la manipulation qu'une plateforme qui génère des revenus par le biais de publicité. Le Chapitre 3 est écrit en collaboration avec Marc Santugini. Dans ce chapitre, nous étudions les effets de la discrimination par les prix au troisième degré en présence de consommateurs non informés qui apprennent sur la qualité d'un produit par le biais de son prix. Dans un environnement stochastique avec deux segments de marché, nous démontrons que la discrimination par les prix peut nuire à la firme et être bénéfique pour les consommateurs. D'un côté, la discrimination par les prix diminue l'incertitude à laquelle font face les consommateurs, c.-à-d., la variance des croyances postérieures est plus faible avec discrimination qu'avec un prix uniforme. En effet, le fait d'observer deux prix (avec discrimination) procure plus d'information aux consommateurs, et ce, même si individuellement chacun de ces prix est moins informatif que le prix uniforme. De l'autre côté, il n'est pas toujours optimal pour la firme de faire de la discrimination par les prix puisque la présence de consommateurs non informés lui donne une incitation à s'engager dans du signaling. Si l'avantage procuré par la flexibilité de fixer deux prix différents est contrebalancé par le coût du signaling avec deux prix différents, alors il est optimal pour la firme de fixer un prix uniforme sur le marché. Finalement, le Chapitre 4 est écrit en collaboration avec Sidartha Gordon. Dans ce chapitre, nous étudions une classe de jeux où les joueurs sont contraints dans le nombre de sources d'information qu'ils peuvent choisir pour apprendre sur un paramètre du jeu, mais où ils ont une certaine liberté quant au degré de dépendance de leurs signaux, avant de prendre une action. En introduisant un nouvel ordre de dépendance entre signaux, nous démontrons qu'un joueur préfère de l'information qui est la plus dépendante possible de l'information obtenue par les joueurs pour qui les actions sont soit, compléments stratégiques et isotoniques, soit substituts stratégiques et anti-toniques, avec la sienne. De même, un joueur préfère de l'information qui est la moins dépendante possible de l'information obtenue par les joueurs pour qui les actions sont soit, substituts stratégiques et isotoniques, soit compléments stratégiques et anti-toniques, avec la sienne. Nous établissons également des conditions suffisantes pour qu'une structure d'information donnée, information publique ou privée par exemple, soit possible à l'équilibre.
Resumo:
Inadequate links between researchers and farmers has resulted in low uptake of research advances recommended to improve food security in the central highlands of Kenya. Access to timely and accurate information by extension agents and farmers is paramount in dissemination of soil fertility management practices. Hence, the study sought to investigate the effect of education levels on communication channels used to disseminate soil fertility technologies in the Central highlands of Kenya. Questionnaires were used to elicit information from 105 extension agents and 240 farmers. About 50.5% of the extension officers were certificate holders while 29.5% were diploma holders from agricultural institutes. Majority of the farmers had attained primary education (59.6%) while 25.8% and 9.2% had attained secondary and post secondary education, respectively. Research institutions were the most accessible sources of information on soil fertility management practices by extension agents while internet and scientific conferences were the least scored as accessible sources of soil fertility management information by extension agents. Education levels significantly influenced preference of individual approach methods by farmers. There was a significant positive relationship between education and accessibility of internet as a source of information on green manure. The implication of the study was that education levels influenced the mode of communication used in the transfer of soil fertility research outputs to the end users. Consequently, it is extremely important to consider education levels in selection of dissemination pathways used in agriculture.
Resumo:
The paper looks at the setting up of media services for British Army forces engaged in Operation Granby, the British participation in Operation Desert Storm, the liberation of Kuwait from Iraqi occupation in 1991. In particular, the piece focuses on the creation of a force newspaper and radio station.There is far more involved in the organisation of modern war than simply battles themselves One aspect is the importance of creating and maintaining morale, which can be achieved in a number of ways, especially by keeping troops informed of their progress, purpose and mission as operations unfold. This was specifically recognised by commanders in the Gulf War, who ensured the smooth transmission of information by the provision of media services for forces personnel throughout the War. Here S J Anglim concentrates on the difficulties inherent in setting up and running the Gulf Forces newspaper, The Sandy Times, a Radio service and producing video reviews, all of which played their part in ensunng high morale. He concludes that there are lessons to be learnt and implemented for the success of any future conflicts.
Resumo:
Feedlot consulting nutritionists were invited to participate in a survey of feedlot nutritional and management practices in Brazil. Thirty-one nutritionists completed the survey on a Web site that was designed for collection of survey data. The survey consisted of 94 questions that included general information (n = 10); commodity information (n = 12); and questions about the use of coproducts (n = 5), roughage source and levels (n = 5), finishing diet adaptation methods (n = 7), supplements and micronutrients (n = 8), feed mixers (n = 6), feeding management (n = 3), cattle management and type of cattle fed (n = 16), formulation practices (n = 17), information resources used for nutritional recommendations (n = 2), and 2 additional questions. One final question addressed the primary challenges associated with applying nutritional recommendations in practice. The number of animals serviced yearly by each nutritionist averaged 121,682 (minimum = 2,000; maximum = 1,500,000; mode = 120,000; total = 3,163,750). Twenty-two respondents (71%) worked with feedlots that feed less than 5,000 animals/yr. Labor, along with availability and precision of equipment, seemed to be the main challenges for the nutritionists surveyed. Most of the nutritionists surveyed used TDN as the primary energy unit for formulation. More than 50% of the clients serviced by the 31 nutritionists did not manage feed bunks to control the quantity of feed offered per pen, and 36.6% fed cattle more than 4 times daily. The NRC (1996) and Journal of Animal Science were the most used sources of information by these nutritionists. Overall, general practices and nutritional recommendations provided by the 31 nutritionists surveyed were fairly consistent. Present data should aid in development of new research, future National Research Council models, and recommendations for Brazilian feeding systems in which Bos indicus cattle predominate.
Resumo:
Three-dimensional kinematic analysis of line of gaze, arm and ball was used to describe the visual and motor behaviour of male adolescents diagnosed with attention deficit hyperactivity disorder (ADHD). The ADHD participants were tested when both on (ADHD-On) and off (ADHD-Off) their medication and compared to age-matched normal controls in a modified table tennis task that required tracking the ball and hitting to cued right and left targets. Long-duration information was provided by a pre-cue, in which the target was illuminated approximately 2 s before the serve, and short-duration information by an early-cue illuminated about 350 ms after the serve, leaving -500 ms to select the target and perform the action. The ADHD groups differed significantly from the control group in both the pre-cue and early-cue conditions in being less accurate, in having a later onset and duration of pursuit tracking, and a higher frequency of gaze on and off the ball. The use of medication significantly reduced the gaze frequency of the ADHD participants, but surprisingly this did not lead to an increase in pursuit tracking, suggesting a barrier was reached beyond which ball flight information could not be processed. The control and ADHD groups did not differ in arm movement onset, duration and velocity in the short-duration early-cue condition; in the long-duration pre-cue condition, however, the ADHD group's movement time onset and arm velocity differed significantly from controls. The results show that the ADHD groups were able to process short-duration information without experiencing adverse effects on their motor behaviour; however, long-duration information contributed to irregular movement control.
Resumo:
The aim of this study was to explore female community health agents’ views about the value of recording qualitative information on contextual health issues they observe during home visits, data that are not officially required to be documented for the Brazilian System of Primary Healthcare Information. Background: The study was conducted in community primary healthcare centres located in the cities of Araçatuba and Coroados (state of São Paulo) and Rio de Janeiro (state of Rio de Janeiro), Brazil. Methods: The design was a qualitative, exploratory study. The purposeful sampling criteria were being female, with a minimum of three years of continuous service in the same location. Data collection with 62 participants was conducted via 11 focus groups (in 2007 and 2008). Audio files were transcribed and submitted to the method of thematic analysis. Four themes guided the analysis: working with qualitative information and undocumented observation; reflecting on qualitative information; integrating/analysing quantitative and qualitative information; and information-sharing with agents and family health teams. In 2010, 25 community health agents verified the final interpretation of the findings. Findings: Participants valued the recording of qualitative, contextual information to expand understanding of primary healthcare issues and as an indicator of clients’ improved health behaviour and health literacy. While participants initiated the recording of additional health information, they generally did not inform the family health team about these findings. They perceived that team members devalued this type of information by considering it a reflection of the clientele’s social conditions or problems beyond the scope of medical concerns. Documentation of qualitative evidence can account for the effectiveness of health education in two ways: by improving preventative care, and by amplifying the voices of underprivileged clients who live in poverty to ensure the most appropriate and best quality primary healthcare for them.
Resumo:
Analisam-se as características de ofertas de emprego do setor da Informação-Documentação. A mostra foi obtida de anúncios na Internet, feitos por empresas e instituições privadas brasileiras, durante o ano de 2010. A atividade foi realizada no âmbito de pesquisa maior, que tem como objetivo geral construir uma taxonomia que classifique e hierarquize as competências e conhecimentos exigidos pelas empresas, bem como as características dos postos oferecidos. Extraiu-se a informação de 200 ofertas de emprego obtidas nos portais Catho e Portal do bibliotecário. Cada entrada foi mantida em seu contexto de incidência e com a expressão original. Os termos obtidos passaram por um processo de normalização. A análise busca classificar e valorar as atividades, conhecimentos e tarefas mais solicitados, bem como as aptidões, competências e habilidades, além das incidências das unidades e dos profissionais de informação.
Resumo:
Understanding and controlling the mechanism of the diffusion of small molecules, macromolecules and nanoparticles in heterogeneous environments is of paramount fundamental and technological importance. The aim of the thesis is to show, how by studying the tracer diffusion in complex systems, one can obtain information about the tracer itself, and the system where the tracer is diffusing. rnIn the first part of my thesis I will introduce the Fluorescence Correlation Spectroscopy (FCS) which is a powerful tool to investigate the diffusion of fluorescent species in various environments. By using the main advantage of FCS namely the very small probing volume (<1µm3) I was able to track the kinetics of phase separation in polymer blends at late stages by looking on the molecular tracer diffusion in individual domains of the heterogeneous structure of the blend. The phase separation process at intermediate stages was monitored with laser scanning confocal microscopy (LSCM) in real time providing images of droplet coalescence and growth. rnIn a further project described in my thesis I will show that even when the length scale of the heterogeneities becomes smaller than the FCS probing volume one can still obtain important microscopic information by studying small tracer diffusion. To do so, I will introduce a system of star shaped polymer solutions and will demonstrate that the mobility of small molecular tracers on microscopic level is nearly not affected by the transition of the polymer system to a “glassy” macroscopic state. rnIn the last part of the thesis I will introduce and describe a new stimuli responsive system which I have developed, that combines two levels of nanoporosity. The system is based on poly-N-isopropylacrylamide (PNIPAM) and silica inverse opals (iOpals), and allows controlling the diffusion of tracer molecules. rn
Resumo:
For smart cities applications, a key requirement is to disseminate data collected from both scalar and multimedia wireless sensor networks to thousands of end-users. Furthermore, the information must be delivered to non-specialist users in a simple, intuitive and transparent manner. In this context, we present Sensor4Cities, a user-friendly tool that enables data dissemination to large audiences, by using using social networks, or/and web pages. The user can request and receive monitored information by using social networks, e.g., Twitter and Facebook, due to their popularity, user-friendly interfaces and easy dissemination. Additionally, the user can collect or share information from smart cities services, by using web pages, which also include a mobile version for smartphones. Finally, the tool could be configured to periodically monitor the environmental conditions, specific behaviors or abnormal events, and notify users in an asynchronous manner. Sensor4Cities improves the data delivery for individuals or groups of users of smart cities applications and encourages the development of new user-friendly services.
Resumo:
The hippocampus receives input from upper levels of the association cortex and is implicated in many mnemonic processes, but the exact mechanisms by which it codes and stores information is an unresolved topic. This work examines the flow of information through the hippocampal formation while attempting to determine the computations that each of the hippocampal subfields performs in learning and memory. The formation, storage, and recall of hippocampal-dependent memories theoretically utilize an autoassociative attractor network that functions by implementing two competitive, yet complementary, processes. Pattern separation, hypothesized to occur in the dentate gyrus (DG), refers to the ability to decrease the similarity among incoming information by producing output patterns that overlap less than the inputs. In contrast, pattern completion, hypothesized to occur in the CA3 region, refers to the ability to reproduce a previously stored output pattern from a partial or degraded input pattern. Prior to addressing the functional role of the DG and CA3 subfields, the spatial firing properties of neurons in the dentate gyrus were examined. The principal cell of the dentate gyrus, the granule cell, has spatially selective place fields; however, the behavioral correlates of another excitatory cell, the mossy cell of the dentate polymorphic layer, are unknown. This report shows that putative mossy cells have spatially selective firing that consists of multiple fields similar to previously reported properties of granule cells. Other cells recorded from the DG had single place fields. Compared to cells with multiple fields, cells with single fields fired at a lower rate during sleep, were less likely to burst, and were more likely to be recorded simultaneously with a large population of neurons that were active during sleep and silent during behavior. These data suggest that single-field and multiple-field cells constitute at least two distinct cell classes in the DG. Based on these characteristics, we propose that putative mossy cells tend to fire in multiple, distinct locations in an environment, whereas putative granule cells tend to fire in single locations, similar to place fields of the CA1 and CA3 regions. Experimental evidence supporting the theories of pattern separation and pattern completion comes from both behavioral and electrophysiological tests. These studies specifically focused on the function of each subregion and made implicit assumptions about how environmental manipulations changed the representations encoded by the hippocampal inputs. However, the cell populations that provided these inputs were in most cases not directly examined. We conducted a series of studies to investigate the neural activity in the entorhinal cortex, dentate gyrus, and CA3 in the same experimental conditions, which allowed a direct comparison between the input and output representations. The results show that the dentate gyrus representation changes between the familiar and cue altered environments more than its input representations, whereas the CA3 representation changes less than its input representations. These findings are consistent with longstanding computational models proposing that (1) CA3 is an associative memory system performing pattern completion in order to recall previous memories from partial inputs, and (2) the dentate gyrus performs pattern separation to help store different memories in ways that reduce interference when the memories are subsequently recalled.
Resumo:
Eight surface observation sites providing quasi-continuous measurements of atmospheric methane mixingratios have been operated since the mid-2000’s in Siberia. For the first time in a single work, we assimilate 1 year of these in situ observations in an atmospheric inversion. Our objective is to quantify methane surface fluxes from anthropogenic and wetland sources at the mesoscale in the Siberian lowlands for the year 2010. To do so, we first inquire about the way the inversion uses the observations and the way the fluxes are constrained by the observation sites. As atmospheric inver- sions at the mesoscale suffer from mis-quantified sources of uncertainties, we follow recent innovations in inversion techniques and use a new inversion approach which quantifies the uncertainties more objectively than the previous inversion systems. We find that, due to errors in the representation of the atmospheric transport and redundant pieces of information, only one observation every few days is found valuable by the inversion. The remaining high-resolution quasi-continuous signal is representative of very local emission patterns difficult to analyse with a mesoscale system. An analysis of the use of information by the inversion also reveals that the observation sites constrain methane emissions within a radius of 500 km. More observation sites than the ones currently in operation are then necessary to constrain the whole Siberian lowlands. Still, the fluxes within the constrained areas are quantified with objectified uncertainties. Finally, the tolerance intervals for posterior methane fluxes are of roughly 20 % (resp. 50 %) of the fluxes for anthropogenic (resp. wetland) sources. About 50–70 % of Siberian lowlands emissions are constrained by the inversion on average on an annual basis. Extrapolating the figures on the constrained areas to the whole Siberian lowlands, we find a regional methane budget of 5–28 TgCH4 for the year 2010, i.e. 1–5 % of the global methane emissions. As very few in situ observations are available in the region of interest, observations of methane total columns from the Greenhouse Gas Observing SATellite (GOSAT) are tentatively used for the evaluation of the inversion results, but they exhibit only a marginal signal from the fluxes within the region of interest.