866 resultados para topic, web information gathering, web personalization
Resumo:
Current-day web search engines (e.g., Google) do not crawl and index a significant portion of theWeb and, hence, web users relying on search engines only are unable to discover and access a large amount of information from the non-indexable part of the Web. Specifically, dynamic pages generated based on parameters provided by a user via web search forms (or search interfaces) are not indexed by search engines and cannot be found in searchers’ results. Such search interfaces provide web users with an online access to myriads of databases on the Web. In order to obtain some information from a web database of interest, a user issues his/her query by specifying query terms in a search form and receives the query results, a set of dynamic pages that embed required information from a database. At the same time, issuing a query via an arbitrary search interface is an extremely complex task for any kind of automatic agents including web crawlers, which, at least up to the present day, do not even attempt to pass through web forms on a large scale. In this thesis, our primary and key object of study is a huge portion of the Web (hereafter referred as the deep Web) hidden behind web search interfaces. We concentrate on three classes of problems around the deep Web: characterization of deep Web, finding and classifying deep web resources, and querying web databases. Characterizing deep Web: Though the term deep Web was coined in 2000, which is sufficiently long ago for any web-related concept/technology, we still do not know many important characteristics of the deep Web. Another matter of concern is that surveys of the deep Web existing so far are predominantly based on study of deep web sites in English. One can then expect that findings from these surveys may be biased, especially owing to a steady increase in non-English web content. In this way, surveying of national segments of the deep Web is of interest not only to national communities but to the whole web community as well. In this thesis, we propose two new methods for estimating the main parameters of deep Web. We use the suggested methods to estimate the scale of one specific national segment of the Web and report our findings. We also build and make publicly available a dataset describing more than 200 web databases from the national segment of the Web. Finding deep web resources: The deep Web has been growing at a very fast pace. It has been estimated that there are hundred thousands of deep web sites. Due to the huge volume of information in the deep Web, there has been a significant interest to approaches that allow users and computer applications to leverage this information. Most approaches assumed that search interfaces to web databases of interest are already discovered and known to query systems. However, such assumptions do not hold true mostly because of the large scale of the deep Web – indeed, for any given domain of interest there are too many web databases with relevant content. Thus, the ability to locate search interfaces to web databases becomes a key requirement for any application accessing the deep Web. In this thesis, we describe the architecture of the I-Crawler, a system for finding and classifying search interfaces. Specifically, the I-Crawler is intentionally designed to be used in deepWeb characterization studies and for constructing directories of deep web resources. Unlike almost all other approaches to the deep Web existing so far, the I-Crawler is able to recognize and analyze JavaScript-rich and non-HTML searchable forms. Querying web databases: Retrieving information by filling out web search forms is a typical task for a web user. This is all the more so as interfaces of conventional search engines are also web forms. At present, a user needs to manually provide input values to search interfaces and then extract required data from the pages with results. The manual filling out forms is not feasible and cumbersome in cases of complex queries but such kind of queries are essential for many web searches especially in the area of e-commerce. In this way, the automation of querying and retrieving data behind search interfaces is desirable and essential for such tasks as building domain-independent deep web crawlers and automated web agents, searching for domain-specific information (vertical search engines), and for extraction and integration of information from various deep web resources. We present a data model for representing search interfaces and discuss techniques for extracting field labels, client-side scripts and structured data from HTML pages. We also describe a representation of result pages and discuss how to extract and store results of form queries. Besides, we present a user-friendly and expressive form query language that allows one to retrieve information behind search interfaces and extract useful data from the result pages based on specified conditions. We implement a prototype system for querying web databases and describe its architecture and components design.
Resumo:
This paper presents the current state and development of a prototype web-GIS (Geographic Information System) decision support platform intended for application in natural hazards and risk management, mainly for floods and landslides. This web platform uses open-source geospatial software and technologies, particularly the Boundless (formerly OpenGeo) framework and its client side software development kit (SDK). The main purpose of the platform is to assist the experts and stakeholders in the decision-making process for evaluation and selection of different risk management strategies through an interactive participation approach, integrating web-GIS interface with decision support tool based on a compromise programming approach. The access rights and functionality of the platform are varied depending on the roles and responsibilities of stakeholders in managing the risk. The application of the prototype platform is demonstrated based on an example case study site: Malborghetto Valbruna municipality of North-Eastern Italy where flash floods and landslides are frequent with major events having occurred in 2003. The preliminary feedback collected from the stakeholders in the region is discussed to understand the perspectives of stakeholders on the proposed prototype platform.
Resumo:
This paper presents a prototype of an interactive web-GIS tool for risk analysis of natural hazards, in particular for floods and landslides, based on open-source geospatial software and technologies. The aim of the presented tool is to assist the experts (risk managers) in analysing the impacts and consequences of a certain hazard event in a considered region, providing an essential input to the decision-making process in the selection of risk management strategies by responsible authorities and decision makers. This tool is based on the Boundless (OpenGeo Suite) framework and its client-side environment for prototype development, and it is one of the main modules of a web-based collaborative decision support platform in risk management. Within this platform, the users can import necessary maps and information to analyse areas at risk. Based on provided information and parameters, loss scenarios (amount of damages and number of fatalities) of a hazard event are generated on the fly and visualized interactively within the web-GIS interface of the platform. The annualized risk is calculated based on the combination of resultant loss scenarios with different return periods of the hazard event. The application of this developed prototype is demonstrated using a regional data set from one of the case study sites, Fella River of northeastern Italy, of the Marie Curie ITN CHANGES project.
Resumo:
Occupational hygiene practitioners typically assess the risk posed by occupational exposure by comparing exposure measurements to regulatory occupational exposure limits (OELs). In most jurisdictions, OELs are only available for exposure by the inhalation pathway. Skin notations are used to indicate substances for which dermal exposure may lead to health effects. However, these notations are either present or absent and provide no indication of acceptable levels of exposure. Furthermore, the methodology and framework for assigning skin notation differ widely across jurisdictions resulting in inconsistencies in the substances that carry notations. The UPERCUT tool was developed in response to these limitations. It helps occupational health stakeholders to assess the hazard associated with dermal exposure to chemicals. UPERCUT integrates dermal quantitative structure-activity relationships (QSARs) and toxicological data to provide users with a skin hazard index called the dermal hazard ratio (DHR) for the substance and scenario of interest. The DHR is the ratio between the estimated 'received' dose and the 'acceptable' dose. The 'received' dose is estimated using physico-chemical data and information on the exposure scenario provided by the user (body parts exposure and exposure duration), and the 'acceptable' dose is estimated using inhalation OELs and toxicological data. The uncertainty surrounding the DHR is estimated with Monte Carlo simulation. Additional information on the selected substances includes intrinsic skin permeation potential of the substance and the existence of skin notations. UPERCUT is the only available tool that estimates the absorbed dose and compares this to an acceptable dose. In the absence of dermal OELs it provides a systematic and simple approach for screening dermal exposure scenarios for 1686 substances.
Resumo:
In this paper we describe a browsing and searching personalization system for digitallibraries based on the use of ontologies for describing the relationships between all theelements which take part in a digital library scenario of use. The main goal of thisproject is to help the users of a digital library to improve their experience of use bymeans of two complementary strategies: first, by maintaining a complete history recordof his or her browsing and searching activities, which is part of a navigational userprofile which includes preferences and all the aspects related to community involvement; and second, by reusing all the knowledge which has been extracted from previous usage from other users with similar profiles. This can be accomplished in terms of narrowing and focusing the search results and browsing options through the use of a recommendation system which organizes such results in the most appropriatemanner, using ontologies and concepts drawn from the semantic web field. The complete integration of the experience of use of a digital library in the learning process is also pursued. Both the usage and information organization can be also exploited to extract useful knowledge from the way users interact with a digital library, knowledge that can be used to improve several design aspects of the library, ranging from internal organization aspects to human factors and user interfaces. Although this project is still on an early development stage, it is possible to identify all the desired functionalities and requirements that are necessary to fully integrate the use of a digital library in an e-learning environment.
Resumo:
E-learning, understood as the intensive use of Information and Communication Technologies in mainly but not only) distance education, has radically changed the meaning of the latter. E-learning is an overused term which has been applied to any use of technology in education. Today, the most widely accepted meaning ofe-learning coincides with the fourth generation described by Taylor (1999), where there is an asynchronousprocess that allows students and teachers to interact in an educational process expressly designed in accordance with these principles. We prefer to speak of Internet-Based Learning or, better still, Web-Based Learning, for example, to explain the fact that distance education is carried out using the Internet, with the appearance of the virtual learning environment concept, a web space where the teaching and learning process is generated and supported (Sangrà, 2002). This entails overcoming the barriers of space and time of brickand mortar education (although we prefer the term face-to-face) or of classical distance education using broadcasting and adopting a completely asynchronous model that allows access to education by many more users, at any level (including secondary education, but primarily higher education and lifelong learning).
Resumo:
Tässä työssä selvitettiin Ajax-tekniikan tilannetta web-sovellusten kehityksessä. Sitä varten kehitettiin demosovellus, jonka avulla tekniikoiden käyttökelpoisuutta voitiin arvioida. Samalla työssä on esitelty eri tekniikoita, jotka liittyvät kiinteästi Ajax-sovellusten toteuttamiseen. Demosovellus tehtiin vapaalla LAMP (Linux, Apache, MySQL and PHP) -alustalla. Työssä on arvioitu Ajax-tekniikan käyttökelpoisuutta ja ongelmia nykyisen webin, web-kehittäjien, käytössä olevien selainten ja käyttäjien kannalta. Lopussa on myös pohdittu hieman webin tulevaisuutta ja Ajaxin osaa siinä.
Resumo:
El projecte té per objecte explorar les possibilitats dels sistemes d'informació geogràfica en els estudis de vigilància epidemiològica (VE).
Resumo:
Disseny i implementació d’aplicació web destinada a consulta d’informació urbana en els municipis de Mallorca. Localització de punts d’interès. Implementació pilot en el municipi de Santa Eugènia, Mallorca
Resumo:
Tietojärjestelmien integraatio on nykypäivänä tärkeä osa alue yritysten toiminnassa ja kilpailukyvyn ylläpitämisessä. Palvelukeskeinen arkkitehtuuri ja Web palvelut on uusi joustava tapa tehdä tietojärjestelmien välinen integraatio. Web palveluiden yksi ydinkomponentti on UDDI, Universal Description, Discovery and Integration. UDDI toimii palvelurekisterin tavoin. UDDI määrittää tavan julkaista, löytää ja ottaa käyttöön Web palveluja. Web palveluja voidaan hakea UDDI:sta erilaisin kriteerein, kuten esimerkiksi palvelun sijainnin, yrityksen nimen ja toimialan perusteella. UDDI on myös itsessään Web palvelu, joka perustuu XML kuvauskieleen ja SOAP protokollaan. Työssä paneudutaan tarkemmin UDDI:in. UDDI:ta käsitellään tarkemmin myös teknisesti. Oleellinen osa UDDI:ta on ollut julkaisijoiden ja käyttäjien mielestä tietoturvan puute, joka on rajoittanut huomattavasti UDDI:n käyttöä ja käyttöönottamista. Työssä tarkastellaankin tarkemmin juuri tietoturvaan liittyviä asioita ja ratkaisuja sekä myös UDDI:n merkitystä yrityksille.
Resumo:
Las bibliotecas universitarias realizan tutoriales web sobre una gran variedad de temáticas. Muchas veces van dirigidos a la alfabetización informacional (alfin) y a mostrar a los usuarios las mejoras tecnológicas de las interfaces de consulta interactiva. Se ofrece una metodología basada en treinta y cuatro indicadores para elaborar, evaluar y mejorar los tutoriales creados por las bibliotecas de cualquier ámbito y especialidad. Se presentan adicionalmente ejemplos de buenas prácticas que ilustran la descripción del indicador y pueden servir de referencia para los servicios que deseen crear sus propios tutoriales
Resumo:
Due to concerns regarding globalisation and sustainable development, corporate social responsibility (CSR) is topical in the business context and in the field of accounting. The main objective of this study was to review previous academic literature in the field of CSR reporting and develop an insight into CSR reporting in the Web-based environment. The main purpose was to find out what Web-based CSR reporting is like and how companies are utilising the Internet to communicate on responsibility issues. I did not, however, collect empirical research data but limited my study into theoretical and descriptive examination. In order to create an insight into Web-based reporting, I examined the development, motives and current practices of CSR reporting. I concluded that the Internet is a unique, interactive communication channel that is used differently compared with annual reports. The amount of companies engaging in Web-based CSR reporting is increasing and the reporting practices in terms of e.g. content and accessibility of information vary. I also concluded that many companies have not yet discovered the true potential of the Web as an interactive communication medium.
Resumo:
Amb la situació econòmica actual pot ser interessant poder vendre objectes que ja no s’utilitzen i també poder-ne comprar de segona mà. Amb aquesta idea sorgeix el projecte de crear una pàgina de subhastes online on la gent pugui comerciar amb les coses que ja no necessita. Tenint en compte el concepte inicial, el propietari de la pàgina no rebrà cap retribució ni percentatge de cada subhasta, tot l’import serà pel venedor. L’objectiu principal és el de poder oferir un lloc on després de registrar-se, els usuaris puguin veure i pujar per els articles que altres persones estan subhastant i també la possibilitat de crear les seves pròpies subhastes. Cada usuari disposarà d’un espai personal on veure les subhastes amb les que ha interactuat i així no perdre-les de vista i també on poder veure en cada moment l’estat de les subhastes que ha creat. La vista d’una subhasta s’actualitzarà automàticament sense haver de recarregar la pàgina i si algú puja durant l’últim minut la subhasta s’allargarà un minut més per evitar puges a l’últim moment i així maximitzar el preu final. Hi haurà un administrador que serà l’encarregat de gestionar el bon funcionament de la pàgina amb permís per afegir, editar, consultar i eliminar tota la informació disponible. Per portar a terme el projecte s’ha utilitzat PHP per la part de programació i MySQL com a sistema gestor de bases de dades.
Resumo:
Vivim en una era digital on cada vegada més les persones estem connectades a la xarxa, ja sigui a través del mòbil, ordinadors o altres dispositius. Actualment internet és un gran aparador i des de fa temps, les empreses han vist en aquest una manera de treure’n profit. És per això, junt amb el seu baix cost, que fa que qualsevol empresa disposi del seu propi espai a la xarxa. D’aquestes idees sorgeix el projecte de fer la web per AutoSuministres Motor. La principal funció de la web es donar-se a conèixer i ensenyar els seus productes, que en aquest cas, són autocaravanes i caravanes. Tot i així, la pàgina web contindrà un espai de reportatges on l’usuari podrà conèixer més a fons un model concret d’autocaravana o caravana i on també podrà demanar més informació mitjançant un formulari de contacte. Per tant, la pàgina web serà totalment administrable perquè els comercials puguin introduir els vehicles corresponents i l’usuari final pugui consultarlos correctament d’una manera fàcil i senzilla. Per tal de dur a terme aquesta web s’ha tingut en compte utilitzar eines de programació de distribució lliure com és el llenguatge PHP, la base de dades MySQL i jQuery.
Resumo:
We present here the first part of the literature review regarding our study object, the Open Device Labs. The research on ODLs emerges from the observation of worldwide non-profit movement, which, through mutual collaboration, information and devices sharing, proposes a final improvement on user’s experience with the web and app.