12 resultados para web map service (WMS)
em Helda - Digital Repository of University of Helsinki
Resumo:
With the recent increase in interest in service-oriented architectures (SOA) and Web services, developing applications with the Web services paradigm has become feasible. Web services are self-describing, platform-independent computational elements. New applications can be assembled from a set of previously created Web services, which are composed together to make a service that uses its components to perform a certain task. This is the idea of service composition. To bring service composition to a mobile phone, I have created Interactive Service Composer for mobile phones. With Interactive Service Composer, the user is able to build service compositions on his mobile phone, consisting of Web services or services that are available from the mobile phone itself. The service compositions are reusable and can be saved in the phone's memory. Previously saved compositions can also be used in new compositions. While developing applications for mobile phones has been possible for some time, the usability of the solutions is not the same as when developing for desktop computers. When developing for mobile phones, the developer has to more carefully consider the decisions he is going to make with the program he is developing. With the lack of processing power and memory, the applications cannot function as well as on desktop PCs. On the other hand, this does not remove the appeal of developing applications for mobile devices.
Resumo:
In recent years, XML has been accepted as the format of messages for several applications. Prominent examples include SOAP for Web services, XMPP for instant messaging, and RSS and Atom for content syndication. This XML usage is understandable, as the format itself is a well-accepted standard for structured data, and it has excellent support for many popular programming languages, so inventing an application-specific format no longer seems worth the effort. Simultaneously with this XML's rise to prominence there has been an upsurge in the number and capabilities of various mobile devices. These devices are connected through various wireless technologies to larger networks, and a goal of current research is to integrate them seamlessly into these networks. These two developments seem to be at odds with each other. XML as a fully text-based format takes up more processing power and network bandwidth than binary formats would, whereas the battery-powered nature of mobile devices dictates that energy, both in processing and transmitting, be utilized efficiently. This thesis presents the work we have performed to reconcile these two worlds. We present a message transfer service that we have developed to address what we have identified as the three key issues: XML processing at the application level, a more efficient XML serialization format, and the protocol used to transfer messages. Our presentation includes both a high-level architectural view of the whole message transfer service, as well as detailed descriptions of the three new components. These components consist of an API, and an associated data model, for XML processing designed for messaging applications, a binary serialization format for the data model of the API, and a message transfer protocol providing two-way messaging capability with support for client mobility. We also present relevant performance measurements for the service and its components. As a result of this work, we do not consider XML to be inherently incompatible with mobile devices. As the fixed networking world moves toward XML for interoperable data representation, so should the wireless world also do to provide a better-integrated networking infrastructure. However, the problems that XML adoption has touch all of the higher layers of application programming, so instead of concentrating simply on the serialization format we conclude that improvements need to be made in an integrated fashion in all of these layers.
Resumo:
As the virtual world grows more complex, finding a standard way for storing data becomes increasingly important. Ideally, each data item would be brought into the computer system only once. References for data items need to be cryptographically verifiable, so the data can maintain its identity while being passed around. This way there will be only one copy of the users family photo album, while the user can use multiple tools to show or manipulate the album. Copies of users data could be stored on some of his family members computer, some of his computers, but also at some online services which he uses. When all actors operate over one replicated copy of the data, the system automatically avoids a single point of failure. Thus the data will not disappear with one computer breaking, or one service provider going out of business. One shared copy also makes it possible to delete a piece of data from all systems at once, on users request. In our research we tried to find a model that would make data manageable to users, and make it possible to have the same data stored at various locations. We studied three systems, Persona, Freenet, and GNUnet, that suggest different models for protecting user data. The main application areas of the systems studied include securing online social networks, providing anonymous web, and preventing censorship in file-sharing. Each of the systems studied store user data on machines belonging to third parties. The systems differ in measures they take to protect their users from data loss, forged information, censorship, and being monitored. All of the systems use cryptography to secure names used for the content, and to protect the data from outsiders. Based on the gained knowledge, we built a prototype platform called Peerscape, which stores user data in a synchronized, protected database. Data items themselves are protected with cryptography against forgery, but not encrypted as the focus has been disseminating the data directly among family and friends instead of letting third parties store the information. We turned the synchronizing database into peer-to-peer web by revealing its contents through an integrated http server. The REST-like http API supports development of applications in javascript. To evaluate the platform’s suitability for application development we wrote some simple applications, including a public chat room, bittorrent site, and a flower growing game. During our early tests we came to the conclusion that using the platform for simple applications works well. As web standards develop further, writing applications for the platform should become easier. Any system this complex will have its problems, and we are not expecting our platform to replace the existing web, but are fairly impressed with the results and consider our work important from the perspective of managing user data.
Resumo:
As the virtual world grows more complex, finding a standard way for storing data becomes increasingly important. Ideally, each data item would be brought into the computer system only once. References for data items need to be cryptographically verifiable, so the data can maintain its identity while being passed around. This way there will be only one copy of the users family photo album, while the user can use multiple tools to show or manipulate the album. Copies of users data could be stored on some of his family members computer, some of his computers, but also at some online services which he uses. When all actors operate over one replicated copy of the data, the system automatically avoids a single point of failure. Thus the data will not disappear with one computer breaking, or one service provider going out of business. One shared copy also makes it possible to delete a piece of data from all systems at once, on users request. In our research we tried to find a model that would make data manageable to users, and make it possible to have the same data stored at various locations. We studied three systems, Persona, Freenet, and GNUnet, that suggest different models for protecting user data. The main application areas of the systems studied include securing online social networks, providing anonymous web, and preventing censorship in file-sharing. Each of the systems studied store user data on machines belonging to third parties. The systems differ in measures they take to protect their users from data loss, forged information, censorship, and being monitored. All of the systems use cryptography to secure names used for the content, and to protect the data from outsiders. Based on the gained knowledge, we built a prototype platform called Peerscape, which stores user data in a synchronized, protected database. Data items themselves are protected with cryptography against forgery, but not encrypted as the focus has been disseminating the data directly among family and friends instead of letting third parties store the information. We turned the synchronizing database into peer-to-peer web by revealing its contents through an integrated http server. The REST-like http API supports development of applications in javascript. To evaluate the platform s suitability for application development we wrote some simple applications, including a public chat room, bittorrent site, and a flower growing game. During our early tests we came to the conclusion that using the platform for simple applications works well. As web standards develop further, writing applications for the platform should become easier. Any system this complex will have its problems, and we are not expecting our platform to replace the existing web, but are fairly impressed with the results and consider our work important from the perspective of managing user data.
Resumo:
When authors of scholarly articles decide where to submit their manuscripts for peer review and eventual publication, they often base their choice of journals on very incomplete information abouthow well the journals serve the authors’ purposes of informing about their research and advancing their academic careers. The purpose of this study was to develop and test a new method for benchmarking scientific journals, providing more information to prospective authors. The method estimates a number of journal parameters, including readership, scientific prestige, time from submission to publication, acceptance rate and service provided by the journal during the review and publication process. Data directly obtainable from the web, data that can be calculated from such data, data obtained from publishers and editors, and data obtained using surveys with authors are used in the method, which has been tested on three different sets of journals, each from a different discipline. We found a number of problems with the different data acquisition methods, which limit the extent to which the method can be used. Publishers and editors are reluctant to disclose important information they have at hand (i.e. journal circulation, web downloads, acceptance rate). The calculation of some important parameters (for instance average time from submission to publication, regional spread of authorship) can be done but requires quite a lot of work. It can be difficult to get reasonable response rates to surveys with authors. All in all we believe that the method we propose, taking a “service to authors” perspective as a basis for benchmarking scientific journals, is useful and can provide information that is valuable to prospective authors in selected scientific disciplines.
Resumo:
Service researchers and practitioners have repeatedly claimed that customer service experiences are essential to all businesses. Therefore comprehension of how service experience is characterised in research is an essential element for its further development through research. The importance of greater in-depth understanding of the phenomenon of service experience has been acknowledged by several researchers, such as Carú and Cova and Vargo and Lusch. Furthermore, Service-Dominant (S-D) logic has integrated service experience to value by emphasising in its foundational premises that value is phenomenologically (experientially) determined. The present study analyses how the concept of service experience has been characterised in previous research. As such, it puts forward three ways to characterise it in relation to that research: 1) phenomenological service experience relates to the value discussion in S-D logic and interpretative consumer research, 2) process-based service experience relates to understanding service as a process, and 3) outcome-based service experience relates to understanding service experience as one element in models linking a number of variables or attributes to various outcomes. Focusing on the phenomenological service experience, the theoretical purpose of the study is to characterise service experience based on the phenomenological approach. In order to do so, an additional methodological purpose was formulated: to find a suitable methodology for analysing service experience based on the phenomenological approach. The study relates phenomenology to a philosophical Husserlian and social constructionist tradition studying phenomena as they appear in our experience in a social context. The study introduces Event-Based Narrative Inquiry Technique (EBNIT), which combines critical events with narratives and metaphors. EBNIT enabled the analysis of lived and imaginary service experiences as expressed in individual narratives. The study presents findings of eight case studies within service innovation of Web 2.0, mobile service, location aware service and public service in the municipal sector. Customers’ and service managers’ stories about their lived private and working lifeworld were the foundation for their ideal service experiences. In general, the thesis finds that service experiences are (1) subjective, (2) context-specific, (3) cumulative, (4) partially socially constructed, (5) both lived and imaginary, (6) temporally multiple-dimensional, and (7) iteratively related to perceived value. In addition to customer service experience, the thesis brings empirical evidence of managerial service experience of front-line managers experiencing the service they manage and develop in their working lifeworld. The study contributes to S-D logic, service innovation and service marketing and management in general by characterising service experience based on the phenomenological approach and integrating it to the value discussion. Additionally, the study offers a methodological approach for further exploration of service experiences. The study discusses managerial implications in conjunction with the case studies and discusses them in relation to service innovation.
Resumo:
Activity systems are the cognitively linked groups of activities that consumers carry out as a part of their daily life. The aim of this paper is to investigate how consumers experience value through their activities, and how services fit into the context of activity systems. A new technique for illustrating consumers’ activity systems is introduced. The technique consists of identifying a consumer’s activities through an interview, then quantitatively measuring how the consumer evaluates the identified activities on three dimensions: Experienced benefits, sacrifices and frequency. This information is used to create a graphical representation of the consumer’s activity system, an “activityscape map”. Activity systems work as infrastructure for the individual consumer’s value experience. The paper contributes to value and service literature, where there currently are no clearly described standardized techniques for visually mapping out individual consumer activity. Existing approaches are service- or relationship focused, and are mostly used to identify activities, not to understand them. The activityscape representation provides an overview of consumers’ perceptions of their activity patterns and the position of one or several services in this pattern. Comparing different consumers’ activityscapes, it shows the differences between consumers' activity structures, and provides insight into how services are used to create value within them. The paper is conceptual; an empirical illustration is used to indicate the potential in further empirical studies. The technique can be used by businesses to understand contexts for service use, which may uncover potential for business reconfiguration and customer segmentation.
Resumo:
Activity systems are the cognitively linked groups of activities that consumers carry out as a part of their daily life. The aim of this paper is to investigate how consumers experience value through their activities, and how services fit into the context of activity systems. A new technique for illustrating consumers’ activity systems is introduced. The technique consists of identifying a consumer’s activities through an interview, then quantitatively measuring how the consumer evaluates the identified activities on three dimensions: Experienced benefits, sacrifices and frequency. This information is used to create a graphical representation of the consumer’s activity system, an “activityscape map”. Activity systems work as infrastructure for the individual consumer’s value experience. The paper contributes to value and service literature, where there currently are no clearly described standardized techniques for visually mapping out individual consumer activity. Existing approaches are service- or relationship focused, and are mostly used to identify activities, not to understand them. The activityscape representation provides an overview of consumers’ perceptions of their activity patterns and the position of one or several services in this pattern. Comparing different consumers’ activityscapes, it shows the differences between consumers' activity structures, and provides insight into how services are used to create value within them. The paper is conceptual; an empirical illustration is used to indicate the potential in further empirical studies. The technique can be used by businesses to understand contexts for service use, which may uncover potential for business reconfiguration and customer segmentation.
Resumo:
Participation is located in a living and complex environment. Traditional means of participation are only partially able to meet the new environmental requirements. In need are forms of participation which take into account the new opportunities of the environment and residents expertise. Internet map applications are an important channel of participation which potential is in many respects as unexplored and unutilized. They are commonly in inventory the perspectives, bringing out the concerns of the area, and only little for discussing about solutions. Interpretation is usually made by designer. This study focuses on evaluation and development of Internet map applications in strategic land use planning. Subject matter is dealt from designer and the inhabitants point of view. City Planning Department of Helsinki s Esikau-punkien Renessanssi -project and the associated SoftGIS survey acts as the case study. In the beginning of the study I tried to recognize the new environment in which the Internet map applications are placed. The research question is, what kind of challenges and opportunities the e-participation confronts in information society, and what kind of requirements the environmental creates for development of an application. In chapter three I evaluate how successfully these requirements are met in Esikau-punkien Renessanssi -project. I m trying to examine how the application would look like if the environment and the characteristics of the project are met better. The approach is experimental and I try to find new ways to take advantage of Internet maps. I try not to be too limited to current projects and studies. For example, I try to examine how social media and Web 2.0 opportunities can be utilized, and how the learning and shaping nature of planning may be reached in Internet map environment. In chapter four I have developed a new concept for the Esikaupunkien Renessanssi map application, and made images to visualize its operation in practice. I collect all the data in the research and gather it into a new service. The aim is to create a better application for Esikaupunkien Renessanssi -project, which takes into account the living and complex environment of participation and responds to threats and opportunities arising from it. The presented outcome is in many respects different from the current query. In the new service the role of residents is to interact and learn. The traditional standing of the Internet maps and the position of resident as one-sided information donor are questioned. In the concept, the residents innovate and make interpretations too. Influences are taken from a number of modern applications and for example services that make use of social media. The user experience is intended to be interactive, fast and easy. The idea is that the service keeps you up to date with planning matters, not the other way around. Service guides inhabitants, striving to achieve a deeper knowledge of the project's objectives as well as the dynamics and realities that different individuals experience.