960 resultados para User studies
Resumo:
In this thesis, different techniques for image analysis of high density microarrays have been investigated. Most of the existing image analysis techniques require prior knowledge of image specific parameters and direct user intervention for microarray image quantification. The objective of this research work was to develop of a fully automated image analysis method capable of accurately quantifying the intensity information from high density microarrays images. The method should be robust against noise and contaminations that commonly occur in different stages of microarray development.
Resumo:
Self-adaptive software provides a profound solution for adapting applications to changing contexts in dynamic and heterogeneous environments. Having emerged from Autonomic Computing, it incorporates fully autonomous decision making based on predefined structural and behavioural models. The most common approach for architectural runtime adaptation is the MAPE-K adaptation loop implementing an external adaptation manager without manual user control. However, it has turned out that adaptation behaviour lacks acceptance if it does not correspond to a user’s expectations – particularly for Ubiquitous Computing scenarios with user interaction. Adaptations can be irritating and distracting if they are not appropriate for a certain situation. In general, uncertainty during development and at run-time causes problems with users being outside the adaptation loop. In a literature study, we analyse publications about self-adaptive software research. The results show a discrepancy between the motivated application domains, the maturity of examples, and the quality of evaluations on the one hand and the provided solutions on the other hand. Only few publications analysed the impact of their work on the user, but many employ user-oriented examples for motivation and demonstration. To incorporate the user within the adaptation loop and to deal with uncertainty, our proposed solutions enable user participation for interactive selfadaptive software while at the same time maintaining the benefits of intelligent autonomous behaviour. We define three dimensions of user participation, namely temporal, behavioural, and structural user participation. This dissertation contributes solutions for user participation in the temporal and behavioural dimension. The temporal dimension addresses the moment of adaptation which is classically determined by the self-adaptive system. We provide mechanisms allowing users to influence or to define the moment of adaptation. With our solution, users can have full control over the moment of adaptation or the self-adaptive software considers the user’s situation more appropriately. The behavioural dimension addresses the actual adaptation logic and the resulting run-time behaviour. Application behaviour is established during development and does not necessarily match the run-time expectations. Our contributions are three distinct solutions which allow users to make changes to the application’s runtime behaviour: dynamic utility functions, fuzzy-based reasoning, and learning-based reasoning. The foundation of our work is a notification and feedback solution that improves intelligibility and controllability of self-adaptive applications by implementing a bi-directional communication between self-adaptive software and the user. The different mechanisms from the temporal and behavioural participation dimension require the notification and feedback solution to inform users on adaptation actions and to provide a mechanism to influence adaptations. Case studies show the feasibility of the developed solutions. Moreover, an extensive user study with 62 participants was conducted to evaluate the impact of notifications before and after adaptations. Although the study revealed that there is no preference for a particular notification design, participants clearly appreciated intelligibility and controllability over autonomous adaptations.
Resumo:
Eye tracking has become a preponderant technique in the evaluation of user interaction and behaviour with study objects in defined contexts. Common eye tracking related data representation techniques offer valuable input regarding user interaction and eye gaze behaviour, namely through fixations and saccades measurement. However, these and other techniques may be insufficient for the representation of acquired data in specific studies, namely because of the complexity of the study object being analysed. This paper intends to contribute with a summary of data representation and information visualization techniques used in data analysis within different contexts (advertising, websites, television news and video games). Additionally, several methodological approaches are presented in this paper, which resulted from several studies developed and under development at CETAC.MEDIA - Communication Sciences and Technologies Research Centre. In the studies described, traditional data representation techniques were insufficient. As a result, new approaches were necessary and therefore, new forms of representing data, based on common techniques were developed with the objective of improving communication and information strategies. In each of these studies, a brief summary of the contribution to their respective area will be presented, as well as the data representation techniques used and some of the acquired results.
Resumo:
The primary objective of this study was to document the benefits and possible detriments of combining ipsilateral acoustic hearing in the cochlear implant ear of a patient with preserved low frequency residual hearing post cochlear implantation. The secondary aim was to examine the efficacy of various cochlear implant mapping and hearing aid fitting strategies in relation to electro-acoustic benefits.
Resumo:
This paper discusses a study to determine the effectiveness of the Hearing Aid Performance Inventory (HAPI) on hearing aid outcomes.
Resumo:
This paper describes the user modeling component of EPIAIM, a consultation system for data analysis in epidemiology. The component is aimed at representing knowledge of concepts in the domain, so that their explanations can be adapted to user needs. The first part of the paper describes two studies aimed at analysing user requirements. The first one is a questionnaire study which examines the respondents' familiarity with concepts. The second one is an analysis of concept descriptions in textbooks and from expert epidemiologists, which examines how discourse strategies are tailored to the level of experience of the expected audience. The second part of the paper describes how the results of these studies have been used to design the user modeling component of EPIAIM. This module works in a two-step approach. In the first step, a few trigger questions allow the activation of a stereotype that includes a "body" and an "inference component". The body is the representation of the body of knowledge that a class of users is expected to know, along with the probability that the knowledge is known. In the inference component, the learning process of concepts is represented as a belief network. Hence, in the second step the belief network is used to refine the initial default information in the stereotype's body. This is done by asking a few questions on those concepts where it is uncertain whether or not they are known to the user, and propagating this new evidence to revise the whole situation. The system has been implemented on a workstation under UNIX. An example of functioning is presented, and advantages and limitations of the approach are discussed.
Resumo:
The paper is an investigation of the exchange of ideas and information between an architect and building users in the early stages of the building design process before the design brief or any drawings have been produced. The purpose of the research is to gain insight into the type of information users exchange with architects in early design conversations and to better understand the influence the format of design interactions and interactional behaviours have on the exchange of information. We report an empirical study of pre-briefing conversations in which the overwhelming majority of the exchanges were about the functional or structural attributes of space, discussion that touched on the phenomenological, perceptual and the symbolic meanings of space were rare. We explore the contextual features of meetings and the conversational strategies taken by the architect to prompt the users for information and the influence these had on the information provided. Recommendations are made on the format and structure of pre-briefing conversations and on designers' strategies for raising the level of information provided by the user beyond the functional or structural attributes of space.
Resumo:
A manufactured aeration and nanofiltration MBR greywater system was tested during continuous operation at the University of Reading, to demonstrate reliability in delivery of high quality treated greywater. Its treatment performance was evaluated against British Standard criteria [BSI (Greywater Systems—Part 1 Code of Practice: BS8525-1:2010. BS Press, 2010); (Greywater Systems—Part 2 Domestic Greywater Treatment, Requirements and Methods: BS 8525-2:2011. BS Press, 2011)]. The low carbon greywater recycling technology produced excellent analytical results as well as consistency in performance. User acceptance of such reliably treated greywater was then evaluated through user perception studies. The results inform the potential supply of treated greywater to student accommodation. Out of 135 questionnaire replies, 95% demonstrated a lack of aversion in one or more attributes, to using treated, recycled greywater.
Resumo:
The primary objective of this research study is to determine which form of testing, the PEST algorithm or an operator-controlled condition is most accurate and time efficient for administration of the gaze stabilization test
Resumo:
BACKGROUND: Shared decision-making (SDM) is an emergent research topic in the field of mental health care and is considered to be a central component of a recovery-oriented system. Despite the evidence suggesting the benefits of this change in the power relationship between users and practitioners, the method has not been widely implemented in clinical practice. OBJECTIVE: The objective of this study was to investigate decisional and information needs among users with mental illness as a prerequisite for the development of a decision support tool aimed at supporting SDM in community-based mental health services in Sweden. METHODS: Three semi-structured focus group interviews were conducted with 22 adult users with mental illness. The transcribed interviews were analyzed using a directed content analysis. This method was used to develop an in-depth understanding of the decisional process as well as to validate and conceptually extend Elwyn et al.'s model of SDM. RESULTS: The model Elwyn et al. have created for SDM in somatic care fits well for mental health services, both in terms of process and content. However, the results also suggest an extension of the model because decisions related to mental illness are often complex and involve a number of life domains. Issues related to social context and individual recovery point to the need for a preparation phase focused on establishing cooperation and mutual understanding as well as a clear follow-up phase that allows for feedback and adjustments to the decision-making process. CONCLUSIONS AND IMPLICATIONS FOR PRACTICE: The current study contributes to a deeper understanding of decisional and information needs among users of community-based mental health services that may reduce barriers to participation in decision-making. The results also shed light on attitudinal, relationship-based, and cognitive factors that are important to consider in adapting SDM in the mental health system.
Resumo:
The use of wireless local area networks, called WLANs, as well as the proliferation of the use of multimedia applications have grown rapidly in recent years. Some factors affect the quality of service (QoS) received by the user and interference is one of them. This work presents strategies for planning and performance evaluation through an empirical study of the QoS parameters of a voice over Internet Protocol (VoIP) application in an interference network, as well as the relevance in the design of wireless networks to determine the coverage area of an access point, taking into account several parameters such as power, jitter, packet loss, delay, and PMOS. Another strategy is based on a hybrid approach that considers measuring and Bayesian inference applied to wireless networks, taking into consideration QoS parameters. The models take into account a cross layer vision of networks, correlating aspects of the physical environment, on the signal propagation (power or distance) with aspects of VoIP applications (e.g., jitter and packet loss). Case studies were carried out for two indoor environments and two outdoor environments, one of them displaying main characteristics of the Amazon region (e.g., densely arboreous environments). This last test bed was carried out in a real system because the Government of the State of Pará has a digital inclusion program called NAVEGAPARÁ.
Resumo:
Not long ago, most software was written by professional programmers, who could be presumed to have an interest in software engineering methodologies and in tools and techniques for improving software dependability. Today, however, a great deal of software is written not by professionals but by end-users, who create applications such as multimedia simulations, dynamic web pages, and spreadsheets. Applications such as these are often used to guide important decisions or aid in important tasks, and it is important that they be sufficiently dependable, but evidence shows that they frequently are not. For example, studies have shown that a large percentage of the spreadsheets created by end-users contain faults, and stories abound of spreadsheet faults that have led to multi-million dollar losses. Despite such evidence, until recently, relatively little research had been done to help end-users create more dependable software.
Resumo:
Not long ago, most software was written by professional programmers, who could be presumed to have an interest in software engineering methodologies and in tools and techniques for improving software dependability. Today, however, a great deal of software is written not by professionals but by end-users, who create applications such as multimedia simulations, dynamic web pages, and spreadsheets. Applications such as these are often used to guide important decisions or aid in important tasks, and it is important that they be sufficiently dependable, but evidence shows that they frequently are not. For example, studies have shown that a large percentage of the spreadsheets created by end-users contain faults. Despite such evidence, until recently, relatively little research had been done to help end-users create more dependable software. We have been working to address this problem by finding ways to provide at least some of the benefits of formal software engineering techniques to end-user programmers. In this talk, focusing on the spreadsheet application paradigm, I present several of our approaches, focusing on methodologies that utilize source-code-analysis techniques to help end-users build more dependable spreadsheets. Behind the scenes, our methodologies use static analyses such as dataflow analysis and slicing, together with dynamic analyses such as execution monitoring, to support user tasks such as validation and fault localization. I show how, to accommodate the user base of spreadsheet languages, an interface to these methodologies can be provided in a manner that does not require an understanding of the theory behind the analyses, yet supports the interactive, incremental process by which spreadsheets are created. Finally, I present empirical results gathered in the use of our methodologies that highlight several costs and benefits trade-offs, and many opportunities for future work.
Resumo:
The ability to utilize information systems (IS) effectively is becoming a necessity for business professionals. However, individuals differ in their abilities to use IS effectively, with some achieving exceptional performance in IS use and others being unable to do so. Therefore, developing a set of skills and attributes to achieve IS user competency, or the ability to realize the fullest potential and the greatest performance from IS use, is important. Various constructs have been identified in the literature to describe IS users with regard to their intentions to use IS and their frequency of IS usage, but studies to describe the relevant characteristics associated with highly competent IS users, or those who have achieved IS user competency, are lacking. This research develops a model of IS user competency by using the Repertory Grid Technique to identify a broad set of characteristics of highly competent IS users. A qualitative analysis was carried out to identify categories and sub-categories of these characteristics. Then, based on the findings, a subset of the model of IS user competency focusing on the IS-specific factors – domain knowledge of and skills in IS, willingness to try and to explore IS, and perception of IS value – was developed and validated using the survey approach. The survey findings suggest that all three factors are relevant and important to IS user competency, with willingness to try and to explore IS being the most significant factor. This research generates a rich set of factors explaining IS user competency, such as perception of IS value. The results not only highlight characteristics that can be fostered in IS users to improve their performance with IS use, but also present research opportunities for IS training and potential hiring criteria for IS users in organizations.
Resumo:
Volatile amines are prominent indicators of food freshness, as they are produced during many microbiological food degradation processes. Monitoring and indicating the volatile amine concentration within the food package by intelligent packaging solutions might therefore be a simple yet powerful way to control food safety throughout the distribution chain.rnrnIn this context, this work aims to the formation of colourimetric amine sensing surfaces on different substrates, especially transparent PET packaging foil. The colour change of the deposited layers should ideally be discernible by the human eye to facilitate the determination by the end-user. rnrnDifferent tailored zinc(II) and chromium(III) metalloporphyrins have been used as chromophores for the colourimetric detection of volatile amines. A new concept to increase the porphyrins absorbance change upon exposure to amines is introduced. Moreover, the novel porphyrins’ processability during the deposition process is increased by their enhanced solubility in non-polar solvents.rnrnThe porphyrin chromophores have successfully been incorporated into polysiloxane matrices on different substrates via a dielectric barrier discharge enhanced chemical vapour deposition. This process allows the use of nitrogen as a cheap and abundant plasma gas, produces minor amounts of waste and by-products and can be easily introduced into (existing) roll-to-roll production lines. The formed hybrid sensing layers tightly incorporate the porphyrins and moreover form a porous structure to facilitate the amines diffusion to and interaction with the chromophores.rnrnThe work is completed with the thorough analysis of the porphyrins’ amine sensing performance in solution as well as in the hybrid coatings . To reveal the underlying interaction mechanisms, the experimental results are supported by DFT calculations. The deposited layers could be used for the detection of NEt3 concentrations below 10 ppm in the gas phase. Moreover, the coated foils have been tested in preliminary food storage experiments. rnrnThe mechanistic investigations on the interaction of amines with chromium(III) porphyrins revealed a novel pathway to the formation of chromium(IV) oxido porphyrins. This has been used for electrochemical epoxidation reactions with dioxygen as the formal terminal oxidant.rn