912 resultados para Natural user interfaces
Resumo:
Modelling business processes for analysis or redesign usually requires the collaboration of many stakeholders. These stakeholders may be spread across locations or even companies, making co-located collaboration costly and difficult to organize. Modern process modelling technologies support remote collaboration but lack support for visual cues used in co-located collaboration. Previously we presented a prototype 3D virtual world process modelling tool that supports a number of visual cues to facilitate remote collaborative process model creation and validation. However, the added complexity of having to navigate a virtual environment and using an avatar for communication made the tool difficult to use for novice users. We now present an evolved version of the technology that addresses these issues by providing natural user interfaces for non-verbal communication, navigation and model manipulation.
Resumo:
A demo video showing the BPMVM prototype using several natural user interfaces, such as multi-touch input, full-body tracking and virtual reality.
Resumo:
Natural User Interfaces (NUI) offer rich ways for interacting with the digital world that make innovative use of existing human capabilities. They include and often combine different input modalities such as voice, gesture, eye gaze, body interactions, touch and touchless interactions. However much of the focus of NUI research and development has been on enhancing the experience of individuals interacting with technology. Effective NUIs must also acknowledge our innately social characteristics, and support how we communicate with each other, play together, learn together and collaboratively work together. This workshop concerns the social aspects of NUI. The workshop seeks to better understand the social uses and applications of these new NUI technologies -- how we design these technologies for new social practices and how we understand the use of these technologies in key social contexts.
Resumo:
The area of Human-Machine Interface is growing fast due to its high importance in all technological systems. The basic idea behind designing human-machine interfaces is to enrich the communication with the technology in a natural and easy way. Gesture interfaces are a good example of transparent interfaces. Such interfaces must identify properly the action the user wants to perform, so the proper gesture recognition is of the highest importance. However, most of the systems based on gesture recognition use complex methods requiring high-resource devices. In this work, we propose to model gestures capturing their temporal properties, which significantly reduce storage requirements, and use clustering techniques, namely self-organizing maps and unsupervised genetic algorithm, for their classification. We further propose to train a certain number of algorithms with different parameters and combine their decision using majority voting in order to decrease the false positive rate. The main advantage of the approach is its simplicity, which enables the implementation using devices with limited resources, and therefore low cost. The testing results demonstrate its high potential.
Resumo:
This paper examines the issues surrounding the successful design and development of tangible technology for optimal engagement in playful activities. At present there is very little data on how, and in what contexts, tangible interactions with technology promote lasting engagement and immersion. The framework at the core of this paper has been designed to guide the effective design of tangible technology for immersive interaction. The paper investigates the relationship between tangible user interfaces (TUI) characteristics of representation and control, and immersive flow experiences produced through balancing skill and challenge in user interaction.
Resumo:
FEniCS is a collection of software tools for the automated solution of differential equations by finite element methods. In this note, we describe how FEniCS can be used to solve a simple nonlinear model problem with varying levels of automation. At one extreme, FEniCS provides tools for the fully automated and adaptive solution of nonlinear partial differential equations. At the other extreme, FEniCS provides a range of tools that allow the computational scientist to experiment with novel solution algorithms. © 2010 American Institute of Physics.
Resumo:
Older people often find it difficult to learn to use new technology. Although they may want to adopt it, they can find the learning process challenging and frustrating and subsequently lose motivation. This paper looks at how psychological theories of intrinsic motivation could be applied to make the ICT learning process more engaging for older users and describes an experiment set up to test the applicability of these theories to user interface (UI) design. The results of the experiment confirmed that intrinsic motivation theory is a valid lens through which to look at current ICT design and also uncovered significant gender differences in reaction to different kinds of learning tasks. © 2013 Springer-Verlag Berlin Heidelberg.
Resumo:
Interactive intention understanding is important for Pen-based User Interface (PUI). Many works on this topic are reported, and focus on handwriting or sketching recognition algorithms at the lexical layer. But these algorithms cannot totally solve the problem of intention understanding and can not provide the pen-based software with high usability. Hence, a scenario-based interactive intention understanding framework is presented in this paper, and is used to simulate human cognitive mechanisms and cognitive habits. By providing the understanding environment supporting the framework, we can apply the framework to the practical PUI system. The evaluation of the Scientific Training Management System for the Chinese National Diving Team shows that the framework is effective in improving the usability and enhancing the intention understanding capacity of this system.
Resumo:
We present a new technique called‘Tilt Menu’ for better extending selection capabilities of pen-based interfaces.The Tilt Menu is implemented by using 3D orientation information of pen devices while performing selection tasks.The Tilt Menu has the potential to aid traditional onehanded techniques as it simultaneously generates the secondary input (e.g., a command or parameter selection) while drawing/interacting with a pen tip without having to use the second hand or another device. We conduct two experiments to explore the performance of the Tilt Menu. In the first experiment, we analyze the effect of parameters of the Tilt Menu, such as the menu size and orientation of the item, on its usability. Results of the first experiment suggest some design guidelines for the Tilt Menu. In the second experiment, the Tilt Menu is compared to two types of techniques while performing connect-the-dot tasks using freeform drawing mechanism. Results of the second experiment show that the Tilt Menu perform better in comparison to the Tool Palette, and is as good as the Toolglass.