New mobile devices with large multi-touch displays, such as the iPad, have brought revolutionary changes to ways users interact with computers. Instead of traditional input devices such as keyboards, touchpads and mice, multi-touch gestures are used as the primary means of interacting with mobile devices. Surprisingly, body-motion gestures are evolving to become a new, natural, and effective way for game players to interact with game consoles in a very similar fashion: in Kinect for Xbox 360, a controller-free gaming experience is made possible by using body-motion gestures to play games.
This paper outlines research showing a suprizing agreement in the guesability of multitouch gestures on tabletop surfaces between users. It also provides more evidence that crowd sourcing gesture mapping will lead to more complete intuitive gesture set and potential convergence into a standard gesture library.
DATE:
TEAM MEMBERS:
Jacob WobbrockMeredith MorisAndrew Wilson
Most tabletop research presents findings from lab-based user studies, focusing on specific interaction techniques. This means we still know little about how these new interfaces perform in real life settings and how users appropriate them. This paper presents findings from a field study of an existing interactive table in a museum of natural history. Visitors were found to employ a wide variety of gestures for interacting; different interface elements invited different types of gesture. The analysis highlights challenges and design conflicts in the design of tabletop interfaces for public
Classic article from the Journal of Visitor Behavior (1994) which discusses different approaches to exhibit design. The author cites considerable historical research, including one of the earliest visitor studies from 1935 about how visitors engage with exhibits. Very thorough analysis and critique of quantitative and qualitative evaluation techniques and when to apply them during exhibit design. Useful for exhibit design teams and anyone involved with designing exhibits for museums and galleries.
Direct-touch interaction on mobile phones revolves around screens that compete for visual attention with users‟ real-world tasks and activities. This paper investigates the impact of these situational impairments on touch-screen interaction. We probe several design factors for touch-screen gestures, under various levels of environmental demands on attention, in comparison to the status-quo approach of soft buttons. We find that in the presence of environmental distractions, ges-tures can offer significant performance gains and reduced attentional load, while performing as well as soft buttons
DATE:
TEAM MEMBERS:
Andrew BragdonEugene NelsonYang LiKen Hinckley
Zooming user interfaces are increasingly popular on mobile devices with touch screens. Swiping and pinching finger gestures anywhere on the screen manipulate the displayed portion of a page, and taps open objects within the page. This makes navigation easy but limits other manipulations of objects that would be supported naturally by the same gestures, notably cut and paste, multiple selection, and drag and drop. A popular device that suffers from this limitation is Apple’s iPhone. In this paper, we present Bezel Swipe, an interaction technique that supports multiple selection, cut, copy
In recent years, a large amount of software for multitouch interfaces with various degrees of similarity has been written. In order to improve interoperability, we aim to identify the common traits of these systems and present a layered software architecture which abstracts these similarities by defining common interfaces between successive layers. This provides developers with a unified view of the various types of multitouch hardware. Moreover, the layered architecture allows easy integration of existing software, as several alternative implementations for each layer can co-exist. Finally
DATE:
TEAM MEMBERS:
Florian EchtlerGudrun KlinkerJim Spadaccini
Recent advances in touch screen technology have increased the prevalence of touch screens and have prompted a wave of new touch screen-based devices. However, touch screens are still largely inaccessible to blind users, who must adopt error-prone compensatory strategies to use them or find accessible alternatives. This inaccessibility is due to interaction techniques that require the user to visually locate objects on the screen. To address this problem, we introduce Slide Rule, a set of audiobased multi-touch interaction techniques that enable blind users to access touch screen applications
DATE:
TEAM MEMBERS:
Jim SpadacciniJeffrey BighamJacob Wobbrock
Most current multi-touch capable interactive user interfaces for tabletop are built from custom toolkits that are decoupled from, and on top of, the “Desktop” provided by the underlying Operating System. However, this approach requires that each individual touch system build their own suite of touch capable custom applications (such as photo browsers), usually resulting in limited functionality. In this paper, we propose a software architecture for supporting and integrating multi-touch capability on existing desktop systems, where multi-touch and multiple single pointer input can be used
DATE:
TEAM MEMBERS:
Kelvin ChengBenjamin ItzsteinPaul SztajerMarkus Rittenbruch
Respected museum professional and consultant Kathleen McLean examines the shift towards a more participatory culture of display and exhibition in museums. She cites numerous examples from different types of museums and the approaches they have taken in trying to better understand their visitors and become more visitor-centered. She lays open the debate within the museum community about the traditional role of museums and the resistance that advocates of visitor research and evaluation face from more conservative thinking colleagues. Useful background reading for interactive exhibit designers.
It is increasingly common for software and hardware systems to support touch-based interaction. While the technology to support this interaction is still evolving, common protocols for providing consistent communication between hardware and software are available. However, this is not true for gesture recognition – the act of translating a series of strokes or touches into a system recognizable event. Developers often end up writing code for this process from scratch due to the lack of higher-level frameworks for defining new gestures. Gesture recognition can contain a significant amount of
Despite the considerable quantity of research directed towards multitouch technologies, a set of standardized UI components have not been developed. Menu systems provide a particular challenge, as traditional GUI menus require a level of pointing precision inappropriate for direct finger input. Marking menus are a promising alternative, but have yet to be investigated or adapted for use within multitouch systems. In this paper, we first investigate the human capabilities for performing directional chording gestures, to assess the feasibility of multitouch marking menus. Based on the positive
DATE:
TEAM MEMBERS:
Julian LepinskiTovi GrossmanGeorge Fitzmaurice