Skip to main content

Community Repository Search Results

resource research Media and Technology
Finger-based touch input has become a major interaction modality for mobile user interfaces. However, due to the low precision of finger input, small user interface components are often difficult to acquire and operate on a mobile device. It is even harder when the user is on the go and unable to pay close attention to the interface. In this paper, we present Gesture Avatar, a novel interaction technique that allows users to operate existing arbitrary user interfaces using gestures. It leverages the visibility of graphical user interfaces and the casual interaction of gestures. Gesture Avatar
DATE:
TEAM MEMBERS: Hao Lü Yang Li
resource research Media and Technology
Collaborative Information Retrieval (CIR) is the process by which people working together can collaboratively search for, share and navigate through information. Computer support for CIR currently makes use of single-user systems. CIR systems could benefit from the use of multi-user interaction to enable more than one person to collaborate using the same data sources, at the same time and in the same place. Multi-touch interaction has provided the ability for multiple users to interact simultaneously with a multi-touch surface. This paper presents a generalised architecture for multi-touch CIR
DATE:
TEAM MEMBERS: Ivan Sams Janet Wesson Dieter Vogts
resource research Media and Technology
This paper explores the interactive possibilities enabled when the barrel of a digital pen is augmented with a multi-touch sensor. We present a novel multi-touch pen (MTPen) prototype and discuss its alternate uses beyond those of a standard stylus, such as allowing new touch gestures to be performed using the index finger or thumb and detecting how users grip the device as a mechanism for mode switch-ing. We also discuss the hardware and software implemen-tation challenges in realizing our prototype, and showcase how one can combine different grips (tripod, relaxed tripod, sketch, wrap) and
DATE:
TEAM MEMBERS: Jim Spadaccini Hyunyoung Song Hrvoje Benko Francois Guimbretiere Shahram Izadi Xiang Cao Ken Hinckley
resource research
In this paper we describe two projects that utilize reality-based interaction to advance collaborative scientific inquiry and discovery. We discuss the relation between reality-based and embodied interaction, and present findings from an experimental study that illustrate benefits of reality-based tabletop interaction for collaborative inquiry-based learning.
DATE:
TEAM MEMBERS: Orit Shaer
resource research Media and Technology
New mobile devices with large multi-touch displays, such as the iPad, have brought revolutionary changes to ways users interact with computers. Instead of traditional input devices such as keyboards, touchpads and mice, multi-touch gestures are used as the primary means of interacting with mobile devices. Surprisingly, body-motion gestures are evolving to become a new, natural, and effective way for game players to interact with game consoles in a very similar fashion: in Kinect for Xbox 360, a controller-free gaming experience is made possible by using body-motion gestures to play games.
DATE:
TEAM MEMBERS: Yuan Feng Zimu Liu Baochun Li
resource research Media and Technology
This paper outlines research showing a suprizing agreement in the guesability of multitouch gestures on tabletop surfaces between users. It also provides more evidence that crowd sourcing gesture mapping will lead to more complete intuitive gesture set and potential convergence into a standard gesture library.
DATE:
TEAM MEMBERS: Jacob Wobbrock Meredith Moris Andrew Wilson
resource evaluation
Direct-touch interaction on mobile phones revolves around screens that compete for visual attention with users‟ real-world tasks and activities. This paper investigates the impact of these situational impairments on touch-screen interaction. We probe several design factors for touch-screen gestures, under various levels of environmental demands on attention, in comparison to the status-quo approach of soft buttons. We find that in the presence of environmental distractions, ges-tures can offer significant performance gains and reduced attentional load, while performing as well as soft buttons
DATE:
TEAM MEMBERS: Andrew Bragdon Eugene Nelson Yang Li Ken Hinckley
resource evaluation
Recent advances in touch screen technology have increased the prevalence of touch screens and have prompted a wave of new touch screen-based devices. However, touch screens are still largely inaccessible to blind users, who must adopt error-prone compensatory strategies to use them or find accessible alternatives. This inaccessibility is due to interaction techniques that require the user to visually locate objects on the screen. To address this problem, we introduce Slide Rule, a set of audiobased multi-touch interaction techniques that enable blind users to access touch screen applications
DATE:
TEAM MEMBERS: Jim Spadaccini Jeffrey Bigham Jacob Wobbrock
resource research Professional Development, Conferences, and Networks
For document visualization, folding techniques provide a focus-plus-context approach with fairly high legibility on flat sections. To enable richer interaction, we explore the design space of multi-touch document folding. We discuss several design considerations for simple modeless gesturing and compatibility with standard Drag and Pinch gestures. We categorize gesture models along the characteristics of Symmetric/Asymmetric and Serial/Parallel, which yields three gesture models. We built a prototype document workspace application that integrates folding and standard gestures, and a system for
DATE:
TEAM MEMBERS: Patrick Chiu Chunyuan Liao Francine Chen
resource research Media and Technology
Delimiters are great for using gestures to overide application or OS commands. This paper investigates whether the DoubleFlip gesture is easy learn and practical to use as an effective delimter.
DATE:
TEAM MEMBERS: Jaime Ruiz Yang Li
resource research
Many tasks in graphical user interfaces require users to interact with elements at various levels of precision. We present FingerGlass, a bimanual technique designed to improve the precision of graphical tasks on multitouch screens. It enables users to quickly navigate to different locations and across multiple scales of a scene using a single hand. The other hand can simultaneously interact with objects in the scene. Unlike traditional pan-zoom interfaces, FingerGlass retains contextual information during the interaction. We evaluated our technique in the context of precise object selection
DATE:
TEAM MEMBERS: Dominik K¨aser Maneesh Agrawala Mark Pauly
resource research Media and Technology
Modern smartphones contain sophisticated sensors to monitor three-dimensional movement of the device. These sensors permit devices to recognize motion gestures— deliberate movements of the device by end-users to invoke commands. However, little is known about best-practices in motion gesture design for the mobile computing paradigm. To address this issue, we present the results of a guessability study that elicits end-user motion gestures to invoke commands on a smartphone device. We demonstrate that consensus exists among our participants on parameters of movement and on mappings of motion
DATE:
TEAM MEMBERS: Jim Spadaccini Jaime Ruiz Yang Li Edward Lank