Number of co-authors:9
Number of publications with 3 favourite co-authors:Loring Holden:Michael V. Capps:Howard Abrams:
Timothy Miller's 3 most productive colleagues in number of publications:Joseph J. LaViola:29Robert C. Zeleznik:13Robert Zeleznik:12
go to course
The Ultimate Guide to Visual Perception and Design
go to course
User-Centred Design - Module 3
Starts tomorrow LAST CALL!
Marc Hassenzahl explains the fascinating concept of User Experience and Experience Design. Commentaries by Don Norman, Eric Reiss, Mark Blythe, and Whitney Hess
User Experience and Experience Design !
Our Latest Books
The Social Design of Technical Systems: Building technologies for communities. 2nd Edition
by Brian Whitworth and Adnan Ahmad
Gamification at Work: Designing Engaging Business Software
by Janaki Mythily Kumar and Mario Herger
The Social Design of Technical Systems: Building technologies for communities
by Brian Whitworth and Adnan Ahmad
The Encyclopedia of Human-Computer Interaction, 2nd Ed.
by Mads Soegaard and Rikke Friis Dam
Publications by Timothy Miller (bibliography)
Bragdon, Andrew, Zeleznik, Robert, Williamson, Brian, Miller, Timothy and LaViola, Joseph J. (2009): GestureBar: improving the approachability of gesture-based interfaces. In: Proceedings of ACM CHI 2009 Conference on Human Factors in Computing Systems 2009. pp. 2269-2278. Available online
GestureBar is a novel, approachable UI for learning gestural interactions that enables a walk-up-and-use experience which is in the same class as standard menu and toolbar interfaces. GestureBar leverages the familiar, clean look of a common toolbar, but in place of executing commands, richly discloses how to execute commands with gestures, through animated images, detail tips and an out-of-document practice area. GestureBar's simple design is also general enough for use with any recognition technique and for integration with standard, non-gestural UI components. We evaluate GestureBar in a formal experiment showing that users can perform complex, ecologically valid tasks in a purely gestural system without training, introduction, or prior gesture experience when using GestureBar, discovering and learning a high percentage of the gestures needed to perform the tasks optimally, and significantly outperforming a state of the art crib sheet. The relative contribution of the major design elements of GestureBar is also explored. A second experiment shows that GestureBar is preferred to a basic crib sheet and two enhanced crib sheet variations.
© All rights reserved Bragdon et al. and/or ACM Press
Zeleznik, Robert and Miller, Timothy (2006): Fluid inking: augmenting the medium of free-form inking with gestures. In: Proceedings of the 2006 Conference on Graphics Interface 2006. pp. 155-162. Available online
We present Fluid Inking, a generally applicable approach to augmenting the fluid medium of free-form inking with gestural commands. Our approach is characterized by four design criteria, including: 1) pen-based hardware impartiality: all interactions can be performed with a button-free stylus, the minimal input hardware requirement for inking, and the least common denominator device for pen-based systems ranging from PDAs to whiteboards; 2) performability: gestures use short sequences of simple and familiar inking interactions that require minimal targeting; 3) extensibility: gestures are a regular pattern of optional shortcuts for commands in an arbitrarily scalable menu system; and 4) discoverability: gesture shortcuts (analogous to modifier keys) are displayed in the interactive menu and are suggested with dynamic feedback during inking. This paper presents the Fluid Inking techniques in the unified context of a prototype notetaking application and emphasizes how post-fix terminal punctuation and prefix flicks can disambiguate gestures from regular inking. We also discuss how user feedback influenced the Fluid Inking design.
© All rights reserved Zeleznik and Miller and/or Canadian Information Processing Society
Zeleznik, Robert, Miller, Timothy and Forsberg, Andrew (2001): Pop through mouse button interactions. In: Marks, Joe and Mynatt, Elizabeth D. (eds.) Proceedings of the 14th annual ACM symposium on User interface software and technology November 11 - 14, 2001, Orlando, Florida. pp. 195-196. Available online
We present a range of novel interactions enabled by a simple modification in
the design of a computer mouse. By converting each mouse button to pop through
tactile push-buttons, similar to the focus/shutter-release buttons used in many
cameras, users can feel, and the computer can sense, two distinct "clicks"
corresponding to pressing lightly and pressing firmly to pop through. Despite
the prototypical status of our hardware and software implementations, our
current pop through mouse interactions are compelling and warrant further
investigation. In particular, we demonstrate that pop through buttons not only
yield an additional button activation state that is composable with, or even
preferable to, techniques such as double-clicking, but also can endow a
qualitatively novel user experience when meaningfully and consistently applied.
We propose a number of software guidelines that may provide a consistent,
systemic benefit; for example, light pressure may invoke default interaction
(short menu), and firm pressure may supply more detail (long menu).
© All rights reserved Zeleznik et al. and/or ACM Press
Zeleznik, Robert C., Holden, Loring, Capps, Michael V., Abrams, Howard and Miller, Timothy (2000): Collaboration between Heterogeneous Stand-alone 3-D Graphical Applications. In Comput. Graph. Forum, 19 (3) .
Miller, Timothy and Zeleznik, Robert C. (1999): The design of 3D haptic widgets. In: SI3D 1999 1999. pp. 97-102. Available online
Miller, Timothy and Zeleznik, Robert (1998): An Insidious Haptic Invasion: Adding Force Feedback to the X Desktop. In: Mynatt, Elizabeth D. and Jacob, Robert J. K. (eds.) Proceedings of the 11th annual ACM symposium on User interface software and technology November 01 - 04, 1998, San Francisco, California, United States. pp. 59-64. Available online
This paper describes preliminary work in a project to add force feedback to user interface elements of the X Window System in an attempt to add true "feel" to the window system's "look and feel". Additions include adding ridges around icons and menu items to aid interaction, alignment guides for moving windows, and other enhancements to window manipulation. The motivation for this system is the observation that people naturally have many skills for and intuitions about a very rich environment of interaction forces in the non-computer world; however, these skills are largely unused in computer applications. We expect that haptic modifications to conventional graphical user interfaces, such as those we present, can lead to gains in performance, intuition, learnability, and enjoyment of the interface. This paper describes details of the implementation of the haptic window system elements, in addition to higher-level haptic design principles and informal observations of users of the system.
© All rights reserved Miller and Zeleznik and/or ACM Press
Join our community and advance:
Page maintainer: The Editorial Team