posted on 2002-01-01, 00:00authored byBrad Myers, Robert Malkin, Michael Bett, Alex Waibel, Ben Bostwick, Robert C. Miller, Jie Yang, Matthias Denecke, Edgar Seemann, JIe Zhu, Choon Hong Peck, Dave Kong, Jeffrey Nichols, Bill Scherlis
We describe our system which facilitates collaboration
using multiple modalities, including speech, handwriting,
gestures, gaze tracking, direct manipulation, large projected
touch-sensitive displays, laser pointer tracking,
regular monitors with a mouse and keyboard, and wirelessly-
networked handhelds. Our system allows multiple,
geographically dispersed participants to simultaneously
and flexibly mix different modalities using the right interface
at the right time on one or more machines. This paper
discusses each of the modalities provided, how they
were integrated in the system architecture, and how the
user interface enabled one or more people to flexibly use
one or more devices.