Carnegie Mellon University
Browse

Tactile Gestures for Human/Robot Interaction

Download (255.68 kB)
journal contribution
posted on 1995-01-01, 00:00 authored by Richard M. Voyles, Pradeep Khosla

Gesture-Based Programming is a new paradigm to ease the burden of programming robots. By tapping in to the user’s wealth of experience with contact transitions, compliance, uncertainty and operations sequencing, we hope to provide a more intuitive programming environment for complex, real-world tasks based on the expressiveness of non-verbal communication. A requirement for this to be accomplished is the ability to interpret gestures to infer the intentions behind them. As a first step toward this goal, this paper presents an application of distributed perception for inferring a user’s intentions by observing tactile gestures. These gestures consist of sparse, inexact, physical “nudges” applied to the robot’s end effector for the purpose of modifying its trajectory in free space. A set of independent agents - each with its own local, fuzzified, heuristic model of a particular trajectory parameter - observes data from a wrist force/torque sensor to evaluate the gestures. The agents then independently determine the confidence of their respective findings and distributed arbitration resolves the interpretation through voting.

History

Date

1995-01-01

Usage metrics

    Exports

    RefWorks
    BibTeX
    Ref. manager
    Endnote
    DataCite
    NLM
    DC