Carnegie Mellon University
Browse

Distributed Representations of Geographically Situated Language

Download (313.69 kB)
journal contribution
posted on 2014-06-01, 00:00 authored by David Bamman, Chris Dyer, Noah A. Smith

We introduce a model for incorporating contextual information (such as geography) in learning vector-space representations of situated language. In contrast to approaches to multimodal representation learning that have used properties of the object being described (such as its color), our model includes information about the subject (i.e., the speaker), allowing us to learn the contours of a word’s meaning that are shaped by the context in which it is uttered. In a quantitative evaluation on the task of judging geographically informed semantic similarity between representations learned from 1.1 billion words of geo-located tweets, our joint model outperforms comparable independent models that learn meaning in isolation.

History

Publisher Statement

Copyright 2014 Association for Computational Linguistics

Date

2014-06-01

Usage metrics

    Exports

    RefWorks
    BibTeX
    Ref. manager
    Endnote
    DataCite
    NLM
    DC