National Gallery of Art InceptionV3 Features

On October 24-25, 2019 the National Gallery of Art in Washington, D.C.hosted a "data-thon" where multiple teams of art historians and data scientists worked with the museum's open collection data to study questions about the history and composition of the collections.

A joint team from Carnegie Mellon and the University of Pittsburgh used image features from a convolutional neural network to index the National Gallery of Art's images by visual similarity. This allowed the team to compare the visual distribution of different collections within the National Gallery, and with related parts of the Samuel H. Kress collection (distributed in museums around the country) as well as a portion of the Lessig Rosenwald collection (split between the National Gallery and the Library of Congress).

This deposit includes:

- a table of Inception V3 image features computed for images from the National Gallery of Art
- a set of JSON files containing basic curatorial data for each image, along with a list of its 7 nearest neighbors in the Inception V3 vector space
- 2D visualizations of paintings, and of prints and drawings, clustered based on these computed features.
- Colorized visualizations showing the distribution of individual collectors' contributions in this visual similarity space
- Slides presented at the National Gallery

An interface for browsing these artworks by visual neighbors is available at