Scaling Up Neural Datasets: A public fMRI dataset of 5000 scenes - Robotics Institute Carnegie Mellon University

Scaling Up Neural Datasets: A public fMRI dataset of 5000 scenes

Nai Chen Chang, Elissa M. Aminoff, John A. Pyles, Michael J. Tarr, and Abhinav Gupta
Journal Article, Journal of Vision, Vol. 18, No. 10, September, 2018

Abstract

Vision science - and particularly machine vision - is being revolutionized by large-scale datasets. State-of-the-art artificial vision models critically depend on large-scale datasets to achieve high performance. In contrast, although large-scale learning models (e.g., deep learning models such as Alexnet) have been applied to human neuroimaging data, the image datasets used on neural studies often rely on significantly fewer images - typically a few hundred due to time-constrained experimental procedures. The small size of these datasets also translates to a limited diversity of used images covered in image space. The lack of image feature diversity inherently limits the degree to which neural data can act as supervisory signals and our ability to compare model and measured neural representations. Here we dramatically increase the image dataset size deployed in an fMRI study of visual scene processing, scaling the number of images by over an order of magnitude relative to most earlier studies: over 5,000 discrete image stimuli were presented to each of four participants. We believe this boost in dataset size will better connect the field of computer vision to human neuroscience. To further enhance this connection and increase image space overlap with computer vision datasets, we include images from two standard artificial learning datasets in our stimuli: 2,000 images from COCO; 2 images per category from ImageNet (~2,000). Also included are 1,000 hand-curated indoor and outdoor scene images from 250 categories. These three image collections cover a wide variety of image types, thereby enabling fine-grained exploration into visual representations ranging from natural scenes to object categories to human interactions. The scale advantage of our dataset and the use of a slow event-related design enables, for the first time, joint computer vision and fMRI analyses that span a significant and diverse region of image space using high-performing models.

Notes
Meeting abstract presented at 18th Annual VSS '18

BibTeX

@article{Chang-2018-121570,
author = {Nai Chen Chang and Elissa M. Aminoff and John A. Pyles and Michael J. Tarr and Abhinav Gupta},
title = {Scaling Up Neural Datasets: A public fMRI dataset of 5000 scenes},
journal = {Journal of Vision},
year = {2018},
month = {September},
volume = {18},
number = {10},
}