Large-Scale Deep Learning on the YFCC100M Dataset
Lawrence Livernmore National Laboratory, Computational Engineering Division, 7000 East Avenue, Livermore, CA 94550
arXiv:1502.03409 [cs.LG], (11 Feb 2015)
@article{ni2015largescale,
title={Large-Scale Deep Learning on the YFCC100M Dataset},
author={Ni, Karl and Pearce, Roger and Boakye, Kofi and Essen, Brian Van and Borth, Damian and Chen, Barry and Wang, Eric},
year={2015},
month={feb},
archivePrefix={"arXiv"},
primaryClass={cs.LG}
}
We present a work-in-progress snapshot of learning with a 15 billion parameter deep learning network on HPC architectures applied to the largest publicly available natural image and video dataset released to-date. Recent advancements in unsupervised deep neural networks suggest that scaling up such networks in both model and training dataset size can yield significant improvements in the learning of concepts at the highest layers. We train our three-layer deep neural network on the Yahoo! Flickr Creative Commons 100M dataset. The dataset comprises approximately 99.2 million images and 800,000 user-created videos from Yahoo’s Flickr image and video sharing platform. Training of our network takes eight days on 98 GPU nodes at the High Performance Computing Center at Lawrence Livermore National Laboratory. Encouraging preliminary results and future research directions are presented and discussed.
February 13, 2015 by hgpu