Places205-VGGNet Models for Scene Recognition
Shenzhen Institutes of Advanced Technology, CAS, China
arXiv:1508.01667 [cs.CV], (7 Aug 2015)
@article{wang2015placesvggnet,
title={Places205-VGGNet Models for Scene Recognition},
author={Wang, Limin and Guo, Sheng and Huang, Weilin and Qiao, Yu},
year={2015},
month={aug},
archivePrefix={arXiv},
primaryClass={cs.CV}
}
VGGNets have turned out to be effective for object recognition in still images. However, it is unable to yield good performance by directly adapting the VGGNet models trained on the ImageNet dataset for scene recognition. This report describes our implementation of training the VGGNets on the large-scale Places205 dataset. Specifically, we train three VGGNet models, namely VGGNet-11, VGGNet-13, and VGGNet-16, by using a Multi-GPU extension of Caffe toolbox with high computational efficiency. We verify the performance of trained Places205-VGGNet models on three datasets: MIT67, SUN397, and Places205. Our trained models achieve the state-of-the-art performance on these datasets and are made public available.
August 10, 2015 by hgpu