Training Logistic Regression and SVM on 200GB Data Using b-Bit Minwise Hashing and Comparisons with Vowpal Wabbit (VW)
Dept. of Statistical Science, Cornell University, Ithaca, NY 14853
arXiv:1108.3072v1 [cs.LG] (15 Aug 2011)
@article{2011arXiv1108.3072L,
author={Li}, P. and {Shrivastava}, A. and {Konig}, C.},
title={"{Training Logistic Regression and SVM on 200GB Data Using b-Bit Minwise Hashing and Comparisons with Vowpal Wabbit (VW)}"},
journal={ArXiv e-prints},
archivePrefix={"arXiv"},
eprint={1108.3072},
primaryClass={"cs.LG"},
keywords={Computer Science – Learning, Statistics – Methodology, Statistics – Machine Learning},
year={2011},
month={aug},
adsurl={http://adsabs.harvard.edu/abs/2011arXiv1108.3072L},
adsnote={Provided by the SAO/NASA Astrophysics Data System}
}
We generated a dataset of 200 GB with 10^9 features, to test our recent b-bit minwise hashing algorithms for training very large-scale logistic regression and SVM. The results confirm our prior work that, compared with the VW hashing algorithm (which has the same variance as random projections), b-bit minwise hashing is substantially more accurate at the same storage. For example, with merely 30 hashed values per data point, b-bit minwise hashing can achieve similar accuracies as VW with 2^14 hashed values per data point.
August 16, 2011 by hgpu