arXiv Analytics

Sign in

arXiv:2212.05735 [cs.LG]AbstractReferencesReviewsResources

Adaptive Low-Precision Training for Embeddings in Click-Through Rate Prediction

Shiwei Li, Huifeng Guo, Lu Hou, Wei Zhang, Xing Tang, Ruiming Tang, Rui Zhang, Ruixuan Li

Published 2022-12-12Version 1

Embedding tables are usually huge in click-through rate (CTR) prediction models. To train and deploy the CTR models efficiently and economically, it is necessary to compress their embedding tables at the training stage. To this end, we formulate a novel quantization training paradigm to compress the embeddings from the training stage, termed low-precision training (LPT). Also, we provide theoretical analysis on its convergence. The results show that stochastic weight quantization has a faster convergence rate and a smaller convergence error than deterministic weight quantization in LPT. Further, to reduce the accuracy degradation, we propose adaptive low-precision training (ALPT) that learns the step size (i.e., the quantization resolution) through gradient descent. Experiments on two real-world datasets confirm our analysis and show that ALPT can significantly improve the prediction accuracy, especially at extremely low bit widths. For the first time in CTR models, we successfully train 8-bit embeddings without sacrificing prediction accuracy. The code of ALPT is publicly available.

Related articles: Most relevant | Search more
arXiv:1905.09433 [cs.LG] (Published 2019-05-23)
FiBiNET: Combining Feature Importance and Bilinear feature Interaction for Click-Through Rate Prediction
arXiv:2001.03025 [cs.LG] (Published 2020-01-08)
Deep Time-Stream Framework for Click-Through Rate Prediction by Tracking Interest Evolution
arXiv:2007.03519 [cs.LG] (Published 2020-07-06)
GateNet: Gating-Enhanced Deep Network for Click-Through Rate Prediction