Abstract
We address the limitations of Gaussian processes for multiclass classification in the setting where both the number of classes and the number of observations is very large. We propose a scalable approximate inference framework by combining the inducing points method with variational approximations of the likelihood that have been recently proposed in the literature. This leads to a tractable lower bound on the marginal likelihood that decomposes into a sum over both data points and class labels, and hence, is amenable to doubly stochastic optimization. To overcome memory issues when dealing with large datasets, we resort to amortized inference, which coupled with subsampling over classes reduces the computational and the memory footprint without a significant loss in performance. We demonstrate empirically that the proposed algorithm leads to superior performance in terms of test accuracy, and improved detection of tail labels.
Original language | English |
---|---|
Title of host publication | Proceedings of 2020 IEEE 30th International Workshop on Machine Learning for Signal Processing |
Number of pages | 6 |
Publisher | IEEE |
Publication date | 2020 |
ISBN (Electronic) | 978-1-7281-6662-9 |
DOIs | |
Publication status | Published - 2020 |
Event | 2020 IEEE 30th International Workshop on Machine Learning for Signal Processing - Aalto University, Espoo, Finland Duration: 21 Sept 2020 → 24 Sept 2020 Conference number: 30 https://ieeexplore.ieee.org/xpl/conhome/9217888/proceeding |
Conference
Conference | 2020 IEEE 30th International Workshop on Machine Learning for Signal Processing |
---|---|
Number | 30 |
Location | Aalto University |
Country/Territory | Finland |
City | Espoo |
Period | 21/09/2020 → 24/09/2020 |
Internet address |
Series | Machine Learning for Signal Processing |
---|---|
ISSN | 1551-2541 |
Keywords
- Gaussian process classification
- Variational inference
- Augmented model