Abstract
In the past decade, Dropout has emerged as a powerful and simple method for training neural networks preventing co-adaptation by stochastically omitting neurons. Dropout is currently not grounded in explicit modelling assumptions which so far has precluded its adoption in Bayesian modeling. Using Bayesian entropic reasoning we show that dropout can be interpreted as optimal inference under constraints. We demonstrate this on an analytically tractable regression model providing a Bayesian interpretation of its mechanism for regularizing and preventing co-adaptation as well as its connection to other Bayesian techniques, and in our experiments we find that dropout can provide robustness under model misspecification. Our framework roots dropout as a theoretically justified and practical tool for statistical modeling allowing Bayesian practitioners to tap into the benefits of dropout training.
Original language | English |
---|---|
Journal | Procedia Computer Science |
Volume | 201 |
Issue number | C |
Pages (from-to) | 771-776 |
ISSN | 1877-0509 |
DOIs | |
Publication status | Published - 2022 |
Event | 3rd International Workshop on Statistical Methods and Artificial Intelligence - Porto, Portugal Duration: 22 Mar 2022 → 25 Mar 2022 |
Conference
Conference | 3rd International Workshop on Statistical Methods and Artificial Intelligence |
---|---|
Country/Territory | Portugal |
City | Porto |
Period | 22/03/2022 → 25/03/2022 |
Keywords
- Bayesian learning
- Dropout
- Maximum entropy