Loading [MathJax]/jax/output/CommonHTML/jax.js

Hamiltonian Monte Carlo on ReLU Neural Networks is Inefficient

Part of Advances in Neural Information Processing Systems 37 (NeurIPS 2024) Main Conference Track

Bibtex Paper

Authors

Vu C. Dinh, Lam Si Tung Ho, Cuong V. Nguyen

Abstract

We analyze the error rates of the Hamiltonian Monte Carlo algorithm with leapfrog integrator for Bayesian neural network inference. We show that due to the non-differentiability of activation functions in the ReLU family, leapfrog HMC for networks with these activation functions has a large local error rate of Ω(ϵ) rather than the classical error rate of O(ϵ3). This leads to a higher rejection rate of the proposals, making the method inefficient. We then verify our theoretical findings through empirical simulations as well as experiments on a real-world dataset that highlight the inefficiency of HMC inference on ReLU-based neural networks compared to analytical networks.