Non-asymptotic Performance Guarantees for Neural Estimation of f-Divergences

Sreejith Sreekumar, Zhengxin Zhang, Ziv Goldfeld

Research output: Contribution to journalConference articlepeer-review

12 Scopus citations

Abstract

Statistical distances (SDs), which quantify the dissimilarity between probability distributions, are central to machine learning and statistics. A modern method for estimating such distances from data relies on parametrizing a variational form by a neural network (NN) and optimizing it. These estimators are abundantly used in practice, but corresponding performance guarantees are partial and call for further exploration. In particular, there seems to be a fundamental tradeoff between the two sources of error involved: approximation and estimation. While the former needs the NN class to be rich and expressive, the latter relies on controlling complexity. This paper explores this tradeoff by means of non-asymptotic error bounds, focusing on three popular choices of SDs-KullbackLeibler divergence, chi-squared divergence, and squared Hellinger distance. Our analysis relies on non-asymptotic function approximation theorems and tools from empirical process theory. Numerical results validating the theory are also provided.

Original languageEnglish
Pages (from-to)3322-3330
Number of pages9
JournalProceedings of Machine Learning Research
Volume130
StatePublished - 1 Jan 2021
Externally publishedYes
Event24th International Conference on Artificial Intelligence and Statistics, AISTATS 2021 - Virtual, Online, United States
Duration: 13 Apr 202115 Apr 2021

ASJC Scopus subject areas

  • Artificial Intelligence
  • Software
  • Control and Systems Engineering
  • Statistics and Probability

Fingerprint

Dive into the research topics of 'Non-asymptotic Performance Guarantees for Neural Estimation of f-Divergences'. Together they form a unique fingerprint.

Cite this