AIMS Mathematics, 2019, 4(3): 359-383. doi: 10.3934/math.2019.3.359

Research article

Export file:

Format

  • RIS(for EndNote,Reference Manager,ProCite)
  • BibTex
  • Text

Content

  • Citation Only
  • Citation and Abstract

Relative entropy minimization over Hilbert spaces via Robbins-Monro

1 Department of Mathematics, Drexel University, Philadelphia, PA 19104, USA
2 Department of Ocean, Earth, and Atmospheric Sciences, Oregon State University, Corvallis, OR 97331, USA

One way of getting insight into non-Gaussian measures is to first obtain good Gaussian approximations. These best fit Gaussians can then provide a sense of the mean and variance of the distribution of interest. They can also be used to accelerate sampling algorithms. This begs the question of how one should measure optimality, and how to then obtain this optimal approximation. Here, we consider the problem of minimizing the distance between a family of Gaussians and the target measure with respect to relative entropy, or Kullback-Leibler divergence. As we are interested in applications in the infinite dimensional setting, it is desirable to have convergent algorithms that are well posed on abstract Hilbert spaces. We examine this minimization problem by seeking roots of the first variation of relative entropy, taken with respect to the mean of the Gaussian, leaving the covariance fixed. We prove the convergence of Robbins-Monro type root finding algorithms in this context, highlighting the assumptions necessary for convergence to relative entropy minimizers. Numerical examples are included to illustrate the algorithms.
  Figure/Table
  Supplementary
  Article Metrics

References

1.A. E. Albert and L. A. Gardner Jr., Stochastic approximation and nonlinear regression The MIT Press, 1967.

2.C. Andrieu and É. Moulines, On the ergodicity properties of some adaptive MCMC algorithms Ann. Appl. Probab. 16 (2006), 1462-1505.    

3.C. Andrieu and J. Thoms, A tutorial on adaptive MCMC Stat. Comput. 18 (2008), 343-373.    

4.D. M. Blei, A. Kucukelbir and J. D. McAuliffe, Variational Inference: A Review for Statisticians J. Am. Stat. Assoc. 112 (2017), 859-877.    

5.J. R. Blum, Approximation methods which converge with probability one Annals of Mathematical Statistics 25 (1954), 382-386.    

6.J. R. Blum, Multidimensional stochastic approximation methods Annals of Mathematical Statistics 25 (1954), 737-744.    

7.S. D. Chatterji, Martingale convergence and the Radon-Nikodym theorem in Banach spaces Math. Scand. 22 (1968), 21-41.    

8.H.-F. Chen, Stochastic approximation and its applications 2002.

9.G. Da Prato, An Introduction to Infinite-Dimensional Analsysis Springer, 2006.

10.H. Haario, E. Saksman and J. Tamminen, An adaptive Metropolis algorithm Bernoulli 7 (2001), 223-242.    

11.H. J. Kushner and G. Yin, Stochastic Approximation and Recursive Algorithms and Applications Springer, 2003.

12.J. Lelong, Almost sure convergence of randomly truncated stochastic algorithms under verifiable conditions Stat. Probabil. Lett. 78 (2008), 2632-2636.    

13.Y. Lu, A. Stuart and H. Weber, Gaussian approximations for probability measures on r^d SIAM/ASA Journal on Uncertainty Quantification 5 (2017), 1136-1165.    

14.Y. Lu, A. Stuart and H. Weber, Gaussian approximations for transition paths in brownian dynamics SIAM J. Math. Anal. 49 (2017), 3005-3047.    

15.F. J. Pinski, G Simpson, A. M. Stuart, et al. Algorithms for Kullback-Leibler Approximation of Probability Measures in Infinite Dimensions SIAM J. Sci. Comput. 37 (2015), A2733-A2757.    

16.F. J. Pinski, G Simpson, A. M. Stuart, et al. Kullback-Leibler Approximation for Probability Measures on Infinite Dimensional Spaces SIAM J. Math. Anal. 47 (2015), 4091-4122.    

17.H. Robbins and S. Monro, A stochastic approximation method The Annals of Mathematical Statistics (1950), 400-407.

18.G. O. Roberts and J. S. Rosenthal, Coupling and Ergodicity of Adaptive Markov Chain Monte Carlo Algorithms J. Appl. Probab. 44 (2007), 458-475.    

19.D. Sanz-Alonso and A. M. Stuart, Gaussian approximations of small noise diffusions in kullback-leibler divergence Commun. Math. Sci. 15 (2017), 2087-2097.    

20.G. Yin and Y. M. Zhu, On H-valued Robbins-Monro processes J. Multivariate Anal. 34 (1990), 116-140.    

© 2019 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution Licese (http://creativecommons.org/licenses/by/4.0)

Download full text in PDF

Export Citation

Article outline

Show full outline
Copyright © AIMS Press All Rights Reserved