Minimax Optimality of Score-based Diffusion Models: Beyond the Density Lower Bound Assumptions

Kaihong Zhang, Caitlyn H. Yin, Feng Liang, Jingbo Liu

Research output: Contribution to journalConference articlepeer-review

Abstract

We study the asymptotic error of score-based diffusion model sampling in large-sample scenarios from a non-parametric statistics perspective.We show that a kernel-based score estimator achieves an optimal mean square error of Õ (Equation presented) for the score function of p0 ∗ N(0, tId), where n and d represent the sample size and the dimension, t is bounded above and below by polynomials of n, and p0 is an arbitrary sub-Gaussian distribution.As a consequence, this yields an Õ (Equation presented) upper bound for the total variation error of the distribution of the sample generated by the diffusion model under a mere sub-Gaussian assumption.If in addition, p0 belongs to the nonparametric family of the β-Sobolev space with β ≤ 2, by adopting an early stopping strategy, we obtain that the diffusion model is nearly (up to log factors) minimax optimal.This removes the crucial lower bound assumption on p0 in previous proofs of the minimax optimality of the diffusion model for nonparametric families.

Original languageEnglish (US)
Pages (from-to)60134-60178
Number of pages45
JournalProceedings of Machine Learning Research
Volume235
StatePublished - 2024
Event41st International Conference on Machine Learning, ICML 2024 - Vienna, Austria
Duration: Jul 21 2024Jul 27 2024

ASJC Scopus subject areas

  • Artificial Intelligence
  • Software
  • Control and Systems Engineering
  • Statistics and Probability

Fingerprint

Dive into the research topics of 'Minimax Optimality of Score-based Diffusion Models: Beyond the Density Lower Bound Assumptions'. Together they form a unique fingerprint.

Cite this