December 2022 Likelihood estimation of sparse topic distributions in topic models and its applications to Wasserstein document distance calculations
Xin Bing, Florentina Bunea, Seth Strimas-Mackey, Marten Wegkamp
Author Affiliations +
Ann. Statist. 50(6): 3307-3333 (December 2022). DOI: 10.1214/22-AOS2229

Abstract

This paper studies the estimation of high-dimensional, discrete, possibly sparse, mixture models in the context of topic models. The data consists of observed multinomial counts of p words across n independent documents. In topic models, the p×n expected word frequency matrix is assumed to be factorized as a p×K word-topic matrix A and a K×n topic-document matrix T. Since columns of both matrices represent conditional probabilities belonging to probability simplices, columns of A are viewed as p-dimensional mixture components that are common to all documents while columns of T are viewed as the K-dimensional mixture weights that are document specific and are allowed to be sparse.

The main interest is to provide sharp, finite sample, 1-norm convergence rates for estimators of the mixture weights T when A is either known or unknown. For known A, we suggest MLE estimation of T. Our nonstandard analysis of the MLE not only establishes its 1 convergence rate, but also reveals a remarkable property: the MLE, with no extra regularization, can be exactly sparse and contain the true zero pattern of T. We further show that the MLE is both minimax optimal and adaptive to the unknown sparsity in a large class of sparse topic distributions. When A is unknown, we estimate T by optimizing the likelihood function corresponding to a plug in, generic, estimator Aˆ of A. For any estimator Aˆ that satisfies carefully detailed conditions for proximity to A, we show that the resulting estimator of T retains the properties established for the MLE. Our theoretical results allow the ambient dimensions K and p to grow with the sample sizes.

Our main application is to the estimation of 1-Wasserstein distances between document generating distributions. We propose, estimate and analyze new 1-Wasserstein distances between alternative probabilistic document representations, at the word and topic level, respectively. We derive finite sample bounds on the estimated proposed 1-Wasserstein distances. For word level document-distances, we provide contrast with existing rates on the 1-Wasserstein distance between standard empirical frequency estimates. The effectiveness of the proposed 1-Wasserstein distances is illustrated by an analysis of an IMDB movie reviews data set. Finally, our theoretical results are supported by extensive simulation studies.

Funding Statement

Bunea is supported in part by NSF Grant DMS-2015195 and DMS-2210563. Wegkamp is supported in part by NSF Grants DMS-2015195 and DMS-2210557.

Acknowledgments

We thank the Editor, the Associate Editor and two referees for their detailed reviews, which helped to improve the paper substantially.

Citation

Download Citation

Xin Bing. Florentina Bunea. Seth Strimas-Mackey. Marten Wegkamp. "Likelihood estimation of sparse topic distributions in topic models and its applications to Wasserstein document distance calculations." Ann. Statist. 50 (6) 3307 - 3333, December 2022. https://doi.org/10.1214/22-AOS2229

Information

Received: 1 July 2021; Revised: 1 September 2022; Published: December 2022
First available in Project Euclid: 21 December 2022

MathSciNet: MR4524498
zbMATH: 07641127
Digital Object Identifier: 10.1214/22-AOS2229

Subjects:
Primary: 62H12 , 62H30
Secondary: 62F10

Keywords: adaptive estimation , anchor words , high-dimensional estimation , maximum likelihood estimation , minimax estimation , mixture model , multinomial distribution , nonnegative matrix factorization , Sparse estimation , topic models

Rights: Copyright © 2022 Institute of Mathematical Statistics

JOURNAL ARTICLE
27 PAGES

This article is only available to subscribers.
It is not available for individual sale.
+ SAVE TO MY LIBRARY

Vol.50 • No. 6 • December 2022
Back to Top