December 2021 Marginal singularity and the benefits of labels in covariate-shift
Samory Kpotufe, Guillaume Martinet
Author Affiliations +
Ann. Statist. 49(6): 3299-3323 (December 2021). DOI: 10.1214/21-AOS2084

Abstract

Transfer Learning addresses common situations in Machine Leaning where little or no labeled data is available for a target prediction problem—corresponding to a distribution Q, but much labeled data is available from some related but different data distribution P. This work is concerned with the fundamental limits of transfer, that is, the limits in target performance in terms of (1) sample sizes from P and Q, and (2) differences in data distributions P, Q. In particular, we aim to address practical questions such as how much target data from Q is sufficient given a certain amount of related data from P, and how to optimally sample such target data for labeling.

We present new minimax results for transfer in nonparametric classification (i.e., for situations where little is known about the target classifier), under the common assumption that the marginal distributions of covariates differ between P and Q (often termed covariate-shift). Our results are first to concisely capture the relative benefits of source and target labeled data in these settings through information-theoretic limits. Namely, we show that the benefits of target labels are tightly controlled by a transfer-exponent γ that encodes how singular Q is locally with respect to P, and interestingly paints a more favorable picture of transfer than what might be believed from insights from previous work. In fact, while previous work rely largely on refinements of traditional metrics and divergences between distributions, and often only yield a coarse view of when transfer is possible or not, our analysis—in terms of γ—reveals a continuum of new regimes ranging from easy to hard transfer.

We then address the practical question of how to efficiently sample target data to label, by showing that a recently proposed semi-supervised procedure—based on k-NN classification, can be refined to adapt to unknown γ and, therefore, requests target labels only when beneficial, while achieving nearly minimax-optimal transfer rates without knowledge of distributional parameters. Of independent interest, we obtain new minimax-optimality results for vanilla k-NN classification in regimes with nonuniform marginals.

Acknowledgments

Authors are listed in alphabetic order. Samory Kpotufe was at Princeton University, ORFE, and a visiting member at the Institute for Advanced Study, Princeton, for a major portion of the project.

Acknowledgments

Authors are listed in alphabetic order. Samory Kpotufe was at Princeton University, ORFE, and a visiting member at the Institute for Advanced Study, Princeton, for a major portion of the project.

Citation

Download Citation

Samory Kpotufe. Guillaume Martinet. "Marginal singularity and the benefits of labels in covariate-shift." Ann. Statist. 49 (6) 3299 - 3323, December 2021. https://doi.org/10.1214/21-AOS2084

Information

Received: 1 May 2019; Revised: 1 December 2020; Published: December 2021
First available in Project Euclid: 14 December 2021

MathSciNet: MR4352531
zbMATH: 1486.62186
Digital Object Identifier: 10.1214/21-AOS2084

Subjects:
Primary: 62B10 , 62D99 , 62G99 , 68Q32 , 68T05

Keywords: covariate shift , domain adaptation , Nonparametric classification , transfer learning

Rights: Copyright © 2021 Institute of Mathematical Statistics

JOURNAL ARTICLE
25 PAGES

This article is only available to subscribers.
It is not available for individual sale.
+ SAVE TO MY LIBRARY

Vol.49 • No. 6 • December 2021
Back to Top