Some recent insights on transfer-learning
Samory Kpotufe (Columbia)
Abstract: A common situation in Machine Learning is one where training data is not fully representative of a target population due to bias in the sampling mechanism or high costs in sampling the target population; in such situations, we aim to ’transfer’ relevant information from the training data (a.k.a. source data) to the target application. How much information is in the source data? How much target data should we collect if any? These are all practical questions that depend crucially on ‘how far’ the source domain is from the target. However, how to properly measure ‘distance’ between source and target domains remains largely unclear.
In this talk we will argue that much of the traditional notions of ‘distance’ (e.g. KL-divergence, extensions of TV such as D_A discrepancy, density-ratios, Wasserstein distance) can yield an over-pessimistic picture of transferability. Instead, we show that some new notions of ‘relative dimension’ between source and target (which we simply term ‘transfer-exponents’) capture a continuum from easy to hard transfer. Transfer-exponents uncover a rich set of situations where transfer is possible even at fast rates, encode relative benefits of source and target samples, and have interesting implications for related problems such as multi-task or multi-source learning.
In particular, in the case of multi-source learning, we will discuss (if time permits) a strong dichotomy between minimax and adaptive rates: no adaptive procedure can achieve a rate better than single source rates, although minimax (oracle) procedures can.
The talk is based on earlier work with Guillaume Martinet, and ongoing work with Steve Hanneke.
optimization and controlstatistics theory
Audience: researchers in the topic
Series comments: Description: Research seminar on data science
See here for Zoom links to individual seminars, links to recordings, and to subscribe to calendar or mailing list.
| Organizers: | Afonso S. Bandeira*, Joan Bruna, Carlos Fernandez-Granda, Jonathan Niles-Weed, Ilias Zadik |
| *contact for this listing |
