Nilesh Tripuraneni

EECS Department, University of California, Berkeley

Technical Report No. UCB/EECS-2022-182

August 5, 2022

http://www2.eecs.berkeley.edu/Pubs/TechRpts/2022/EECS-2022-182.pdf

Classically, most machine learning (ML) methodology has made an innocuous modeling assumption: data drawn from both the training/test sets has been independently sampled from a pair of identical distributions with nice properties. Yet, in the situations modern ML methods must confront, deviations from this idealized setting are quickly becoming the norm–not the exception. In this thesis, we address the challenges arising in understanding the often unexpected phenomenology in these settings by developing theory in two areas of interest: transfer learning and robust learning. In particular, we focus on identifying what structural conditions/techniques are needed to permit sample-efficient learning in these new settings, in order to answer questions such as why pretraining is so effective and what the limits of learning are for extremely heavy-tailed distributions.

Advisors: Michael Jordan


BibTeX citation:

@phdthesis{Tripuraneni:EECS-2022-182,
    Author= {Tripuraneni, Nilesh},
    Title= {Learning Beyond the Standard Model (of Data)},
    School= {EECS Department, University of California, Berkeley},
    Year= {2022},
    Month= {Aug},
    Url= {http://www2.eecs.berkeley.edu/Pubs/TechRpts/2022/EECS-2022-182.html},
    Number= {UCB/EECS-2022-182},
    Abstract= {Classically, most machine learning (ML) methodology has made an innocuous modeling assumption: data drawn from both the training/test sets has been independently sampled from a pair of identical distributions with nice properties. Yet, in the situations modern ML methods must confront, deviations from this idealized setting are quickly becoming the norm–not the exception. In this thesis, we address the challenges arising in understanding the often unexpected phenomenology in these settings by developing theory in two areas of interest: transfer learning and robust learning. In particular, we focus on identifying what structural conditions/techniques are needed to permit sample-efficient learning in these new settings, in order to answer questions such as why pretraining is so effective and what the limits of learning are for extremely heavy-tailed distributions.},
}

EndNote citation:

%0 Thesis
%A Tripuraneni, Nilesh 
%T Learning Beyond the Standard Model (of Data)
%I EECS Department, University of California, Berkeley
%D 2022
%8 August 5
%@ UCB/EECS-2022-182
%U http://www2.eecs.berkeley.edu/Pubs/TechRpts/2022/EECS-2022-182.html
%F Tripuraneni:EECS-2022-182