Dave Epstein

EECS Department, University of California, Berkeley

Technical Report No. UCB/EECS-2024-65

May 8, 2024

http://www2.eecs.berkeley.edu/Pubs/TechRpts/2024/EECS-2024-65.pdf

Generative modeling promises an elegant solution to learning about high-dimensional data distributions such as images and videos — but how can we expose and utilize the rich structure these models discover? Rather than just drawing new samples, how can an agent actually harness p(x) as a source of knowledge about how our world works? This thesis explores scalable inductive biases that unlock a generative model's understanding of the entities latent in visual data, enabling much richer interaction with the model as a result.

Advisors: Alexei (Alyosha) Efros


BibTeX citation:

@phdthesis{Epstein:EECS-2024-65,
    Author= {Epstein, Dave},
    Title= {Disentangled Visual Generative Models},
    School= {EECS Department, University of California, Berkeley},
    Year= {2024},
    Month= {May},
    Url= {http://www2.eecs.berkeley.edu/Pubs/TechRpts/2024/EECS-2024-65.html},
    Number= {UCB/EECS-2024-65},
    Abstract= {Generative modeling promises an elegant solution to learning about high-dimensional data distributions such as images and videos — but how can we expose and utilize the rich structure these models discover? Rather than just drawing new samples, how can an agent actually harness p(x) as a source of knowledge about how our world works? This thesis explores scalable inductive biases that unlock a generative model's understanding of the entities latent in visual data, enabling much richer interaction with the model as a result.},
}

EndNote citation:

%0 Thesis
%A Epstein, Dave 
%T Disentangled Visual Generative Models
%I EECS Department, University of California, Berkeley
%D 2024
%8 May 8
%@ UCB/EECS-2024-65
%U http://www2.eecs.berkeley.edu/Pubs/TechRpts/2024/EECS-2024-65.html
%F Epstein:EECS-2024-65