Learning Low-Dimensional Structure via Closed-Loop Transcription: Equilibria and Optimization
Druv Pai
EECS Department, University of California, Berkeley
Technical Report No. UCB/EECS-2023-74
May 9, 2023
http://www2.eecs.berkeley.edu/Pubs/TechRpts/2023/EECS-2023-74.pdf
We consider the problem of learning maximally informative representations for data in a high-dimensional space with distribution supported on or around a single or multiple low-dimensional geometric structures, with or without labels. That is, we wish to compute a linear injective map (i.e., an “encoder”) such that the image of the data (i.e., the “representations”) have maximum “information gain”; we also want to compute a a suitable notion of inverse for the encoder (i.e., a “decoder”). We formulate this family of learning problems as a class of two-player games. For a broad notion of game-theoretic equilibria which is learnable via standard gradient-based optimization techniques, we show that the equilibrium solutions to games within the class indeed result in maximally informative representations and a consistent autoencoding. We then apply this framework to several instances of the closed-loop transcription (CTRL) framework, which has been recently proposed for learning discriminative and generative representations for data lying on low- dimensional submanifolds, obtaining desirable representations which provably emulate and extend ones given by classical theory. Finally, we present a novel optimization algorithm to obtain the particular equilibria which our theory desires, and prove its correctness in a restricted case.
Advisors: Yi Ma
BibTeX citation:
@mastersthesis{Pai:EECS-2023-74, Author= {Pai, Druv}, Title= {Learning Low-Dimensional Structure via Closed-Loop Transcription: Equilibria and Optimization}, School= {EECS Department, University of California, Berkeley}, Year= {2023}, Month= {May}, Url= {http://www2.eecs.berkeley.edu/Pubs/TechRpts/2023/EECS-2023-74.html}, Number= {UCB/EECS-2023-74}, Abstract= {We consider the problem of learning maximally informative representations for data in a high-dimensional space with distribution supported on or around a single or multiple low-dimensional geometric structures, with or without labels. That is, we wish to compute a linear injective map (i.e., an “encoder”) such that the image of the data (i.e., the “representations”) have maximum “information gain”; we also want to compute a a suitable notion of inverse for the encoder (i.e., a “decoder”). We formulate this family of learning problems as a class of two-player games. For a broad notion of game-theoretic equilibria which is learnable via standard gradient-based optimization techniques, we show that the equilibrium solutions to games within the class indeed result in maximally informative representations and a consistent autoencoding. We then apply this framework to several instances of the closed-loop transcription (CTRL) framework, which has been recently proposed for learning discriminative and generative representations for data lying on low- dimensional submanifolds, obtaining desirable representations which provably emulate and extend ones given by classical theory. Finally, we present a novel optimization algorithm to obtain the particular equilibria which our theory desires, and prove its correctness in a restricted case.}, }
EndNote citation:
%0 Thesis %A Pai, Druv %T Learning Low-Dimensional Structure via Closed-Loop Transcription: Equilibria and Optimization %I EECS Department, University of California, Berkeley %D 2023 %8 May 9 %@ UCB/EECS-2023-74 %U http://www2.eecs.berkeley.edu/Pubs/TechRpts/2023/EECS-2023-74.html %F Pai:EECS-2023-74