Abstract

Deep learning is effective at large scales due to advances in self-supervised learning, a paradigm that encompasses a broad class of training algorithms capable of learning useful and informative representations end-to-end without explicit data labels. Despite their success, many widespread deep learning models are still limited in their ability to generalize to new tasks and data distributions, and often require large amounts of labeled data to achieve good performance. In this thesis, we explore ways to use regularization in order to improve deep learning models and prevent collapse in their hidden representations, an undesirable scenario in which a network learns trivial or uninformative representations. Throughout the various included works, we take the lens of energy-based models as the learning framework where we apply our regularization techniques. In the first part of the thesis, we focus on sparse coding, a classic self-supervised algorithm for extracting image representations. We extend the original sparse coding algorithm to incorporate a non-linear decoder, then evaluate on different tasks including image classification in the low-data regime. In the second part of the thesis, we focus on building world models through self-supervised video representation learning using joint-embedding predictive architectures as an alternative to generative predictive models. Our study suggests that our approach yields more information-rich video representations. Finally, we present research on improving video representations through variance and covariance regularization in the setting of supervised transfer learning. We hope that our findings spur new research into using regularization techniques to prevent collapse both in the current and in the next generation of deep learning architectures.

Details

Title
Representation Learning With Regularized Energy-Based Models
Author
Drozdov, Katrina Vladimirova
Publication year
2024
Publisher
ProQuest Dissertations & Theses
ISBN
9798342711135
Source type
Dissertation or Thesis
Language of publication
English
ProQuest document ID
3121538548
Copyright
Database copyright ProQuest LLC; ProQuest does not claim copyright in the individual underlying works.