Abstract

Symmetric functions, which take as input an unordered, fixed-size set, find practical application in myriad physical settings based on indistinguishable points or particles, and are also used as intermediate building blocks to construct networks with other invariances. Symmetric functions are known to be universally representable by neural networks that enforce permutation invariance. However the theoretical tools that characterize the approximation, optimization and generalization of typical networks fail to adequately characterize architectures that enforce invariance.

This thesis explores when these tools can be adapted to symmetric architectures, and when the invariance properties lead to new theoretical findings altogether. We study and prove approximation limitations on the extension of symmetric neural networks to infinite-sized inputs, the approximation capabilities of symmetric and antisymmetric networks relative to the interaction between set elements, and the learnability of simple symmetric functions with gradient methods.

Details

Title
Theory of Symmetric Neural Networks
Author
Zweig, Aaron
Publication year
2024
Publisher
ProQuest Dissertations & Theses
ISBN
9798381851311
Source type
Dissertation or Thesis
Language of publication
English
ProQuest document ID
2937439642
Copyright
Database copyright ProQuest LLC; ProQuest does not claim copyright in the individual underlying works.