It appears you don't have support to open PDFs in this web browser. To view this file, Open with your PDF reader
Abstract
Whether it is Google's federated framework for mobile devices or OpenAI's large language models that are capturing the mainstream attention, large-scale learning systems are ubiquitous. While the rate of progress and performance of modern learning systems has been impressive, they are still hampered by many issues. For example, training large-scale learning models is costly both time and resource-wise, making guarantees with respect to an individual sample path imperative. In addition, it was noted that many learning models are impeded by the lack of classical smoothness and induce phenomena such as heavy-tailed gradient noise, necessitating the use of (stochastic) gradient methods with nonlinear mappings, such as sign, clipping or normalization. Moreover, the use of nonlinearly modified gradient methods is known to bring many benefits, such as stabilizing and accelerating training, reducing the size of transmitted messages, as well as enhancing security and privacy in distributed machine learning. Another issue stems from the fact that the data is generated across varied sources, making it difficult to train a single model that caters to the needs of a wide range of users. Toward resolving these issues, the first part of this thesis establishes learning guarantees for a general framework of nonlinear stochastic gradient methods in the presence of heavy-tailed noise. The general framework allows us to subsume many popular nonlinearities, like sign, normalization, clipping and quantization, providing a broad range of guarantees, including large deviation upper bounds and finite-time convergence, both in expectation and high-probability sense. The second part of the thesis is dedicated to studying the multi-model framework in distributed heterogeneous settings and designing algorithms that are able to utilize the wealth of data, while providing communication-efficient models, personalized to individual users.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer






