Content area

Abstract

The advent of large foundation models like ChatGPT and the concept of artificial general intelligence have shifted the machine learning paradigm from “one model per task” to “one large model, many tasks.” On one hand, the prevalent LLM-based foundation models excel at many tasks that can be described by natural language, programming language, and mathematical language. On the other hand, they still struggle with tasks of other modalities, such as DNA data and continuous time-series data. This limitation has led researchers to pursue specialized foundation models for other data domains by transferring the techniques from LLM-based foundation models. This thesis is to look at the concept of foundation models with a focus on the track with less attention from the research community, which is about time series foundation models and their applications to scientific discoveries. Initial attempts of time series foundation models (TSFMs) have demonstrated superior efficacy across various benchmark datasets, outperforming traditional one-model-per-task approaches in forecasting tasks. While these models are pre-trained on hundreds of gigabytes of multi-domain time series data and achieve good performance on benchmarks, there still exist limitations and challenges to developing a better TSFM with more comprehensive abilities. Besides, their potential contribution to scientific discoveries remains largely unexplored. Particularly, questions persist regarding their ability to handle irregularly sampled scientific time series and their effectiveness in domain-specific downstream tasks such as variable star classification in astrophysics.

This thesis looks at the realm of time series foundation models from three perspectives. i) Starting from the oracle of foundation models, LLM-based foundation models, I study the adaptive batch size technique to improve the pre-training efficiency for large models. ii) I study better methodologies for each step in the time series foundation models’ pipeline that contribute to developing a better time series foundation model. iii) I study time series foundation models’ applications to accelerate scientific discoveries in astrophysics.

Details

1010268
Business indexing term
Title
Time Series Foundation Models and Their Applications to Scientific Discoveries
Author
Number of pages
306
Publication year
2025
Degree date
2025
School code
0163
Source
DAI-B 87/3(E), Dissertation Abstracts International
ISBN
9798291587218
Advisor
Committee member
Chen, Yan; Wang, Zhaoran; Li, Manling
University/institution
Northwestern University
Department
Computer Science
University location
United States -- Illinois
Degree
Ph.D.
Source type
Dissertation or Thesis
Language
English
Document type
Dissertation/Thesis
Dissertation/thesis number
32173187
ProQuest document ID
3245357208
Document URL
https://www.proquest.com/dissertations-theses/time-series-foundation-models-their-applications/docview/3245357208/se-2?accountid=208611
Copyright
Database copyright ProQuest LLC; ProQuest does not claim copyright in the individual underlying works.
Database
ProQuest One Academic