Content area

Abstract

This dissertation explores the evolution and application of artificial intelligence techniques across three critical domains: financial modeling, mathematical reasoning, and structured data analysis. The dissertation presents seven research projects that chart a progression from specialized neural architectures to sophisticated large language models (LLMs), contributing novel methodologies and frameworks at each stage.

In the financial domain, the research first introduces TS-Mixer, a MLP-based architecture for time-series forecasting that captures both feature relationships and temporal dependencies through a simple yet effective design, outperforming more complex models in S&P500 index prediction. The dissertation then presents DySTAGE, a dynamic graph representation learning framework that addresses the evolving nature of financial markets by modeling changing asset relationships, demonstrating superior performance in both predictive accuracy and portfolio optimization. Finally, the dissertation proposes a hybrid framework integrating LLMs with reinforcement learning for adaptive margin trading, enabling dynamic risk management through explainable market reasoning.

For mathematical reasoning, the research develops two novel evaluation frameworks that expand beyond traditional correctness metrics: CreativeMath and FaultyMath. CreativeMath assesses LLMs' ability to generate novel, insightful solutions to mathematical problems, introducing a comprehensive benchmark of competition-level problems with multiple human solutions. FaultyMath evaluates logical robustness by testing whether models can identify logically flawed or unsolvable problems, revealing significant gaps in current systems' critical thinking capabilities.

In structured data analysis, the dissertation introduces DataFrame QA, a privacy-preserving framework that enables natural language interaction with tabular data without exposing sensitive information, achieving high accuracy while eliminating data exposure risks. The research also presents TextFlow, a modular approach to flowchart understanding that separates visual extraction from semantic reasoning, demonstrating substantial improvements over end-to-end vision-language models in accuracy and interpretability.

Collectively, these contributions advance AI capabilities across multiple dimensions—efficiency, adaptability, creativity, logical robustness, privacy, and interpretability—while establishing methodologies that leverage the strengths of different AI paradigms for complex analytical tasks. The dissertation provides both theoretical insights and practical frameworks that bridge the gap between specialized neural architectures and general-purpose language models, with applications in finance, education, data science, and beyond.

Details

1010268
Business indexing term
Title
From Neural Networks to Large Language Models: Innovations in Financial AI, Mathematical Reasoning, and Structured Data Representation
Author
Ye, Junyi  VIAFID ORCID Logo 
Number of pages
291
Publication year
2025
Degree date
2025
School code
0152
Source
DAI-A 87/2(E), Dissertation Abstracts International
ISBN
9798290935249
Committee member
Wei, Zhi; Sharma, Shantanu; Liu, Rong; Yin, Wenpeng
University/institution
New Jersey Institute of Technology
Department
Department of Computer Science
University location
United States -- New Jersey
Degree
Ph.D.
Source type
Dissertation or Thesis
Language
English
Document type
Dissertation/Thesis
Dissertation/thesis number
31844448
ProQuest document ID
3237577613
Document URL
https://www.proquest.com/dissertations-theses/neural-networks-large-language-models-innovations/docview/3237577613/se-2?accountid=208611
Copyright
Database copyright ProQuest LLC; ProQuest does not claim copyright in the individual underlying works.
Database
ProQuest One Academic