Title: Uncertainty-Aware and Data-Efficient Fine-Tuning and Application of Foundation Models
Date: April 18th, 2025
Time: 1.00 PM — 2.30 PM, EDT
Location: CODA C1103 Lindberg; also on Zoom: https://gatech.zoom.us/j/94101453773
Yinghao Li
Machine Learning PhD Student
School of Electrical and Computer Engineering
Georgia Institute of Technology
Committee
1. Dr. Chao Zhang (CSE, Advisor)
2. Dr. Rampi Ramprasad (MSE, Co-Advisor)
3. Dr. Tuo Zhao (ISYE)
4. Dr. Srijan Kumar (CSE & Lighthouz AI)
5. Dr. Victor Fung (CSE)
6. Dr. Ali Torkamani (Amazon Web Services)
Abstract
Pre‑trained foundation models power modern natural language processing and scientific workflows, yet their deployment is hindered by training-inference distribution shifts, scarce in‑domain labels, model hallucination, and poorly calibrated model confidence. This thesis tackles these obstacles along two fronts: 1) reliable uncertainty quantification (UQ) and 2) data‑efficient model learning. For reliability, we establish MUBen, a best‑practice UQ benchmark for molecular property prediction, and propose UQAC, which backtracks attention chains to approximate the intractable marginalization over the reasoning space, yielding better calibrated answer probabilities from large language models (LLMs). To boost data efficiency, we introduce CHMM and its sparse variant for weakly supervised named entity recognition, and devise G&O, a zero‑shot information extraction framework that harnesses LLM reasoning. We further present ELREA, a fine‑tuning strategy that clusters input instructions by gradient direction for training task‑specific LoRA experts and inference-time expert ensembling, improving model generalization capability without relying on additional data points. Together, these contributions enhance the trustworthiness, robustness, and adaptability of foundation models in high‑stakes real‑world settings.