Dear faculty members and fellow students,

 

You are cordially invited to my thesis defense on Dec 2nd.

 

Title: Towards Interpretable and Controllable Machine Learning Models via Logic Reasoning

 

Date: 12/02/2024

Time: 11:00AM EST

Location: https://gatech.zoom.us/j/94374751336

 

Yuan Yang

Machine Learning PhD Student

School of Computational Science and Engineering

Georgia Institute of Technology

 

Committee

1 Dr. Faramarz Fekri, School of Electrical and Computer Engineering, Georgia Institute of Technology (Advisor)

2 Dr. Zsolt Kira, School of Interactive Computing, Georgia Institute of Technology

3 Dr. Larry Heck, School of Electrical and Computer Engineering, Georgia Institute of Technology

4 Dr. Viveck Cadambe, School of Electrical and Computer Engineering, Georgia Institute of Technology

5 Dr. Bo Dai, School of Computational Science and Engineering, Georgia Institute of Technology

 

Abstract

Modern machine learning models have provided new capabilities across a spectrum of applications in vision, reasoning, and natural language processing.  However, these models are criticized for being non-interpretable, data-inefficient, and vulnerable to subtle perturbations such as adversarial attacks and distribution shifts. Addressing these issues remains at the center of developing trustworthy ML systems for real-world applications.

 

Our research focuses on providing a principled solution to these issues through logic reasoning formalism.

Specifically, we study the fundamental technique of inductive logic programming (ILP) that learns and represents patterns in knowledge graphs as first-order logic (FOL) rules, providing an interpretable approach to various reasoning tasks on structured data:

(1) we investigate the connection between model explanation and logic formalism and propose frameworks for explaining and defending ML models via logic reasoning;

(2) we formalize logic reasoning methods as a novel data programming paradigm and propose data-efficient frameworks for model training and evaluation;

(3) to improve the expressiveness of the ILP technique, we propose to extend the model to the temporal domain and hypergraphs so that one can generalize FOL rules on complex structures

 

Furthermore, our research explores the integration of large language models (LLMs) with logical reasoning techniques to enhance interpretability, data efficiency, and controllability in machine learning systems. We investigate:

(1) the potential of LLMs in translating natural language to formal logical representations to solve complex reasoning problems;

(2) enhancing LLMs' reasoning capability on open-ended, ambiguous problems by incorporating formal logic reasoning, thereby improving their controllability and robustness beyond narrowly defined domains. 

By combining logic reasoning with the latest advancements in LLMs, our research aims to bridge the gap between powerful ML models and the need for explainable, efficient, and reliable AI systems in real-world applications.