PhD Defense |Towards Interpretable and Controllable Machine Learning Models via Logic Reasoning

Yuan Yang - Machine Learning PhD Student - School of Computational Science and Engineering

 

Date: December 2nd 

Time: 11:00 AM – 1:00 PM EST

Location: Online

Meeting Link: https://gatech.zoom.us/j/94374751336

Committee

1 Dr. Faramarz Fekri, School of Electrical and Computer Engineering, Georgia Institute of Technology (Advisor)

2 Dr. Zsolt Kira, School of Interactive Computing, Georgia Institute of Technology

3 Dr. Larry Heck, School of Electrical and Computer Engineering, Georgia Institute of Technology

4 Dr. Viveck Cadambe, School of Electrical and Computer Engineering, Georgia Institute of Technology

5 Dr. Bo Dai, School of Computational Science and Engineering, Georgia Institute of Technology

Abstract

Modern machine learning models have provided new capabilities across a spectrum of applications in vision, reasoning, and natural language processing.  However, these models are criticized for being non-interpretable, data-inefficient, and vulnerable to subtle perturbations such as adversarial attacks and distribution shifts. Addressing these issues remains at the center of developing trustworthy ML systems for real-world applications.

Our research focuses on providing a principled solution to these issues through logic reasoning formalism.

Specifically, we study the fundamental technique of inductive logic programming (ILP) that learns and represents patterns in knowledge graphs as first-order logic (FOL) rules, providing an interpretable approach to various reasoning tasks on structured data:

  1. we investigate the connection between model explanation and logic formalism and propose frameworks for explaining and defending ML models via logic reasoning;
  2. we formalize logic reasoning methods as a novel data programming paradigm and propose data-efficient frameworks for model training and evaluation;
  3. to improve the expressiveness of the ILP technique, we propose to extend the model to the temporal domain and hypergraphs so that one can generalize FOL rules on complex structures

Furthermore, our research explores the integration of large language models (LLMs) with logical reasoning techniques to enhance interpretability, data efficiency, and controllability in machine learning systems. We investigate:

  1. the potential of LLMs in translating natural language to formal logical representations to solve complex reasoning problems;
  2. enhancing LLMs' reasoning capability on open-ended, ambiguous problems by incorporating formal logic reasoning, thereby improving their controllability and robustness beyond narrowly defined domains. 

By combining logic reasoning with the latest advancements in LLMs, our research aims to bridge the gap between powerful ML models and the need for explainable, efficient, and reliable AI systems in real-world applications.

Event Details

Date/Time:

  • Monday, December 2, 2024
    11:00 am - 1:00 pm
Location: https://gatech.zoom.us/j/94374751336