Augmenting interpretable models with large language models during training

Abstract Recent large language models (LLMs), such as ChatGPT, have demonstrated remarkable prediction performance for a growing array of tasks. However, their proliferation into high-stakes domains and compute-limited settings has created a burgeoning need for interpretability and efficiency. We ad...

Full description

Bibliographic Details
Main Authors: Chandan Singh, Armin Askari, Rich Caruana, Jianfeng Gao
Format: Article
Language:English
Published: Nature Portfolio 2023-11-01
Series:Nature Communications
Online Access:https://doi.org/10.1038/s41467-023-43713-1