What is it about?

We provide a comprehensive overview of existing interpretability techniques for medium-sized models and large language models respectively. In the part of medium-sized models, interpretability methodologies are introduced in two classes: local and global explanation. Unlike local explanation is less meaningful to LLMs, global explanation is still a major tool. Besides, we introduce some interpretability approaches that are specifically applying to LLMs.

Featured Image

Why is it important?

In the era of LLMs, it is a quick start for practitioners to learn what tools they can use to interpret their learning models. We also provide some ideas on why some techniques applicable to medium-sized models are no long useful with LLMs.

Read the Original

This page is a summary of: Explainability for Large Language Models: A Survey, ACM Transactions on Intelligent Systems and Technology, January 2024, ACM (Association for Computing Machinery),
DOI: 10.1145/3639372.
You can read the full text:

Read

Contributors

The following have contributed to this page