Mechanistic Interpretability Workshop

ICML 2026

July 10 or 11, 2026 (TBC)
COEX Convention & Exhibition Center · Seoul, South Korea

Call for papers and schedule coming soon

As neural networks grow in influence and capability, understanding the mechanisms behind their decisions remains a fundamental scientific challenge. This gap between performance and understanding limits our ability to predict model behavior, ensure reliability, and detect sophisticated adversarial or deceptive behavior. Many of the deepest scientific mysteries in machine learning may remain out of reach if we cannot look inside the black box.

Mechanistic interpretability addresses this challenge by developing principled methods to analyze and understand a model’s internals–weights and activations–and to use this understanding to gain greater insight into its behavior, and the computation underlying it.

The field has grown rapidly, with sizable communities in academia, industry and independent research, dedicated startups, and a rich ecosystem of tools and techniques. Following our workshops at ICML 2024 and NeurIPS 2025, this edition at ICML 2026 aims to bring together diverse perspectives from the community to discuss recent advances, build common understanding and chart future directions.

Call for papers coming soon.

Paper submission deadline — mid May 2026.

ICML 2024 Workshop ICML 2024 Social

The first Mechanistic Interpretability Workshop (ICML 2024).

Organizing Committee

Neel Nanda

Neel Nanda

Google DeepMind

Andrew Lee

Andrew Lee

Harvard

Andy Arditi

Andy Arditi

Northeastern University

Stefan Heimersheim

Stefan Heimersheim

Google DeepMind

Anna Soligo

Anna Soligo

Imperial College London

J Rosser

University of Oxford

Questions? Email icml2026@mechinterpworkshop.com

Curve detector visualization

What are those beautiful rainbow flower things?

These are visualizations of "curve detector" neurons from early mechanistic interpretability research. Learn more in the Curve Detectors article on Distill.