Scaling Interpretability for LLM Agents | Gabriele Sarti
Home
About me
Publications
Blog
Talks
Projects
Activities
CV
Academic CV
Short CV
Tools
Inseq
LangLearn
Communities
AI2S
AISIG
Scaling Interpretability for LLM Agents
Gabriele Sarti
Natural Language Processing
,
Academic
Code
Project
Project
Slides
Date
Mar 27, 2026
Event
Seminar at the BauLab Group of Northeastern University
Location
177 Huntington Ave, 22nd Floor
Boston, MA, USA
Natural Language Processing
Interpretability
Sequence-to-sequence
Language Modeling
Feature Attribution
Retrieval-augmented Generation
NDIF
Mechanistic Interpretability
Agents
Goal-directedness
Related
Attribution: Tracing Influence to Inputs and Model Components
Interpretability for Language Models: Current Trends and Applications
Interpreting Context Usage in Generative Language Models
Interpreting Context Usage in Generative Language Models
Interpreting Context Usage in Generative Language Models
Cite
×