Interpreting Latent Features in Large Language Models | Gabriele Sarti
Home
About me
Publications
Blog
Talks
Projects
Activities
CV
Communities
AI2S
AISIG
Interpreting Latent Features in Large Language Models
Gabriele Sarti
Natural Language Processing
,
Academic
PDF
Slides
Date
May 22, 2025
Event
Paper Presentation at InCLoW Reading Group
Location
Room 1313.0344, Harmonie Building, University of Groningen
Groningen, Netherlands
Natural Language Processing
Interpretability
Language Modeling
Feature Attribution
Concept-based Interpretability
Related
Interpretability for Language Models: Current Trends and Applications
Interpreting Context Usage in Generative Language Models
QE4PE: Word-level Quality Estimation for Human Post-Editing
Interpreting Context Usage in Generative Language Models with Inseq, PECoRe and MIRAGE
Interpretability for Language Models: Current Trends and Applications
Cite
×