Gabriele Sarti
Avatar

Gabriele Sarti

PhD in Natural Language Processing

CLCG, University of Groningen

About me

Welcome to my website! 👋 I am a PhD student at the Computational Linguistics Group of the University of Groningen and member of the NWO-funded consortium InDeep: Interpreting Deep Learning Models for Text and Sound, focusing on interpretability for neural machine translation. I am supervised by Arianna Bisazza, Malvina Nissim and Grzegorz Chrupała.

Previously, I was a research intern at Amazon Translate, a research scientist at Aindo, a student in the Data Science MSc at University of Trieste & SISSA and a founding member of the AI Student Society. My master’s thesis with the ItaliaNLP Lab concerned the study of linguistic complexity using gaze recordings and neural language models.

My research focuses on interpretability for NLP models, in particular to the benefit of end-users and by leveraging human behavioral signals. I am also passionate about social applications of machine learning, ethical AI, and open source collaboration.

Interests

  • Conditional Text Generation
  • Interpretability for Deep Learning
  • Behavioral Data for NLP
  • Uncertainty Estimation in NLP

Education

Experience

🗞️ News

 

Selected Publications

 

DivEMT: Neural Machine Translation Post-Editing Effort Across Typologically Diverse Languages

We introduce DivEMT, the first publicly available post-editing study of Neural Machine Translation over a typologically diverse set of …

IT5: Large-scale Text-to-text Pretraining for Italian Language Understanding and Generation

We present IT5, the first family of encoder-decoder transformer models pretrained specifically on Italian on more than 40 billion …

Contrastive Language-Image Pre-training for the Italian Language

We present the first CLIP model for the Italian Language (CLIP-Italian), trained on more than 1.4 million image-text pairs.

Blog posts

 

ICLR 2020 Trends: Better & Faster Transformers for Natural Language Processing

A summary of promising directions from ICLR 2020 for better and faster pretrained tranformers language models.

Recent & Upcoming Talks

Towards User-centric Interpretability of Machine Translation Models
Towards User-centric Interpretability of NLP Models
Empowering Human Translators via Interpretable Interactive Neural Machine Translation

Projects

 

Contrastive Image-Text Pretraining for Italian

The first CLIP model pretrained on the Italian language.

Covid-19 Semantic Browser

A semantic browser for SARS-CoV-2 and COVID-19 powered by neural language models.

AItalo Svevo: Letters from an Artificial Intelligence

Generating letters with a neural language model in the style of Italo Svevo, a famous italian writer of the 20th century.

Histopathologic Cancer Detection with Neural Networks

A journey into the state of the art of histopathologic cancer detection approaches.