Skip to content

Latest commit

 

History

History
52 lines (36 loc) · 2.75 KB

README.md

File metadata and controls

52 lines (36 loc) · 2.75 KB

SignAttention: On the Interpretability of Transformer Models for Sign Language Translation

Official PyTorch implementation of SignAttention: On the Interpretability of Transformer Models for Sign Language Translation
Accepted at IAI Workshop @ NeurIPS 2024

Authors:
Pedro Alejandro Dal Bianco (*), Oscar Agustín Stanchi (*), Facundo Manuel Quiroga, Franco Ronchetti, Enzo Ferrante.
(* denotes equal contribution)

Overview

This repository provides the official implementation of our paper "SignAttention: On the Interpretability of Transformer Models for Sign Language Translation". The paper presents the first comprehensive interpretability analysis of a Transformer-based model for translating Greek Sign Language videos into glosses and text, leveraging attention mechanisms to provide insights into the model's decision-making processes.

Dependencies

The code has been written in Python (3.10) and requires PyTorch (2.0). Install the required dependencies with:

pip install -r requirements.txt

Data

We utilize the Greek Sign Language Dataset, which contains 10,290 samples of RGB videos, gloss annotations, and Greek language translations. This dataset is ideal for interpretability analysis due to its detailed gloss annotations and repetition of sentences. Download the dataset here.

Experiments

You can view the poster we presented at NeurIPS 2024, which includes sample visualizations of the attention mechanisms analyzed in the paper, by clicking the link below:

Poster for NeurIPS 2024

Usage

- src/train.ipynb                          : Train the model

- src/get_interp_weights.ipynb             : Obtain attention weights

- src/interp.ipynb                         : Visualize attention for a single sample

- src/interp_all.ipynb                     : Generate global attention visualizations

Citation

If you find this repository useful for your work, please cite:

@article{bianco2024signattention,
  title={SignAttention: On the Interpretability of Transformer Models for Sign Language Translation},
  author={Bianco, Pedro Alejandro Dal and Stanchi, Oscar Agust{\'\i}n and Quiroga, Facundo Manuel and Ronchetti, Franco and Ferrante, Enzo},
  journal={arXiv preprint arXiv:2410.14506},
  year={2024}
}

For queries regarding the paper or the code, please contact [email protected] and [email protected], or open an issue on this repository.