Skip to content

ShesterG/SHuBERT

Repository files navigation

SHuBERT: Self-Supervised Sign Language Representation Learning via Multi-Stream Cluster Prediction [ACL 2025]

This repository contains research code for the paper SHuBERT: Self-Supervised Sign Language Representation Learning via Multi-Stream Cluster Prediction.

SHuBERT Overview

We introduce SHuBERT (Sign Hidden-Unit BERT), a self-supervised contextual representation model learned from approximately 1,000 hours of American Sign Language video. SHuBERT adapts masked token prediction objectives to multi-stream visual sign language input, learning to predict multiple targets corresponding to clustered hand, face, and body pose streams. SHuBERT achieves state-of-the-art performance across multiple tasks including sign language translation, isolated sign language recognition, and fingerspelling detection.


Installation

We provide installation and inference instructions in QUICKSTART.md.

Usage

1. Preparing the data

We describe how to prepare the datasets in DATASETS.md.

2. Model Weights

Please download the weight of SHuBERT (as well as the DINO Face and Hand) weights link.

3. Feature Extraction

We describe how to extract features from the pretrained model in FEATURES.md.

4. Pretraining

  • sbatch train_shubert.sh

5. Fine-tuning on Downstream Tasks

TODO


Citing our work

If you find our work useful in your research, please consider citing:

@inproceedings{gueuwou-etal-2025-shubert,
    title = "SHuBERT: Self-Supervised Sign Language Representation Learning via Multi-Stream Cluster Prediction",
    author = "Gueuwou, Shester and Du, Xiaodan and Shakhnarovich, Greg and Livescu, Karen and Liu, Alexander H.",
    booktitle = "Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)",
    year = "2025",
    address = "Vienna, Austria",
    publisher = "Association for Computational Linguistics",
}

References

This codebase is heavily influenced by the DinoSR and Fairseq repositories.

License

This project is primarily under the MIT license.

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 2

  •  
  •