[2025/1/25] This paper is accepted by ICLR 2025
🎉🎉!
[2025/2/24] Release CSL-News dataset and code implementation.
We suggest to create a new conda environment.
# create environment
conda create --name Uni-Sign python=3.9
conda activate Uni-Sign
# install other relevant dependencies
pip install -r requirements.txt
Please follow the instructions provided in DATASET.md for data preparation.
All scripts must be executed within the Uni-Sign directory.
Stage 1: pose-only pre-training.
bash ./script/train_stage1.sh
Stage 2: RGB-pose pre-training.
bash ./script/train_stage2.sh
Stage 3: downstream fine-tuning.
bash ./script/train_stage3.sh
After completing stage 3 fine-tuning, performance evaluation on a single GPU can be performed using the following command:
bash ./script/eval_stage3.sh
- Release CSL-News dataset
- Release Uni-Sign implementation
If you have any questions, please feel free to contact Zecheng Li ([email protected]). Thank you.
The codebase of Uni-Sign is adapted from GFSLT-VLP. We are also grateful for the following projects our Uni-Sign arise from:
- 🤟SSVP-SLT: a excellent sign language translation framework!
- 🏃️MMPose: an open-source toolbox for pose estimation.
- 🤠FUNASR: a high-performance speech-to-text toolkit.
If you find Uni-Sign useful for your research and applications, please cite using this BibTeX:
@article{li2025uni,
title={Uni-Sign: Toward Unified Sign Language Understanding at Scale},
author={Li, Zecheng and Zhou, Wengang and Zhao, Weichao and Wu, Kepeng and Hu, Hezhen and Li, Houqiang},
journal={arXiv preprint arXiv:2501.15187},
year={2025}
}