GuangxueWang / SLGTformer

An Attention Based Approach to Sign Language Recognition | SOTA on WLASL Joints | https://arxiv.org/abs/2212.10746

Home Page:https://arxiv.org/abs/2212.10746

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

SLGTformer: An Attention Based Approach to Sign Language Recognition

If you find this code useful for your research, consider citing:

@misc{https://doi.org/10.48550/arxiv.2212.10746,
  doi = {10.48550/ARXIV.2212.10746},
  url = {https://arxiv.org/abs/2212.10746},
  author = {Song, Neil},
  keywords = {Computer Vision and Pattern Recognition (cs.CV), FOS: Computer and information sciences, FOS: Computer and information sciences},
  title = {SLGTformer: An Attention-Based Approach to Sign Language Recognition},
  publisher = {arXiv},
  year = {2022},
  copyright = {Creative Commons Attribution 4.0 International}
}

Data preparation

  1. Please download and use the preprocessed skeleton data for WLASL by Skeleton Aware Multi-modal Sign Language Recognition. Please be sure to follow their rules and agreements when using the preprocessed data.
    ./download.sh
    

Pretrained models

Pretrained models are provided here.

Usage

Train WLASL:

./train.sh

Test:

./test.sh

Acknowledgements

SAM-SLR-v2

This code is based on SAM-SLR-v2. Huge thank you to the authors for open sourcing their code.

IRVL

Thank you to @yuxng for his advice and guidance throughout this project. Shout-out to his lab @IRVL for the RTX A5000s and all the fun conversations while models were training.

About

An Attention Based Approach to Sign Language Recognition | SOTA on WLASL Joints | https://arxiv.org/abs/2212.10746

https://arxiv.org/abs/2212.10746

License:Other


Languages

Language:Python 99.0%Language:Shell 1.0%