Skip to content

JSALT2024/slt-demo

Repository files navigation

slt-demo

Sign Language Translation - Gradio web app for Video to Text

This demo is a proof of concept for the translation system of ASL. It uses YoloV8, MediaPipe and T5 in the backend, trained on the YoutubeASL dataset by the CV team at UWB. See T5_for_SLT and PoseEstimation.

Find the demo at HuggingFace spaces

How to run locally

  1. Clone repository
git clone https://github.com/JSALT2024/slt-demo.git
  1. Install dependencies
pip install -r requirements.txt
  1. Run the demo
python app.py
  1. Open http://127.0.0.1:7860/

  2. You are good to go! Upload any video or use your webcam 📷

Code structure

slt-demo/
├── backend.py
├── predict_pose.py
├── sltGradio.py
├── checkpoints/
│   ├── pose/
│   │   └── [pose model files]
│   └── t5-v1_1-base/
│       └── [T5 model files]
├── configs/
│   └── predict_config_demo.yaml
├── dataset/
│   └── generic_sl_dataset.py
├── model/
│   ├── configuration_t5.py
│   └── modeling_t5.py
├── utils/
│   └── translation.py
└── video/
    └── [example videos]

About

Sign Language Recognition - Gradio web app for Video to Caption

Topics

Resources

License

Stars

Watchers

Forks