Skip to content

An encoder-decoder deep learning model (with/without attention mechanism) where the input is an arabic sign-language video and the output is its translation in text format.

Notifications You must be signed in to change notification settings

AI-14/video-captioning-for-arabic-sign-language-recognition-at-sentence-level

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

9 Commits
 
 
 
 
 
 

Repository files navigation

Video Captioning For Arabic Sign Language Recognition At Sentence Level

Table of Contents

  1. Desription
  2. Dataset
  3. Installation and Usage

Description

An encoder-decoder deep learning model (with/without attention mechanism) where the input is an arabic sign-language video and the output is its translation in text format.

Note: For a detailed model architecture and preprocesing, refer Video Captioning.ipynb file.

Dataset

The dataset consists of

  1. Total number of video samples are 534.
  2. 10 different sentences performed by three signers.
  3. Each video sample is already normalized to 80 frames.

Installation and Usage

  • Requirements
    • python >= 3.6
  • git clone https://github.com/AI-14/video-captioning-for-arabic-sign-language-recognition-at-sentence-level.git - clones the repository
  • cd video-captioning-for-arabic-sign-language-recognition-at-sentence-level
  • py -m venv yourVenvName - creates a virtual environment
  • pip install -r requirements.txt - installs all modules

About

An encoder-decoder deep learning model (with/without attention mechanism) where the input is an arabic sign-language video and the output is its translation in text format.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published