Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Sign language is a cardinal element for communication between deaf and dumb community. Sign language has its own grammatical structure and gesticulation nature. Research on SLRT focuses a lot of attention in gesture identification. Sign language comprises of manual gestures performed by hand poses and non-manual features expressed through eye, mouth and gaze movements. The sentence-level completely labelled Indian Sign Language dataset for Sign Language Translation and Recognition (SLTR) research is developed. The ISL-CSLTR dataset assists the research community to explore intuitive insights and to build the SLTR framework for establishing communication with the deaf and dumb community using advanced deep learning and computer vision methods for SLTR purposes. This ISL-CSLTR dataset aims in contributing to the sentence level dataset created with two native signers from Navajeevan, Residential School for the Deaf, College of Spl. D.Ed & B.Ed, Vocational Centre, and Child Care & Learning Centre, Ayyalurimetta, Andhra Pradesh, India and four student volunteers from SASTRA Deemed University, Thanjavur, Tamilnadu. The ISL-CSLTR corpus consists of a large vocabulary of 700 fully annotated videos, 18863 Sentence level frames, and 1036 word level images for 100 Spoken language Sentences performed by 7 different Signers. This corpus is arranged based on signer variants and time boundaries with fully annotated details and it is made available publicly. The main objective of creating this sentence level ISL-CSLRT corpus is to explore more research outcomes in the area of SLTR. This completely labelled video corpus assists the researchers to build framework for converting spoken language sentences into sign language and vice versa. This corpus has been created to address the various challenges faced by the researchers in SLRT and significantly improves translation and recognition performance. The videos are annotated with relevant spoken language sentences provide clear and easy understanding of the corpus data. Acknowledgements: The research was funded by the Science and Engineering Research Board (SERB), India under Start-up Research Grant (SRG)/2019–2021 (Grant no. SRG/2019/001338). And also, we thank all the signers for their contribution in collecting the sign videos and the successful completion of the ISL-CSLTR corpus. We would like to thank Navajeevan, Residential School for the Deaf, College of Spl. D.Ed & B.Ed, Vocational Centre, and Child Care & Learning Centre, Ayyalurimetta, Andhra Pradesh, India for their support and contribution.
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
## Overview
Indian Sign Language Detection is a dataset for object detection tasks - it contains A Z 0 9 annotations for 1,748 images.
## Getting Started
You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
## License
This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).
Facebook
TwitterAttribution-ShareAlike 4.0 (CC BY-SA 4.0)https://creativecommons.org/licenses/by-sa/4.0/
License information was derived automatically
Indian Sign Language (ISL) is a complete language with its own grammar, syntax, vocabulary and several unique linguistic attributes. It is used by over 5 million deaf people in India. Currently, there is no publicly available dataset on ISL to evaluate Sign Language Recognition (SLR) approaches. In this work, we present the Indian Lexicon Sign Language Dataset - INCLUDE - an ISL dataset that contains 0.27 million frames across 4,287 videos over 263 word signs from 15 different word categories. INCLUDE is recorded with the help of experienced signers to provide close resemblance to natural conditions. This dataset is further modified by adding Hand-Landmarks and Pose-Landmarks and cropping it to fit the specifications of google/vivit-b-16x2
NOTE: Only 80 word signs are uploaded for now, mode word signs will be updated soon.
Original Dataset can be obtained at: * https://zenodo.org/records/4010759 * https://huggingface.co/datasets/ai4bharat/INCLUDE
Citation [1]A. Sridhar, R. G. Ganesan, P. Kumarand M. Khapra, ‘INCLUDE: A Large Scale Dataset for Indian Sign Language Recognition’. Zenodo, Oct. 16, 2020. doi: 10.1145/3394171.3413528.
Facebook
TwitterCC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
License information was derived automatically
A curated Indian Sign Language (ISL) dataset featuring hand gestures, facial expressions, and motion sequences designed for AI/ML sign-language recognition, gesture classification, accessibility tools, and human–computer interaction systems.
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Abstract: Indian Sign Language (ISL) is a complete language with its own grammar, syntax, vocabulary and several unique linguistic attributes. It is used by over 5 million deaf people in India. Currently, there is no publicly available dataset on ISL to evaluate Sign Language Recognition (SLR) approaches. In this work, we present the Indian Lexicon Sign Language Dataset - INCLUDE - an ISL dataset that contains 0.27 million frames across 4,287 videos over 263 word signs from 15 different word categories. INCLUDE is recorded with the help of experienced signers to provide close resemblance to natural conditions. A subset of 50 word signs is chosen across word categories to define INCLUDE-50 for rapid evaluation of SLR methods with hyperparameter tuning. The best performing model achieves an accuracy of 94.5% on the INCLUDE-50 dataset and 85.6% on the INCLUDE dataset. Download Instructions: For ease of access, we have prepared a Shell Script to download all the parts of the dataset and extract them to form the complete INCLUDE dataset.You can find the script here: http://bit.ly/include_dl
Facebook
TwitterAttribution-NoDerivs 4.0 (CC BY-ND 4.0)https://creativecommons.org/licenses/by-nd/4.0/
License information was derived automatically
The Indian Sign Language Alphabet Dataset is a curated collection of hand gesture images representing the alphabets (A–Z) in Indian Sign Language (ISL). It was specifically created to support the development of Gesture-Speech, a groundbreaking project that translates Indian Sign Language gestures into spoken words, making communication more accessible for the hearing and speech-impaired communities.
The dataset is intended for researchers, developers, and educators to build robust machine learning models for recognizing ISL gestures and advancing assistive technologies.
The dataset was developed as part of the Gesture-Speech project, an open-source initiative hosted on GitHub. Gesture-Speech focuses on real-time translation of Indian Sign Language gestures into spoken language using deep learning and computer vision. The project’s mission is to: - Empower differently-abled individuals by breaking communication barriers. - Promote inclusivity through assistive AI-driven technologies. - Enable the use of ISL in everyday interactions.
The dataset serves as the backbone of Gesture-Speech’s training pipeline, allowing the development of accurate and efficient recognition models.
Dataset Highlights:
Integration with Gesture-Speech:
The dataset follows a directory-based structure for classification tasks:
Indian_Sign_Language_Alphabet/
│── a/
│ ├── 0.jpg
│ ├── 1.jpg
│ ├── ...
│── b/
│ ├── 0.jpg
│ ├── 1.jpg
│ ├── ...
│── ...
│── z/
384x384 pixels for consistency..jpg format.The Gesture-Speech project can be found on GitHub. Key highlights of the project: - Uses TensorFlow and OpenCV to process hand gestures in real-time. - Translates ISL gestures into spoken English or Hindi phrases using a deep learning pipeline. - Open-source and extensible, encouraging contributions and further development.
This dataset is released under the Creative Commons Attribution 4.0 International (CC BY 4.0) license. Users can: - Share: Copy and redistribute the material in any medium or format. - Adapt: Remix, transform, and build upon the material for any purpose, even commercially. - Attribution: Proper credit must be given to the dataset creator.
The dataset and the Gesture-Speech project are dedicated to promoting inclusivity and enabling new opportunities for the differently-abled. Special thanks to contributors and volunteers who helped create this dataset by providing diverse samples of ISL gestures.
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
The dataset contains the RGB images of hand gestures of twenty ISL words, namely, ‘afraid’,’agree’,’assistance’,’bad’,’become’,’college’,’doctor’,’from’,’pain’,’pray’, ’secondary’, ’skin’, ’small’, ‘specific’, ‘stand’, ’today’, ‘warn’, ‘which’, ‘work’, ‘you’’ which are commonly used to convey messages or seek support during medical situations. All the words included in this dataset are static. The images were captured from 8 individuals including 6 males and 2 females in the age group of 9 years to 30 years. The dataset contains a 18000 images in jpg format. The images are labelled using the format ISLword_X_YYYY_Z, where: • ISLword corresponds to the words ‘afraid’, ‘agree’, ‘assistance’, ‘bad’, ‘become’, ‘college’, ‘doctor’ ,‘from’, ’pray’, ‘pain’, ‘secondary’, ‘skin’, ‘small’, ‘specific’, ‘stand’, ‘today’, ‘warn’, ‘which’, ‘work’, ‘you’. • X is an image number in the range 1 to 900. • YYYY is an identifier of the participant and is in the range of 1 to 6. • Z corresponds to 01 or 02 that identifies the sample number for each subject. For example, the file named afraid_1_user1_1 is the image sequence of the first sample of the ISL gesture of the word ‘afraid’ presented by the 1st user.
Facebook
TwitterMIT Licensehttps://opensource.org/licenses/MIT
License information was derived automatically
The dataset contains word-level Indian Sign Language videos in .mp4 format. These signs were sourced from the official Indian Sign Language YouTube channel, maintained by the government body. Four contributors created 60 unique signs, each with 60 corresponding videos. Additionally, 30 videos of still signs were included to improve model accuracy and avoid misinterpretation.
Paper link - https://link.springer.com/chapter/10.1007/978-981-97-6992-6_19
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Recent advancements in sign language recognition technology have significantly improved communication for individuals who are deaf or hard of hearing. Despite these advancements, many people who use sign language still face challenges in everyday interactions due to widespread unfamiliarity with sign language. However, new technologies have greatly enhanced our ability to recognize and interpret sign language, making communication more accessible and inclusive.
This dataset includes images of common phrases in both Indian Sign Language (ISL) and American Sign Language (ASL). The images were captured using a standard laptop webcam with a resolution of 680x480 pixels and a bit depth of 24 pixels. The dataset covers 44 different phrases, each represented by 40 images. All images are stored in PNG format. Note that this dataset includes static signs only and does not contain any dynamic sign language gestures.
Facebook
TwitterOpen Database License (ODbL) v1.0https://www.opendatacommons.org/licenses/odbl/1.0/
License information was derived automatically
This dataset contains a comprehensive collection of Indian Sign Language (ISL) hand gestures representing the 26 letters of the English alphabet, captured using the MediaPipe pose estimation technology. The dataset includes over 50,000 high-quality images of hand gestures captured from various angles, lighting conditions, and skin tones. The dataset is ideal for researchers and developers working in the field of computer vision, machine learning, and sign language recognition.
Facebook
TwitterMIT Licensehttps://opensource.org/licenses/MIT
License information was derived automatically
## Overview
Indian Sign Language is a dataset for object detection tasks - it contains Sign_Language annotations for 1,866 images.
## Getting Started
You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
## License
This dataset is available under the [MIT license](https://creativecommons.org/licenses/MIT).
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
This dataset consists of static hand gestures and lips movement for each character in the English alphabet, eight Hindi Vowels and ten Numerals as represented in Indian sign language (ISL). The dataset consists of 1,02,470 images of subjects from different age groups presenting static gestures under varied backgrounds and illumination conditions. The dataset is structured into three folders namely Kids, Teenagers and Adults. Each folder consists of sub-folders namely Full Sleeves and Half Sleeves indicating the type of clothing that the subject has worn at the time of image acquisition. In each sub-folders, images for the English alphabet, Hindi Vowels and Numerals are stored respectively in the sub-folders named with that specific character. However, for the English alphabet 'E' and Numeral '9' we have captured two different signs for each (that are used interchangeably), and it is contained in the folder namely E1 and E2 for alphabet 'E' and 9a and 9b for Numeral '9'. For the English alphabet, wherever a character is represented by a dynamic sign, the last frame of the sign is captured. For example, this is typically a case with English characters like 'J', 'H' and 'Y'. The images are stored in .jpeg format and have resolutions varying from 300 x 500 to 800 x 600, and the size is less than 100KB. The dataset is captured by a team pursuing research at Chandigarh College of Engineering and Technology, Chandigarh. The subjects have been informed about the research and Informed Participant Consent has been obtained prior to image acquisition. This dataset can be used only for research purposes either as it is or after cropping the static gestures from the image after duly referencing it. Any other use of the dataset is strictly prohibited and any illegal use is subject to the Indian court of law.
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
## Overview
Indian Sign Language is a dataset for object detection tasks - it contains Signs For Help annotations for 478 images.
## Getting Started
You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
## License
This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
y
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
The dataset includes videos files of the hand gestures of eight words (accident, call, doctor, help, hot, lose, pain, thief) from Indian sign language (ISL), commonly used to communicate during emergency situations. The data is useful for the researchers working on vision based automatic sign language recognition as well as hand gesture recognition.
All the words included in the dataset, except the word doctor are dynamic hand gestures. The videos in this dataset were collected by asking the participants to stand comfortably behind a black colored board and present the hand gestures, in front of the board. A Sony cyber shot DSC-W810 digital camera with 20.1 mega pixel resolution has been used for capturing the videos.
The videos have been collected from 26 individuals including 12 males and 14 females in the age group of 22 to 26 years. Two sample videos have been captured from each participant in an indoor environment under normal lighting conditions by placing the camera at a fixed distance. The dataset is presented in two folders with the original raw video sequences in one folder, and the cropped and downsampled video sequences in the other folder.
Facebook
TwitterThis dataset contains 36 classes representing Indian Sign Language (ISL) characters, including digits (0–9) and alphabets (A–Z). Each class has 1,000 images, resulting in a total of 36,000 labeled samples.
The dataset is designed to support research and development in:
Computer Vision: Hand gesture recognition
Deep Learning: Image classification and CNN-based models
Human-Computer Interaction: Enabling communication tools for the deaf and hard-of-hearing community
Sign Language Translation Systems
Dataset Details:
Classes: 36 (0–9, A–Z)
Images per class: 1,000(some might need pre-processing)
Format: JPG
Use cases: Training, validation, and testing of sign language recognition models
This dataset can be used to build, train, and benchmark machine learning models for gesture recognition tasks. It contributes to bridging the communication gap by empowering developers and researchers to create real-world applications such as sign language interpreters, accessibility tools, and educational platforms.
Facebook
TwitterThis dataset consists of the Indian sign language of all the alphabets and numbers in Indian hand recognition given by ISRTC(Indian Sign Research and Training Center). This dataset is in black white background for faster computing and for getting better accuracy while training the dataset.
Please give credit to this dataset if you download it.
Facebook
TwitterThis dataset was created by Harsh0239
Facebook
Twitterhttps://creativecommons.org/publicdomain/zero/1.0/https://creativecommons.org/publicdomain/zero/1.0/
This dataset contains Indian Sign Language (ISL) hand gesture images for alphabets A to Z. Each class is stored in separate folders and includes multiple image samples captured under consistent lighting conditions. The dataset is suitable for training machine learning and deep learning models such as CNN, LSTM, and hybrid architectures for sign language recognition.
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
This dataset contains depth data collected through Intel RealSense Depth Camera D435i. Data corresponding to Indian Sign Language (ISL) gesture of Weekdays (Sunday-Saturday) is used.
Data is stored as comma separated values. Each line corresponds to a sign.
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Sign language is a cardinal element for communication between deaf and dumb community. Sign language has its own grammatical structure and gesticulation nature. Research on SLRT focuses a lot of attention in gesture identification. Sign language comprises of manual gestures performed by hand poses and non-manual features expressed through eye, mouth and gaze movements. The sentence-level completely labelled Indian Sign Language dataset for Sign Language Translation and Recognition (SLTR) research is developed. The ISL-CSLTR dataset assists the research community to explore intuitive insights and to build the SLTR framework for establishing communication with the deaf and dumb community using advanced deep learning and computer vision methods for SLTR purposes. This ISL-CSLTR dataset aims in contributing to the sentence level dataset created with two native signers from Navajeevan, Residential School for the Deaf, College of Spl. D.Ed & B.Ed, Vocational Centre, and Child Care & Learning Centre, Ayyalurimetta, Andhra Pradesh, India and four student volunteers from SASTRA Deemed University, Thanjavur, Tamilnadu. The ISL-CSLTR corpus consists of a large vocabulary of 700 fully annotated videos, 18863 Sentence level frames, and 1036 word level images for 100 Spoken language Sentences performed by 7 different Signers. This corpus is arranged based on signer variants and time boundaries with fully annotated details and it is made available publicly. The main objective of creating this sentence level ISL-CSLRT corpus is to explore more research outcomes in the area of SLTR. This completely labelled video corpus assists the researchers to build framework for converting spoken language sentences into sign language and vice versa. This corpus has been created to address the various challenges faced by the researchers in SLRT and significantly improves translation and recognition performance. The videos are annotated with relevant spoken language sentences provide clear and easy understanding of the corpus data. Acknowledgements: The research was funded by the Science and Engineering Research Board (SERB), India under Start-up Research Grant (SRG)/2019–2021 (Grant no. SRG/2019/001338). And also, we thank all the signers for their contribution in collecting the sign videos and the successful completion of the ISL-CSLTR corpus. We would like to thank Navajeevan, Residential School for the Deaf, College of Spl. D.Ed & B.Ed, Vocational Centre, and Child Care & Learning Centre, Ayyalurimetta, Andhra Pradesh, India for their support and contribution.