Attribution-NonCommercial-NoDerivs 4.0 (CC BY-NC-ND 4.0)https://creativecommons.org/licenses/by-nc-nd/4.0/
License information was derived automatically
Dataset comprises 199,955 images featuring 28,565 individuals displaying a variety of facial expressions. It is designed for research in emotion recognition and facial expression analysis across diverse races, genders, and ages.
By utilizing this dataset, researchers and developers can enhance their understanding of facial recognition technology and improve the accuracy of emotion classification systems. - Get the data
https://www.googleapis.com/download/storage/v1/b/kaggle-user-content/o/inbox%2F22059654%2F22472a4de7d505ff4962b7eaa14071bf%2F1.png?generation=1740432470830146&alt=media" alt="">
This dataset includes images that capture different emotions, such as happiness, sadness, surprise, anger, disgust, and fear, allowing researchers to develop and evaluate recognition algorithms and detection methods.
https://www.googleapis.com/download/storage/v1/b/kaggle-user-content/o/inbox%2F22059654%2F8cfad327bf19d7f6fad22ae2cc021a5b%2FFrame%201%20(2).png?generation=1740432926933026&alt=media" alt="">
Researchers can leverage this dataset to explore various learning methods and algorithms aimed at improving emotion detection and facial expression recognition.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Computer Vision Scientist-Collected Dataset:This facial emotion recognition dataset has been meticulously curated by computer vision scientists using mobile phone cameras to capture candid moments of individuals expressing a spectrum of emotions, including Happy, Sad, Fear, and Humor. The dataset comprises a rich collection of images with diverse angles and backgrounds, providing a realistic portrayal of human emotional expression.Marketing Expert-Collected Dataset:The ethnicity-focused dataset for facial recognition has been meticulously assembled by marketing experts, aiming to shed light on the vital aspect of ethnicity variations in computer vision. With a dedicated focus on ethnicity, this dataset provides a unique perspective for training and testing facial recognition models in an ethnically diverse context.This dataset comprises a rich collection of images capturing individuals from various ethnic backgrounds. It emphasises the importance of ethnicity in the field of computer vision and includes a wide range of facial features, expressions, and poses, thereby enriching the dataset's diversity.By offering insights into the critical area of ethnicity in computer vision, this dataset is a valuable addition to the toolkit of researchers and practitioners, facilitating the development of more inclusive and accurate facial recognition models.Researchers and experts in the fields of computer vision and marketing are encouraged to explore these datasets for their research, model development, and the advancement of understanding in these respective domains.
https://www.futurebeeai.com/policies/ai-data-license-agreementhttps://www.futurebeeai.com/policies/ai-data-license-agreement
Welcome to the East Asian Facial Expression Image Dataset, curated to support the development of advanced facial expression recognition systems, biometric identification models, KYC verification processes, and a wide range of facial analysis applications. This dataset is ideal for training robust emotion-aware AI solutions.
The dataset includes over 2000 high-quality facial expression images, grouped into participant-wise sets. Each participant contributes:
To ensure generalizability and robustness in model training, images were captured under varied real-world conditions:
Each participant's image set is accompanied by detailed metadata, enabling precise filtering and training:
This metadata helps in building expression recognition models that are both accurate and inclusive.
This dataset is ideal for a variety of AI and computer vision applications, including:
To support evolving AI development needs, this dataset is regularly updated and can be tailored to project-specific requirements. Custom options include:
Facial Expression Recognition dataset helps AI interpret human emotions for improved sentiment analysis and recognition
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Interactive Facial Expression and Emotion Detection (IFEED) is an annotated dataset that can be used to train, validate, and test Deep Learning models for facial expression and emotion recognition. It contains pre-filtered and analysed images of the interactions between the six main characters of the Friends television series, obtained from the video recordings of the Multimodal EmotionLines Dataset (MELD).
The images were obtained by decomposing the videos into multiple frames and extracting the facial expression of the correctly identified characters. A team composed of 14 researchers manually verified and annotated the processed data into several classes: Angry, Sad, Happy, Fearful, Disgusted, Surprised and Neutral.
IFEED can be valuable for the development of intelligent facial expression recognition solutions and emotion detection software, enabling binary or multi-class classification, or even anomaly detection or clustering tasks. The images with ambiguous or very subtle facial expressions can be repurposed for adversarial learning. The dataset can be combined with additional data recordings to create more complete and extensive datasets and improve the generalization of robust deep learning models.
https://www.futurebeeai.com/policies/ai-data-license-agreementhttps://www.futurebeeai.com/policies/ai-data-license-agreement
Welcome to the African Facial Expression Image Dataset, curated to support the development of advanced facial expression recognition systems, biometric identification models, KYC verification processes, and a wide range of facial analysis applications. This dataset is ideal for training robust emotion-aware AI solutions.
The dataset includes over 2000 high-quality facial expression images, grouped into participant-wise sets. Each participant contributes:
To ensure generalizability and robustness in model training, images were captured under varied real-world conditions:
Each participant's image set is accompanied by detailed metadata, enabling precise filtering and training:
This metadata helps in building expression recognition models that are both accurate and inclusive.
This dataset is ideal for a variety of AI and computer vision applications, including:
To support evolving AI development needs, this dataset is regularly updated and can be tailored to project-specific requirements. Custom options include:
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
The Facial Emotion Detection Dataset is a collection of images of individuals with two different emotions - happy and sad. The dataset was captured using a mobile phone camera and contains photos taken from different angles and backgrounds.
The dataset contains a total of 637 photos with an additional dataset of 127 from previous work. Out of the total, 402 images are of happy faces, and 366 images are of sad faces. Each individual had a minimum of 10 images of both expressions.
The project faced challenges in terms of time constraints and people's constraints, which limited the number of individuals who participated. Despite the limitations, the dataset can be used for deep learning projects and real-time emotion detection models. Future work can expand the dataset by capturing more images to improve the accuracy of the model. The dataset can also be used to create a custom object detection model to evaluate other types of emotional expressions.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
etc.)
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Emotion Detection Model for Facial Expressions
Project Description:
In this project, we developed an Emotion Detection Model using a curated dataset of 715 facial images, aiming to accurately recognize and categorize expressions into five distinct emotion classes. The emotion classes include Happy, Sad, Fearful, Angry, and Neutral.
Objectives: - Train a robust machine learning model capable of accurately detecting and classifying facial expressions in real-time. - Implement emotion detection to enhance user experience in applications such as human-computer interaction, virtual assistants, and emotion-aware systems.
Methodology: 1. Data Collection and Preprocessing: - Assembled a diverse dataset of 715 images featuring individuals expressing different emotions. - Employed Roboflow for efficient data preprocessing, handling image augmentation and normalization.
Model Architecture:
Training and Validation:
Model Evaluation:
Deployment and Integration:
Results: The developed Emotion Detection Model demonstrates high accuracy in recognizing and classifying facial expressions across the defined emotion classes. This project lays the foundation for integrating emotion-aware systems into various applications, fostering more intuitive and responsive interactions.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
## Overview
Facial Expression Recognition is a dataset for classification tasks - it contains Emotions annotations for 7,939 images.
## Getting Started
You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
## License
This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).
CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
License information was derived automatically
40 sub-folders are further divided in this directory, each sub-folder contains the data of all the facial expression per participant. The sub-folders are named after the participant ID and include 4 sub-sub folders which are central RGB (CRGB) facial expression data, left RGB (LRGB) facial expression data, right RGB (RRGB) facial expression data, and central infrared (CIR) facial expression data. Each folder contains multiple MP4 files, and each MP4 file corresponds to valid emotional driving.
The JAFFE images may be used only for non-commercial scientific research.
The source and background of the dataset must be acknowledged by citing the following two articles. Users should read both carefully.
Michael J. Lyons, Miyuki Kamachi, Jiro Gyoba.
Coding Facial Expressions with Gabor Wavelets (IVC Special Issue)
arXiv:2009.05938 (2020) https://arxiv.org/pdf/2009.05938.pdf
Michael J. Lyons
"Excavating AI" Re-excavated: Debunking a Fallacious Account of the JAFFE Dataset
arXiv: 2107.13998 (2021) https://arxiv.org/abs/2107.13998
The following is not allowed:
A few sample images (not more than 10) may be displayed in scientific publications.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Facial Expression Recognition DatasetThis dataset supports research on facial expression recognition using visible and infrared modalities. It includes data for various facial expressions from two publicly available datasets: VIRI (five expressions: angry, happy, neutral, sad, and surprised) and NVIE (three expressions: fear, disgust, and happy). The dataset has been processed and prepared for training and evaluation of machine learning models.The dataset is designed for use with deep learning frameworks like PyTorch and supports experiments in feature extraction, model evaluation, and early fusion approaches for visible and infrared modalities.For details on the methodology, preprocessing steps, and evaluation metrics, please refer to the linked GitHub repository: https://github.com/naseemmuhammadtahir/raw-data.This dataset facilitates reproducibility and exploration of advanced models for facial expression recognition tasks in diverse modalities.
disgust
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
This dataset is a meticulously curated dataset designed for infant facial emotion recognition, featuring four primary emotional expressions: Angry, Cry, Laugh, and Normal. The dataset aims to facilitate research in machine learning, deep learning, affective computing, and human-computer interaction by providing a large collection of labeled infant facial images.
Primary Data (1600 Images): - Angry: 400 - Cry: 400 - Laugh: 400 - Normal: 400
Data Augmentation & Expanded Dataset (26,143 Images): To enhance the dataset's robustness and expand the dataset, 20 augmentation techniques (including HorizontalFlip, VerticalFlip, Rotate, ShiftScaleRotate, BrightnessContrast, GaussNoise, GaussianBlur, Sharpen, HueSaturationValue, CLAHE, GridDistortion, ElasticTransform, GammaCorrection, MotionBlur, ColorJitter, Emboss, Equalize, Posterize, FogEffect, and RainEffect) were applied randomly. This resulted in a significantly larger dataset with:
Data Collection & Ethical Considerations: The dataset was collected under strict ethical guidelines to ensure compliance with privacy and data protection laws. Key ethical considerations include: 1. Ethical Approval: The study was reviewed and approved by the Institutional Review Board (IRB) of Daffodil International University under Reference No: REC-FSIT-2024-11-10. 2. Informed Parental Consent: Written consent was obtained from parents before capturing and utilizing infant facial images for research purposes. 3. Privacy Protection: No personally identifiable information (PII) is included in the dataset, and images are strictly used for research in AI-driven emotion recognition.
Data Collection Locations & Geographical Diversity: To ensure diversity in infant facial expressions, data collection was conducted across multiple locations in Bangladesh, covering healthcare centers and educational institutions:
Face Detection Methodology: To extract the facial regions efficiently, RetinaNet—a deep learning-based object detection model—was employed. The use of RetinaNet ensures precise facial cropping while minimizing background noise and occlusions.
Potential Applications: 1. Affective Computing: Understanding infant emotions for smart healthcare and early childhood development. 2. Computer Vision: Training deep learning models for automated infant facial expression recognition. 3. Pediatric & Mental Health Research: Assisting in early autism screening and emotion-aware AI for child psychology. 4. Human-Computer Interaction (HCI): Designing AI-powered assistive technologies for infants.
CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
License information was derived automatically
The Indoor Facial 75 Expressions Dataset enriches the internet, media, entertainment, and mobile sectors with an in-depth exploration of human emotions. It features 60 individuals in indoor settings, showcasing a balanced gender representation and varied postures, with 75 distinct facial expressions per person. This dataset is tagged with facial expression categories, making it an invaluable tool for emotion recognition and interactive applications.
IntroductionPerceptual learning of facial expression is shown specific to the train expression, indicating separate encoding of the emotional contents in different expressions. However, little is known about the specificity of emotional recognition training with the visual search paradigm and the sensitivity of learning to near-threshold stimuli.MethodsIn the present study, we adopted a visual search paradigm to measure the recognition of facial expressions. In Experiment 1 (Exp1), Experiment 2 (Exp2), and Experiment 3 (Exp3), subjects were trained for 8 days to search for a target expression in an array of faces presented for 950 ms, 350 ms, and 50 ms, respectively. In Experiment 4 (Exp4), we trained subjects to search for a target of a triangle, and tested them with the task of facial expression search. Before and after the training, subjects were tested on the trained and untrained facial expressions which were presented for 950 ms, 650 ms, 350 ms, or 50 ms.ResultsThe results showed that training led to large improvements in the recognition of facial emotions only if the faces were presented long enough (Exp1: 85.89%; Exp2: 46.05%). Furthermore, the training effect could transfer to the untrained expression. However, when the faces were presented briefly (Exp3), the training effect was small (6.38%). In Exp4, the results indicated that the training effect could not transfer across categories.DiscussionOur findings revealed cross-emotion transfer for facial expression recognition training in a visual search task. In addition, learning hardly affects the recognition of near-threshold expressions.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Recognition rate of the proposed FER system using IMFDB dataset of facial expressions (Unit: %).
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
a good solution is to augment the DBs with appropriate techniques
Video dataset capturing diverse facial expressions and emotions from 1000+ people, suitable for emotion recognition AI training
Attribution-NonCommercial-NoDerivs 4.0 (CC BY-NC-ND 4.0)https://creativecommons.org/licenses/by-nc-nd/4.0/
License information was derived automatically
Dataset comprises 199,955 images featuring 28,565 individuals displaying a variety of facial expressions. It is designed for research in emotion recognition and facial expression analysis across diverse races, genders, and ages.
By utilizing this dataset, researchers and developers can enhance their understanding of facial recognition technology and improve the accuracy of emotion classification systems. - Get the data
https://www.googleapis.com/download/storage/v1/b/kaggle-user-content/o/inbox%2F22059654%2F22472a4de7d505ff4962b7eaa14071bf%2F1.png?generation=1740432470830146&alt=media" alt="">
This dataset includes images that capture different emotions, such as happiness, sadness, surprise, anger, disgust, and fear, allowing researchers to develop and evaluate recognition algorithms and detection methods.
https://www.googleapis.com/download/storage/v1/b/kaggle-user-content/o/inbox%2F22059654%2F8cfad327bf19d7f6fad22ae2cc021a5b%2FFrame%201%20(2).png?generation=1740432926933026&alt=media" alt="">
Researchers can leverage this dataset to explore various learning methods and algorithms aimed at improving emotion detection and facial expression recognition.