100+ datasets found
  1. h

    facial-expression-recognition-dataset

    • huggingface.co
    Updated Mar 31, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Unidata (2025). facial-expression-recognition-dataset [Dataset]. https://huggingface.co/datasets/UniDataPro/facial-expression-recognition-dataset
    Explore at:
    Dataset updated
    Mar 31, 2025
    Authors
    Unidata
    License

    Attribution-NonCommercial-NoDerivs 4.0 (CC BY-NC-ND 4.0)https://creativecommons.org/licenses/by-nc-nd/4.0/
    License information was derived automatically

    Description

    Emotion recognition Dataset

    Dataset comprises 199,955 images featuring 28,565 individuals displaying a variety of facial expressions. It is designed for research in emotion recognition and facial expression analysis across diverse races, genders, and ages. By utilizing this dataset, researchers and developers can enhance their understanding of facial recognition technology and improve the accuracy of emotion classification systems. - Get the data

      Examples of data
    

    This… See the full description on the dataset page: https://huggingface.co/datasets/UniDataPro/facial-expression-recognition-dataset.

  2. F

    South Asian Facial Expression Image Dataset

    • futurebeeai.com
    wav
    Updated Aug 1, 2022
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    FutureBee AI (2022). South Asian Facial Expression Image Dataset [Dataset]. https://www.futurebeeai.com/dataset/image-dataset/facial-images-expression-south-asian
    Explore at:
    wavAvailable download formats
    Dataset updated
    Aug 1, 2022
    Dataset provided by
    FutureBeeAI
    Authors
    FutureBee AI
    License

    https://www.futurebeeai.com/policies/ai-data-license-agreementhttps://www.futurebeeai.com/policies/ai-data-license-agreement

    Area covered
    South Asia
    Dataset funded by
    FutureBeeAI
    Description

    Introduction

    Welcome to the South Asian Facial Expression Image Dataset, curated to support the development of advanced facial expression recognition systems, biometric identification models, KYC verification processes, and a wide range of facial analysis applications. This dataset is ideal for training robust emotion-aware AI solutions.

    Facial Expression Data

    The dataset includes over 2000 high-quality facial expression images, grouped into participant-wise sets. Each participant contributes:

    Expression Images: 5 distinct facial images capturing common human emotions: Happy, Sad, Angry, Shocked, and Neutral

    Diversity & Representation

    Geographical Coverage: Individuals from South Asian countries including India, Pakistan, Bangladesh, Nepal, Sri Lanka, Bhutan, Maldives, and more
    Demographics: Participants aged 18 to 70 years, with a gender distribution of 60% male and 40% female
    File Formats: All images are available in JPEG and HEIC formats

    Image Quality & Capture Conditions

    To ensure generalizability and robustness in model training, images were captured under varied real-world conditions:

    Lighting Conditions: Natural and artificial lighting to represent diverse scenarios
    Background Variability: Indoor and outdoor backgrounds to enhance model adaptability
    Device Quality: Captured using modern smartphones to ensure clarity and consistency

    Metadata

    Each participant's image set is accompanied by detailed metadata, enabling precise filtering and training:

    Unique Participant ID
    File Name
    Age
    Gender
    Country
    Facial Expression Label
    Demographic Information
    File Format

    This metadata helps in building expression recognition models that are both accurate and inclusive.

    Use Cases & Applications

    This dataset is ideal for a variety of AI and computer vision applications, including:

    Facial Expression Recognition: Improve accuracy in detecting emotions like happiness, anger, or surprise
    Biometric & Identity Systems: Enhance facial biometric authentication with expression variation handling
    KYC & Identity Verification: Validate facial consistency in ID documents and selfies despite varied expressions
    Generative AI Training: Support expression generation and animation in AI-generated facial images
    Emotion-Aware Systems: Power human-computer interaction, mental health assessment, and adaptive learning apps

    Secure & Ethical Collection

    Data Security: All data is securely processed and stored on FutureBeeAI’s proprietary platform
    Ethical Standards: Collection followed strict ethical guidelines ensuring participant privacy and informed consent
    Informed Consent: All participants were made aware of the data use and provided written consent

    Dataset Updates & Customization

    To support evolving AI development needs, this dataset is regularly updated and can be tailored to project-specific requirements. Custom options include:

    <div style="margin-top:10px; margin-bottom: 10px; padding-left: 30px; display: flex; gap: 16px; align-items:

  3. f

    The MPI Facial Expression Database — A Validated Database of Emotional and...

    • plos.figshare.com
    • datasetcatalog.nlm.nih.gov
    pdf
    Updated Jun 1, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Kathrin Kaulard; Douglas W. Cunningham; Heinrich H. Bülthoff; Christian Wallraven (2023). The MPI Facial Expression Database — A Validated Database of Emotional and Conversational Facial Expressions [Dataset]. http://doi.org/10.1371/journal.pone.0032321
    Explore at:
    pdfAvailable download formats
    Dataset updated
    Jun 1, 2023
    Dataset provided by
    PLOS ONE
    Authors
    Kathrin Kaulard; Douglas W. Cunningham; Heinrich H. Bülthoff; Christian Wallraven
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    The ability to communicate is one of the core aspects of human life. For this, we use not only verbal but also nonverbal signals of remarkable complexity. Among the latter, facial expressions belong to the most important information channels. Despite the large variety of facial expressions we use in daily life, research on facial expressions has so far mostly focused on the emotional aspect. Consequently, most databases of facial expressions available to the research community also include only emotional expressions, neglecting the largely unexplored aspect of conversational expressions. To fill this gap, we present the MPI facial expression database, which contains a large variety of natural emotional and conversational expressions. The database contains 55 different facial expressions performed by 19 German participants. Expressions were elicited with the help of a method-acting protocol, which guarantees both well-defined and natural facial expressions. The method-acting protocol was based on every-day scenarios, which are used to define the necessary context information for each expression. All facial expressions are available in three repetitions, in two intensities, as well as from three different camera angles. A detailed frame annotation is provided, from which a dynamic and a static version of the database have been created. In addition to describing the database in detail, we also present the results of an experiment with two conditions that serve to validate the context scenarios as well as the naturalness and recognizability of the video sequences. Our results provide clear evidence that conversational expressions can be recognized surprisingly well from visual information alone. The MPI facial expression database will enable researchers from different research fields (including the perceptual and cognitive sciences, but also affective computing, as well as computer vision) to investigate the processing of a wider range of natural facial expressions.

  4. f

    Table_3_East Asian Young and Older Adult Perceptions of Emotional Faces From...

    • frontiersin.figshare.com
    xlsx
    Updated Jun 4, 2023
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Yu-Zhen Tu; Dong-Wei Lin; Atsunobu Suzuki; Joshua Oon Soo Goh (2023). Table_3_East Asian Young and Older Adult Perceptions of Emotional Faces From an Age- and Sex-Fair East Asian Facial Expression Database.XLSX [Dataset]. http://doi.org/10.3389/fpsyg.2018.02358.s005
    Explore at:
    xlsxAvailable download formats
    Dataset updated
    Jun 4, 2023
    Dataset provided by
    Frontiers
    Authors
    Yu-Zhen Tu; Dong-Wei Lin; Atsunobu Suzuki; Joshua Oon Soo Goh
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    There is increasing interest in clarifying how different face emotion expressions are perceived by people from different cultures, of different ages and sex. However, scant availability of well-controlled emotional face stimuli from non-Western populations limit the evaluation of cultural differences in face emotion perception and how this might be modulated by age and sex differences. We present a database of East Asian face expression stimuli, enacted by young and older, male and female, Taiwanese using the Facial Action Coding System (FACS). Combined with a prior database, this present database consists of 90 identities with happy, sad, angry, fearful, disgusted, surprised and neutral expressions amounting to 628 photographs. Twenty young and 24 older East Asian raters scored the photographs for intensities of multiple-dimensions of emotions and induced affect. Multivariate analyses characterized the dimensionality of perceived emotions and quantified effects of age and sex. We also applied commercial software to extract computer-based metrics of emotions in photographs. Taiwanese raters perceived happy faces as one category, sad, angry, and disgusted expressions as one category, and fearful and surprised expressions as one category. Younger females were more sensitive to face emotions than younger males. Whereas, older males showed reduced face emotion sensitivity, older female sensitivity was similar or accentuated relative to young females. Commercial software dissociated six emotions according to the FACS demonstrating that defining visual features were present. Our findings show that East Asians perceive a different dimensionality of emotions than Western-based definitions in face recognition software, regardless of age and sex. Critically, stimuli with detailed cultural norms are indispensable in interpreting neural and behavioral responses involving human facial expression processing. To this end, we add to the tools, which are available upon request, for conducting such research.

  5. i

    Facial Expression Dataset (Sri Lankan)

    • ieee-dataport.org
    Updated Jul 29, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Amod Pathirana (2025). Facial Expression Dataset (Sri Lankan) [Dataset]. https://ieee-dataport.org/documents/facial-expression-dataset-sri-lankan
    Explore at:
    Dataset updated
    Jul 29, 2025
    Authors
    Amod Pathirana
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Area covered
    Sri Lanka
    Description

    often based on foreign samples

  6. Z

    IFEED: Interactive Facial Expression and Emotion Detection Dataset

    • data.niaid.nih.gov
    • zenodo.org
    Updated May 26, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Oliveira, Jorge (2023). IFEED: Interactive Facial Expression and Emotion Detection Dataset [Dataset]. https://data.niaid.nih.gov/resources?id=zenodo_7963451
    Explore at:
    Dataset updated
    May 26, 2023
    Dataset provided by
    Oliveira, Nuno
    Oliveira, Jorge
    Praça, Isabel
    Vitorino, João
    Dias, Tiago
    Maia, Eva
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Interactive Facial Expression and Emotion Detection (IFEED) is an annotated dataset that can be used to train, validate, and test Deep Learning models for facial expression and emotion recognition. It contains pre-filtered and analysed images of the interactions between the six main characters of the Friends television series, obtained from the video recordings of the Multimodal EmotionLines Dataset (MELD).

    The images were obtained by decomposing the videos into multiple frames and extracting the facial expression of the correctly identified characters. A team composed of 14 researchers manually verified and annotated the processed data into several classes: Angry, Sad, Happy, Fearful, Disgusted, Surprised and Neutral.

    IFEED can be valuable for the development of intelligent facial expression recognition solutions and emotion detection software, enabling binary or multi-class classification, or even anomaly detection or clustering tasks. The images with ambiguous or very subtle facial expressions can be repurposed for adversarial learning. The dataset can be combined with additional data recordings to create more complete and extensive datasets and improve the generalization of robust deep learning models.

  7. u

    Facial Expression Recognition Dataset

    • unidata.pro
    jpg/jpeg, png
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Unidata L.L.C-FZ, Facial Expression Recognition Dataset [Dataset]. https://unidata.pro/datasets/facial-expression-recognition-dataset/
    Explore at:
    jpg/jpeg, pngAvailable download formats
    Dataset authored and provided by
    Unidata L.L.C-FZ
    Description

    Facial Expression Recognition dataset helps AI interpret human emotions for improved sentiment analysis and recognition

  8. Facial Expression Recognition Dataset – 1,142 People, 7 Emotions, Online...

    • nexdata.ai
    Updated Oct 8, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Nexdata (2023). Facial Expression Recognition Dataset – 1,142 People, 7 Emotions, Online Conference Scenes [Dataset]. https://www.nexdata.ai/datasets/computervision/1281
    Explore at:
    Dataset updated
    Oct 8, 2023
    Dataset authored and provided by
    Nexdata
    Variables measured
    Data size, Data format, Accuracy rate, Age distribution, Race distribution, Collection content, Gender distribution, Collection diversity, Collection equipment, Collection environment
    Description

    This dataset contains facial expression recognition data from 1,142 people in online conference scenes. Participants include Asian, Caucasian, Black, and Brown individuals, mainly young and middle-aged adults. Data was collected across a variety of indoor office scenes, covering meeting rooms, coffee shops, libraries , bedroom, etc., Each participant performed seven key expressions: normal, happy, surprised, sad, angry, disgusted, and fearful. The dataset is suitable for tasks such as facial expression recognition, emotion recognition, human-computer interaction, and video conferencing AI applications.

  9. g

    Facial Expression Image Data AFFECTNET YOLO Format

    • gts.ai
    json
    Updated Mar 20, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    GLOBOSE TECHNOLOGY SOLUTIONS PRIVATE LIMITED (2024). Facial Expression Image Data AFFECTNET YOLO Format [Dataset]. https://gts.ai/dataset-download/facial-expression-image-data-affectnet-yolo-format/
    Explore at:
    jsonAvailable download formats
    Dataset updated
    Mar 20, 2024
    Dataset authored and provided by
    GLOBOSE TECHNOLOGY SOLUTIONS PRIVATE LIMITED
    License

    CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
    License information was derived automatically

    Description

    A comprehensive facial expression dataset containing 0.4 million images labeled with 8 emotions (neutral, happy, angry, sad, fear, surprise, disgust, contempt) including valence and arousal scores. Images are resized to 96x96 pixels and organized for YOLO training with separate folders for training, testing, and validation.

  10. Facial Expression Data in Online Conference Scenes

    • kaggle.com
    Updated Jun 7, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Frank Wong (2024). Facial Expression Data in Online Conference Scenes [Dataset]. https://www.kaggle.com/datasets/nexdatafrank/facial-expression-data-in-online-conference-scenes
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Jun 7, 2024
    Dataset provided by
    Kagglehttp://kaggle.com/
    Authors
    Frank Wong
    License

    Attribution-NonCommercial 4.0 (CC BY-NC 4.0)https://creativecommons.org/licenses/by-nc/4.0/
    License information was derived automatically

    Description

    Facial Expression Data in Online Conference Scenes

    Description

    1,142 people facial expression data in online conference scenes, including Asian, Caucasian, black, and brown multitasking, mainly young and middle-aged, collected a variety of indoor office scenes, covering meeting rooms, coffee shops, libraries , bedroom, etc., each collector collected 7 kinds of expressions: normal, happy, surprised, sad, angry, disgusted, and fearful. For more details, please refer to the link: https://www.nexdata.ai/datasets/computervision/1281?source=Kaggle

    Data size

    1,142 people, each person collects 7 videos

    Race distribution

    153 Asians, 889 Caucasians, 66 blacks, 34 brown people

    Gender distribution

    535 males, 607 females

    Age distribution

    from teenagers to the elderly, mainly young and middle-aged

    Collection environment

    indoor office scenes, such as meeting rooms, coffee shops, libraries, bedrooms, etc.

    Collection diversity

    different facial expressions, different races, different age groups, different meeting scenes

    Collection equipment

    cellphone, using the cellphone to simulate the perspective of the laptop camera in online conference scenes

    Collection content

    collecting the expression data in online conference scenes

    Data format

    .mp4, .mov

    Accuracy rate

    the accuracy exceeds 97% based on the accuracy of the expressions; the accuracy of expression naming

    Licensing Information

    Commercial License

  11. Z

    Facial Expression and Landmark Tracking (FELT) dataset

    • data.niaid.nih.gov
    Updated Oct 19, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Liao, Zhenghao; Livingstone, Steven; Russo, Frank A. (2024). Facial Expression and Landmark Tracking (FELT) dataset [Dataset]. https://data.niaid.nih.gov/resources?id=zenodo_13243599
    Explore at:
    Dataset updated
    Oct 19, 2024
    Dataset provided by
    Ontario Tech University
    Toronto Metropolitan University
    Authors
    Liao, Zhenghao; Livingstone, Steven; Russo, Frank A.
    License

    Attribution-NonCommercial-ShareAlike 4.0 (CC BY-NC-SA 4.0)https://creativecommons.org/licenses/by-nc-sa/4.0/
    License information was derived automatically

    Description

    Contact Information

    If you would like further information about the Facial expression and landmark tracking data set, or if you experience any issues downloading files, please contact us at ravdess@gmail.com.

    Facial Expression examples

    Watch a sample of the facial expression tracking results.

    Commercial Licenses

    Commercial licenses for this dataset can be purchased. For more information, please contact us at ravdess@gmail.com.

    Description

    The Facial Expression and Landmark Tracking (FELT) dataset dataset contains tracked facial expression movements and animated videos from the Ryerson Audio-Visual Database of Emotional Speech and Song (RAVDESS) [RAVDESS Zenodo page]. Tracking data and videos were produced by Py-Feat 0.6.2 (2024-03-29 release) (Cheong, J.H., Jolly, E., Xie, T. et al. Py-Feat: Python Facial Expression Analysis Toolbox. Affec Sci 4, 781–796 (2023). https://doi.org/10.1007/s42761-023-00191-4) and custom code (github repo). Tracked information includes: facial emotion classification estimates, facial landmark detection (68 points), head pose estimation (yaw, pitch, roll, x, y), and facial Action Unit (AU) recognition. Videos include: landmark overlay videos, AU activation animations, and landmark plot animations.

    The FELT dataset was created at the Affective Data Science Lab.

    This dataset contains tracking data and videos for all 2452 RAVDESS trials. Raw and smoothed tracking data are provided. All tracking movement data are contained in the following archives: raw_motion_speech.zip, smoothed_motion_speech.zip, raw_motion_song.zip, and smoothed_motion_song.zip. Each actor has 104 tracked trials (60 speech, 44 song). Note, there are no song files for Actor 18.

    Total Tracked Files = (24 Actors x 60 Speech trials) + (23 Actors x 44 Song trials) = 2452 CSV files.

    Tracking results for each trial are provided as individual comma separated value files (CSV format). File naming convention of raw and smoothed tracked files is identical to that of the RAVDESS. For example, smoothed tracked file "01-01-01-01-01-01-01.csv" corresponds to RAVDESS audio-video file "01-01-01-01-01-01-01.mp4". For a complete description of the RAVDESS file naming convention and experimental manipulations, please see the RAVDESS Zenodo page.

    Landmark overlays, AU activation, and landmark plot videos for all trials are also provided (720p h264, .mp4). Landmark overlays present tracked landmarks and head pose overlaid on the original RAVDESS actor video. As the RAVDESS does not contain "ground truth" facial landmark locations, the overlay videos provide a visual 'sanity check' for researchers to confirm the general accuracy of the tracking results. Landmark plot animations present landmarks only, anchored to the top left corner of the head bounding box with translational head motion removed. AU activation animations visualize intensity of AU activations (0-1 normalized) as a heatmap over time. The file naming convention of all videos also matches that of the RAVDESS. For example, "Landmark_Overlay/01-01-01-01-01-01-01.mp4", "Landmark_Plot/01-01-01-01-01-01-01.mp4", "ActionUnit_Animation/01-01-01-01-01-01-01.mp4", all correspond to RAVDESS audio-video file "01-01-01-01-01-01-01.mp4".

    Smoothing procedure

    Raw tracking data were first low-pass filtered with a 5th order butterworth filter (cutoff_freq = 6, sampling_freq = 29.97, order = 5) to remove high-frequency noise. Data were then smoothed with a Savitzky-Golay filter (window_length = 11, poly_order = 5). Scipy.signal (v 1.13.1) was used for both procedures.

    Landmark Tracking models

    Six separate machine learning models were used by Py-Feat to perform various aspects of tracking and classification. Video outputs generated by different combinations of ML models were visually compared, with final model choice determined by voting of first and second authors. Models were specified in the call to Detector class (described here). Exact function call as follows:

    Detector(face_model='img2pose',  landmark_model='mobilenet',  au_model='xgb',  emotion_model='resmasknet',  facepose_model='img2pose-c',  identity_model='facenet',  device='cuda',  n_jobs=1,  verbose=False,  )
    

    Default Py_feat parameters to each model were used in most cases. Non-defaults were specified in the call to detect_video function (described here). Exact function call as follows: (video_path, skip_frames=None, output_size=(720, 1280), batch_size=5, num_workers=0, pin_memory=False, face_detection_threshold=0.83, face_identity_threshold=0.8 )

    Tracking File Output Format

    This data set retained Py-Feat's data output format. The resolution of all input videos was 1280x720. Tracking output units are in pixels, their range of values is (0,0) (top left corner) to (1280,720) (bottom right corner).

    Column 1 = Timing information

    1. frame - The number of the frame (source videos 29.97 fps), range = 1 to n

    Columns 2-5 = Head bounding box

    2-3. FaceRectX, FaceRectY - X and Y coordinates of top-left corner of head bounding box (pixels)

    4-5. FaceRectWidth, FaceRectHeightF - Width and Height of head bounding box (pixels)

    Column 6 = Face detection confidence

    FaceScore - Confidence level that a human face was deteceted, range = 0 to 1

    Columns 7-142 = Facial landmark locations in 2D

    7-142. x_0, ..., x_67, y_0,...y_67 - Location of 2D landmarks in pixels. A figure describing the landmark index can be found here.

    Columns 143-145 = Head pose

    143-145. Pitch, Roll, Yaw - Rotation of the head in degrees (described here). The rotation is in world coordinates with the camera being located at the origin.

    Columns 146-165 = Facial Action Units

    Facial Action Units (AUs) are a way to describe human facial movements (Ekman, Friesen, and Hager, 2002) [wiki link]. More information on Py-Feat's implementation of AUs can be found here.

    145-150, 152-153, 155-158, 160-165. AU01, AU02, AU04, AU05, AU06, AU09, AU10, AU12, AU14, AU15, AU17, AU23, AU24, AU25, AU26, AU28, AU43 - Intensity of AU movement, range from 0 (no muscle contraction) to 1 (maximal muscle contraction).

    151, 154, 159. AU07, AU11, AU20 - Presence or absence of AUs, range 0 (absent, not detected) to 1 (present, detected).

    Columns 166-172 = Emotion classification confidence

    162-172. anger, disgust, fear, happiness, sadness, surprise, neutral - Confidence of classified emotion category, range 0 (0%) to 1 (100%) confidence.

    Columns 173-685 = Face identity score

    Identity of faces contained in the video were classified using the FaceNet model (described here). This procedure generates at 512 dimension Euclidean embedding space.

    1. Identity - Predicated individual identifyed in the RAVDESS video. Note, value is always Person_0, as each video only contains a single actor at all times (categorical).

    174-685. Identity_1, ..., Identity_512 - Face embedding vector used by FaceNet to perform facial identity matching.

    Column 686 = Input video

    1. frame - The number of the frame (source videos 29.97 fps), range = 1 to n

    Columns 687-688 = Timing information

    1. frame.1 - The number of the frame (source videos 29.97 fps), duplicated column, range = 1 to n

    2. approx_time - Approximate time of current frame (0.0 to x.x, in seconds)

    Tracking videos

    Landmark Overlay and Landmark Plot videos were produced with plot_detections function call (described here). This function generated invidual images for each frame, which were then compiled into a video using the imageio library (described here).

    AU Activation videos were produced with plot_face function call (described here). This function also generated invidual images for each frame, which were then compiled into a video using the imageio library. Some frames could not be correctly generated by Py-Feat, producing only the AU heatmap but failing to plot/locate facial landmarks. These frames were dropped prior to compositing the output video. Drop rate was approximately 10% of all frames, in each video. Dropped frames were distributed evenly across the video timeline (i.e. no apparent clustering).

    License information

    The RAVDESS Facial expression and landmark tracking data set is released under a Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License, CC BY-NA-SC 4.0.

    How to cite the RAVDESS Facial Tracking data set

    Academic citation If you use the RAVDESS Facial Tracking data set in an academic publication, please cite both references:

    Liao, Z., Livingstone, SR., & Russo, FA. (2024). RAVDESS Facial expression and landmark tracking (Version 1.0.0) [Data set]. Zenodo. http://doi.org/10.5281/zenodo.13243600

    Livingstone SR, Russo FA (2018). The Ryerson Audio-Visual Database of Emotional Speech and Song (RAVDESS): A dynamic, multimodal set of facial and vocal expressions in North American English. PLoS ONE 13(5): e0196391. https://doi.org/10.1371/journal.pone.0196391.

    All other attributions If you use the RAVDESS Facial expression and landmark tracking dataset in a form other than an academic publication, such as in a blog post, data science project or competition, school project, or non-commercial product, please use the following attribution: "RAVDESS Facial expression and landmark tracking" by Liao, Livingstone, & Russo is licensed under CC BY-NA-SC 4.0.

    Related Data sets

    The Ryerson Audio-Visual Database of Emotional Speech and Song [Zenodo project page].

  12. F

    African Facial Expression Image Dataset

    • futurebeeai.com
    wav
    Updated Aug 1, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    FutureBee AI (2022). African Facial Expression Image Dataset [Dataset]. https://www.futurebeeai.com/dataset/image-dataset/facial-images-expression-african
    Explore at:
    wavAvailable download formats
    Dataset updated
    Aug 1, 2022
    Dataset provided by
    FutureBeeAI
    Authors
    FutureBee AI
    License

    https://www.futurebeeai.com/policies/ai-data-license-agreementhttps://www.futurebeeai.com/policies/ai-data-license-agreement

    Dataset funded by
    FutureBeeAI
    Description

    Introduction

    Welcome to the African Facial Expression Image Dataset, curated to support the development of advanced facial expression recognition systems, biometric identification models, KYC verification processes, and a wide range of facial analysis applications. This dataset is ideal for training robust emotion-aware AI solutions.

    Facial Expression Data

    The dataset includes over 2000 high-quality facial expression images, grouped into participant-wise sets. Each participant contributes:

    Expression Images: 5 distinct facial images capturing common human emotions: Happy, Sad, Angry, Shocked, and Neutral

    Diversity & Representation

    Geographical Coverage: Individuals from African countries including Kenya, Malawi, Nigeria, Ethiopia, Benin, Somalia, Uganda, and more
    Demographics: Participants aged 18 to 70 years, with a gender distribution of 60% male and 40% female
    File Formats: All images are available in JPEG and HEIC formats

    Image Quality & Capture Conditions

    To ensure generalizability and robustness in model training, images were captured under varied real-world conditions:

    Lighting Conditions: Natural and artificial lighting to represent diverse scenarios
    Background Variability: Indoor and outdoor backgrounds to enhance model adaptability
    Device Quality: Captured using modern smartphones to ensure clarity and consistency

    Metadata

    Each participant's image set is accompanied by detailed metadata, enabling precise filtering and training:

    Unique Participant ID
    File Name
    Age
    Gender
    Country
    Facial Expression Label
    Demographic Information
    File Format

    This metadata helps in building expression recognition models that are both accurate and inclusive.

    Use Cases & Applications

    This dataset is ideal for a variety of AI and computer vision applications, including:

    Facial Expression Recognition: Improve accuracy in detecting emotions like happiness, anger, or surprise
    Biometric & Identity Systems: Enhance facial biometric authentication with expression variation handling
    KYC & Identity Verification: Validate facial consistency in ID documents and selfies despite varied expressions
    Generative AI Training: Support expression generation and animation in AI-generated facial images
    Emotion-Aware Systems: Power human-computer interaction, mental health assessment, and adaptive learning apps

    Secure & Ethical Collection

    Data Security: All data is securely processed and stored on FutureBeeAI’s proprietary platform
    Ethical Standards: Collection followed strict ethical guidelines ensuring participant privacy and informed consent
    Informed Consent: All participants were made aware of the data use and provided written consent

    Dataset Updates & Customization

    To support evolving AI development needs, this dataset is regularly updated and can be tailored to project-specific requirements. Custom options include:

  13. m

    Cry, Laugh, or Angry? A Benchmark Dataset for Computer Vision-Based Approach...

    • data.mendeley.com
    Updated Mar 10, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Md. Mehedi Hasan (2025). Cry, Laugh, or Angry? A Benchmark Dataset for Computer Vision-Based Approach to Infant Facial Emotion Recognition [Dataset]. http://doi.org/10.17632/hy969mrx9p.1
    Explore at:
    Dataset updated
    Mar 10, 2025
    Authors
    Md. Mehedi Hasan
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    This dataset is a meticulously curated dataset designed for infant facial emotion recognition, featuring four primary emotional expressions: Angry, Cry, Laugh, and Normal. The dataset aims to facilitate research in machine learning, deep learning, affective computing, and human-computer interaction by providing a large collection of labeled infant facial images.

    Primary Data (1600 Images): - Angry: 400 - Cry: 400 - Laugh: 400 - Normal: 400

    Data Augmentation & Expanded Dataset (26,143 Images): To enhance the dataset's robustness and expand the dataset, 20 augmentation techniques (including HorizontalFlip, VerticalFlip, Rotate, ShiftScaleRotate, BrightnessContrast, GaussNoise, GaussianBlur, Sharpen, HueSaturationValue, CLAHE, GridDistortion, ElasticTransform, GammaCorrection, MotionBlur, ColorJitter, Emboss, Equalize, Posterize, FogEffect, and RainEffect) were applied randomly. This resulted in a significantly larger dataset with:

    • Angry: 5,781
    • Cry: 6,930
    • Laugh: 6,870
    • Normal: 6,562

    Data Collection & Ethical Considerations: The dataset was collected under strict ethical guidelines to ensure compliance with privacy and data protection laws. Key ethical considerations include: 1. Ethical Approval: The study was reviewed and approved by the Institutional Review Board (IRB) of Daffodil International University under Reference No: REC-FSIT-2024-11-10. 2. Informed Parental Consent: Written consent was obtained from parents before capturing and utilizing infant facial images for research purposes. 3. Privacy Protection: No personally identifiable information (PII) is included in the dataset, and images are strictly used for research in AI-driven emotion recognition.

    Data Collection Locations & Geographical Diversity: To ensure diversity in infant facial expressions, data collection was conducted across multiple locations in Bangladesh, covering healthcare centers and educational institutions:

    1. 250-bed District Sadar Hospital, Sherpur (Latitude: 25.019405 & Longitude: 90.013733)
    2. Upazila Health Complex, Baraigram, Natore (Latitude: 24.3083 & Longitude: 89.1700)
    3. Char Bhabna Community Clinic, Sherpur (Latitude: 25.0188 & Longitude: 90.0175)
    4. Jamiatul Amin Mohammad Al-Islamia Cadet Madrasa, Khagan, Dhaka (Latitude: 23.872856 & Longitude: 90.318947)

    Face Detection Methodology: To extract the facial regions efficiently, RetinaNet—a deep learning-based object detection model—was employed. The use of RetinaNet ensures precise facial cropping while minimizing background noise and occlusions.

    Potential Applications: 1. Affective Computing: Understanding infant emotions for smart healthcare and early childhood development. 2. Computer Vision: Training deep learning models for automated infant facial expression recognition. 3. Pediatric & Mental Health Research: Assisting in early autism screening and emotion-aware AI for child psychology. 4. Human-Computer Interaction (HCI): Designing AI-powered assistive technologies for infants.

  14. The Japanese Female Facial Expression (JAFFE) Dataset

    • zenodo.org
    • data.niaid.nih.gov
    txt, zip
    Updated Mar 5, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Michael Lyons; Michael Lyons; Miyuki Kamachi; Miyuki Kamachi; Jiro Gyoba; Jiro Gyoba (2025). The Japanese Female Facial Expression (JAFFE) Dataset [Dataset]. http://doi.org/10.5281/zenodo.14974867
    Explore at:
    zip, txtAvailable download formats
    Dataset updated
    Mar 5, 2025
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Michael Lyons; Michael Lyons; Miyuki Kamachi; Miyuki Kamachi; Jiro Gyoba; Jiro Gyoba
    Time period covered
    1997
    Description

    The JAFFE images may be used only for non-commercial scientific research.

    The source and background of the dataset must be acknowledged by citing the following two articles. Users should read both carefully.

    Michael J. Lyons, Miyuki Kamachi, Jiro Gyoba.
    Coding Facial Expressions with Gabor Wavelets (IVC Special Issue)
    arXiv:2009.05938 (2020) https://arxiv.org/pdf/2009.05938.pdf

    Michael J. Lyons
    "Excavating AI" Re-excavated: Debunking a Fallacious Account of the JAFFE Dataset
    arXiv: 2107.13998 (2021) https://arxiv.org/abs/2107.13998

    The following is not allowed:

    • Redistribution of the JAFFE dataset (incl. via Github, Kaggle, Colaboratory, GitCafe, CSDN etc.)
    • Posting JAFFE images on the web and social media
    • Public exhibition of JAFFE images in museums/galleries etc.
    • Broadcast in the mass media (tv shows, films, etc.)

    A few sample images (not more than 10) may be displayed in scientific publications.

  15. Data from: An fMRI dataset in response to large-scale short natural dynamic...

    • openneuro.org
    Updated Oct 15, 2024
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Panpan Chen; Chi Zhang; Bao Li; Li Tong; Linyuan Wang; Shuxiao Ma; Long Cao; Ziya Yu; Bin Yan (2024). An fMRI dataset in response to large-scale short natural dynamic facial expression videos [Dataset]. http://doi.org/10.18112/openneuro.ds005047.v1.0.7
    Explore at:
    Dataset updated
    Oct 15, 2024
    Dataset provided by
    OpenNeurohttps://openneuro.org/
    Authors
    Panpan Chen; Chi Zhang; Bao Li; Li Tong; Linyuan Wang; Shuxiao Ma; Long Cao; Ziya Yu; Bin Yan
    License

    CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
    License information was derived automatically

    Description

    Summary

    Facial expression is among the most natural methods for human beings to convey their emotional information in daily life. Although the neural mechanisms of facial expression have been extensively studied employing lab-controlled images and a small number of lab-controlled video stimuli, how the human brain processes natural dynamic facial expression videos still needs to be investigated. To our knowledge, this type of data specifically on large-scale natural facial expression videos is currently missing. We describe here the natural Facial Expressions Dataset (NFED), an fMRI dataset including responses to 1,320 short (3-second) natural facial expression video clips. These video clips are annotated with three types of labels: emotion, gender, and ethnicity, along with accompanying metadata. We validate that the dataset has good quality within and across participants and, notably, can capture temporal and spatial stimuli features. NFED provides researchers with fMRI data for understanding of the visual processing of large number of natural facial expression videos.

    Data Records

    Data Records The data, which is structured following the BIDS format53, were accessible at https://openneuro.org/datasets/ds00504754. The “sub-

    Stimulus. Distinct folders store the stimuli for distinct fMRI experiments: "stimuli/face-video", "stimuli/floc", and "stimuli/prf" (Fig. 2b). The category labels and metadata corresponding to video stimuli are stored in the "videos-stimuli_category_metadata.tsv”. The “videos-stimuli_description.json” file describes category and metadata information of video stimuli(Fig. 2b).

    Raw MRI data. Each participant's folder is comprised of 11 session folders: “sub-

    Volume data from pre-processing. The pre-processed volume-based fMRI data were in the folder named “pre-processed_volume_data/sub-

    Surface data from pre-processing. The pre-processed surface-based data were stored in a file named “volumetosurface/sub-

    FreeSurfer recon-all. The results of reconstructing the cortical surface are saved as “recon-all-FreeSurfer/sub-

    Surface-based GLM analysis data. We have conducted GLMsingle on the data of the main experiment. There is a file named “sub--

    Validation. The code of technical validation was saved in the “derivatives/validation/code” folder. The results of technical validation were saved in the “derivatives/validation/results” folder (Fig. 2h). The “README.md” describes the detailed information of code and results.

  16. H

    Data from: Facial Expression Phoenix (FePh): An Annotated Sequenced Dataset...

    • dataverse.harvard.edu
    Updated Sep 11, 2020
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Marie Alaghband; Niloofar Yousefi; Ivan Garibay (2020). Facial Expression Phoenix (FePh): An Annotated Sequenced Dataset for Facial and Emotion-Specified Expressions in Sign Language [Dataset]. http://doi.org/10.7910/DVN/358QMQ
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Sep 11, 2020
    Dataset provided by
    Harvard Dataverse
    Authors
    Marie Alaghband; Niloofar Yousefi; Ivan Garibay
    License

    https://dataverse.harvard.edu/api/datasets/:persistentId/versions/2.2/customlicense?persistentId=doi:10.7910/DVN/358QMQhttps://dataverse.harvard.edu/api/datasets/:persistentId/versions/2.2/customlicense?persistentId=doi:10.7910/DVN/358QMQ

    Description

    Facial expressions are important parts of both gesture and sign language recognition systems. Despite the recent advances in both fields, annotated facial expression datasets in the context of sign language are still scarce resources. In this manuscript, we introduce an annotated sequenced facial expression dataset in the context of sign language, comprising over 3000 facial images extracted from the daily news and weather forecast of the public tv-station PHOENIX. Unlike the majority of currently existing facial expression datasets, FePh provides sequenced semi-blurry facial images with different head poses, orientations, and movements. In addition, in the majority of images, identities are mouthing the words, which makes the data more challenging. To annotate this dataset we consider primary, secondary, and tertiary dyads of seven basic emotions of "sad", "surprise", "fear", "angry", "neutral", "disgust", and "happy". We also considered the "None" class if the image's facial expression could not be described by any of the aforementioned emotions. Although we provide FePh as a facial expression dataset of signers in sign language, it has a wider application in gesture recognition and Human Computer Interaction (HCI) systems.

  17. Z

    Data from: UIBVFEDPlus-Light: Virtual facial expression dataset with...

    • data.niaid.nih.gov
    Updated Jul 8, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Mascaró Oliver, Miquel (2024). UIBVFEDPlus-Light: Virtual facial expression dataset with lighting [Dataset]. https://data.niaid.nih.gov/resources?id=zenodo_10377462
    Explore at:
    Dataset updated
    Jul 8, 2024
    Authors
    Mascaró Oliver, Miquel
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    This database, named UIBVFEDPlus-Light, is an extension of the previously published UIBVFED virtual facial expression dataset. It includes 100 characters, four lighting configurations and 13200 images. Images are in png format with a resolution of 1080x1920 RGB, without alpha channel and an average size of 2.0 MB.The images represent virtual characters reproducing FACS-based facial expressions. Expressions are classified based on the six universal emotions (Anger, Disgust, Fear, Joy, Sadness, and Surprise) labeled according to Faigin’s classification.The dataset aims to give researchers access to data they may use to support their research and generate new knowledge. In particular, to study the effect of lighting conditions in the fields of facial expression and emotion recognition.

  18. Pre Trained Model For Emotion Detection

    • kaggle.com
    Updated Jan 30, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Abhishek Singh (2024). Pre Trained Model For Emotion Detection [Dataset]. http://doi.org/10.34740/kaggle/ds/4374471
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Jan 30, 2024
    Dataset provided by
    Kaggle
    Authors
    Abhishek Singh
    License

    https://creativecommons.org/publicdomain/zero/1.0/https://creativecommons.org/publicdomain/zero/1.0/

    Description

    FER2013 (Facial Expression Recognition 2013) dataset is a widely used dataset for training and evaluating facial expression recognition models. Here are key details about the FER2013 dataset:

    Overview:

    FER2013 is a dataset designed for facial expression recognition tasks, particularly the classification of facial expressions into seven different emotion categories. The dataset was introduced for the Emotion Recognition in the Wild (EmotiW) Challenge in 2013.

    Emotion Categories:

    The dataset consists of images labeled with seven emotion categories: Angry, Disgust, Fear, Happy, Sad, Surprise, and Neutral.

    Image Size:

    Each image in the FER2013 dataset is grayscale and has a resolution of 48x48 pixels.

    Number of Images:

    The dataset contains a total of 35,887 labeled images, with approximately 5,000 images per emotion category. Partitioning:

    FER2013 is often divided into training, validation, and test sets. The original split has 28,709 images for training, 3,589 images for validation, and 3,589 images for testing.

    Usage in Research:

    FER2013 has been widely used in research for benchmarking and training facial expression recognition models, particularly deep learning models. It provides a standard dataset for evaluating the performance of models on real-world facial expressions. Challenges:

    The FER2013 dataset is known for its relatively simple and posed facial expressions. In real-world scenarios, facial expressions can be more complex and spontaneous, and there are datasets addressing these challenges.

    Challenges and Criticisms:

    Some criticisms of the dataset include its relatively small size, limited diversity in facial expressions, and the fact that some expressions (e.g., "Disgust") are challenging to recognize accurately.

    This pre trained machine model implements a Convolutional Neural Network (CNN) for emotion detection using the TensorFlow and Keras frameworks. The model architecture includes convolutional layers, batch normalization, and dropout for effective feature extraction and classification. The training process utilizes an ImageDataGenerator for data augmentation, enhancing the model's ability to generalize to various facial expressions.

    Key Steps:

    Model Training: The CNN model is trained on an emotion dataset using an ImageDataGenerator for dynamic data augmentation. Training is performed over a specified number of epochs with a reduced batch size for efficient learning.

    Model Checkpoint: ModelCheckpoint is employed to save the best-performing model during training, ensuring that the most accurate model is retained.

    Save Model and Memory Cleanup: The trained model is saved in both HDF5 and JSON formats. Memory is efficiently managed by deallocating resources, clearing the Keras session, and performing garbage collection.

  19. h

    facial-emotion-recognition-dataset

    • huggingface.co
    Updated Jul 22, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Unique Data (2023). facial-emotion-recognition-dataset [Dataset]. https://huggingface.co/datasets/UniqueData/facial-emotion-recognition-dataset
    Explore at:
    Dataset updated
    Jul 22, 2023
    Authors
    Unique Data
    License

    Attribution-NonCommercial-NoDerivs 4.0 (CC BY-NC-ND 4.0)https://creativecommons.org/licenses/by-nc-nd/4.0/
    License information was derived automatically

    Description

    The dataset consists of images capturing people displaying 7 distinct emotions (anger, contempt, disgust, fear, happiness, sadness and surprise). Each image in the dataset represents one of these specific emotions, enabling researchers and machine learning practitioners to study and develop models for emotion recognition and analysis. The images encompass a diverse range of individuals, including different genders, ethnicities, and age groups*. The dataset aims to provide a comprehensive representation of human emotions, allowing for a wide range of use cases.

  20. s

    Indoor Facial 75 Expressions Dataset

    • shaip.com
    • st.shaip.com
    • +9more
    json
    Updated Nov 26, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Shaip (2024). Indoor Facial 75 Expressions Dataset [Dataset]. https://www.shaip.com/offerings/facial-body-part-segmentation-and-recognition-datasets/
    Explore at:
    jsonAvailable download formats
    Dataset updated
    Nov 26, 2024
    Dataset authored and provided by
    Shaip
    License

    CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
    License information was derived automatically

    Description

    The Indoor Facial 75 Expressions Dataset enriches the internet, media, entertainment, and mobile sectors with an in-depth exploration of human emotions. It features 60 individuals in indoor settings, showcasing a balanced gender representation and varied postures, with 75 distinct facial expressions per person. This dataset is tagged with facial expression categories, making it an invaluable tool for emotion recognition and interactive applications.

Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
Unidata (2025). facial-expression-recognition-dataset [Dataset]. https://huggingface.co/datasets/UniDataPro/facial-expression-recognition-dataset

facial-expression-recognition-dataset

UniDataPro/facial-expression-recognition-dataset

Explore at:
388 scholarly articles cite this dataset (View in Google Scholar)
Dataset updated
Mar 31, 2025
Authors
Unidata
License

Attribution-NonCommercial-NoDerivs 4.0 (CC BY-NC-ND 4.0)https://creativecommons.org/licenses/by-nc-nd/4.0/
License information was derived automatically

Description

Emotion recognition Dataset

Dataset comprises 199,955 images featuring 28,565 individuals displaying a variety of facial expressions. It is designed for research in emotion recognition and facial expression analysis across diverse races, genders, and ages. By utilizing this dataset, researchers and developers can enhance their understanding of facial recognition technology and improve the accuracy of emotion classification systems. - Get the data

  Examples of data

This… See the full description on the dataset page: https://huggingface.co/datasets/UniDataPro/facial-expression-recognition-dataset.

Search
Clear search
Close search
Google apps
Main menu