Apache License, v2.0https://www.apache.org/licenses/LICENSE-2.0
License information was derived automatically
Leonardo6/memotion dataset hosted on Hugging Face and contributed by the HF Datasets community
MIT Licensehttps://opensource.org/licenses/MIT
License information was derived automatically
Dataset Title Memotion-Expanded: Multimodal Sarcasm Detection with AI-Human Annotation Comparisons
Author: Sirojiddin Bobokulov, University of Bremen License: CC-BY 4.0 (inherits original Memotion 7k permissions)
Dataset Description This dataset extends the Memotion 7k benchmark with AI-generated annotations and explanations for sarcasm detection. It contains 6,905 entries comparing:
Human annotations (original multimodal labels from Memotion 7k)
AI annotations for both multimodal (text + image) and unimodal (text-only) conditions
Model explanations (≤20 words) justifying AI predictions
Key Attributes Column Description Example image_name Filename of meme image image_1.jpg text_corrected Transcribed/cleaned meme text "LOOK THERE MY FRIEND LIGHTYEAR..." multimodal_annotation_ai AI sarcasm label (multimodal) general multimodal_explanation_ai AI rationale (text + image) "Making fun of Facebook trends..." unimodal_annotation_ai AI sarcasm label (text-only) twisted_meaning unimodal_explanation_ai AI rationale (text-only) "Uses exaggeration to mock..." multimodal_annotation_humans Original human label general Sarcasm Categories:
general, twisted_meaning, very_twisted, non-sarcastic
Key Contributions AI-Human Comparison: Directly compare multimodal AI vs. unimodal AI vs. human sarcasm judgments.
Explanation Alignment: Study how AI rationales (e.g., "exaggeration to mock") align with human annotations.
Modality Impact: Analyze performance differences between text-only vs. text+image conditions.
Use Cases Sarcasm Detection: Train models using human/AI annotations as silver labels.
Explainability Research: Evaluate if AI explanations match human sarcasm perception.
Modality Studies: Quantify how visual context affects sarcasm detection accuracy.
Dataset Structure python Copy Sample Entry: { "image_name": "image_1.jpg", "text_corrected": "LOOK THERE MY FRIEND...FACEBOOK imgflip.com", "multimodal_annotation_ai": "general", "multimodal_explanation_ai": "Making fun of Facebook trends and followers", "unimodal_annotation_ai": "twisted_meaning", "unimodal_explanation_ai": "The text uses exaggeration to sarcastically mock...", "multimodal_annotation_humans": "general" } Ethical Considerations Image Licensing: Host only anonymized image URLs (no direct redistribution).
Bias Mitigation: Annotate model confidence scores for low-frequency categories.
Citations Original Memotion 7k:
bibtex Copy @inproceedings{chhavi2020memotion, title={Memotion Analysis 1.0 @SemEval 2020}, author={Sharma, Chhavi et al.}, booktitle={Proceedings of SemEval-2020}, year={2020} } This Extension:
bibtex Copy @dataset{bobokulov2024memotion_ai, author = {Sirojiddin Bobokulov}, title = {Memotion-Expanded: AI-Human Sarcasm Annotation Comparisons}, year = {2024}, note = {Extended from Memotion 7k (https://www.kaggle.com/datasets/williamscott701/memotion-dataset-7k)} }
A multimodal dataset for sentiment analysis on internet memes.
MIT Licensehttps://opensource.org/licenses/MIT
License information was derived automatically
Dataset Card for "emotion"
Dataset Summary
Emotion is a dataset of English Twitter messages with six basic emotions: anger, fear, joy, love, sadness, and surprise. For more detailed information please refer to the paper.
Dataset Structure
Data Instances
An example looks as follows. { "text": "im feeling quite sad and sorry for myself but ill snap out of it soon", "label": 0 }
Data Fields
The data fields are:
text: a string feature.… See the full description on the dataset page: https://huggingface.co/datasets/XuehangCang/emotion-417k.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Emotion Detection Model for Facial Expressions
Project Description:
In this project, we developed an Emotion Detection Model using a curated dataset of 715 facial images, aiming to accurately recognize and categorize expressions into five distinct emotion classes. The emotion classes include Happy, Sad, Fearful, Angry, and Neutral.
Objectives: - Train a robust machine learning model capable of accurately detecting and classifying facial expressions in real-time. - Implement emotion detection to enhance user experience in applications such as human-computer interaction, virtual assistants, and emotion-aware systems.
Methodology: 1. Data Collection and Preprocessing: - Assembled a diverse dataset of 715 images featuring individuals expressing different emotions. - Employed Roboflow for efficient data preprocessing, handling image augmentation and normalization.
Model Architecture:
Training and Validation:
Model Evaluation:
Deployment and Integration:
Results: The developed Emotion Detection Model demonstrates high accuracy in recognizing and classifying facial expressions across the defined emotion classes. This project lays the foundation for integrating emotion-aware systems into various applications, fostering more intuitive and responsive interactions.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
This dataset is collected and annotated for the SMILE project http://www.culturesmile.org. This collection of tweets mentioning 13 Twitter handles associated with British museums was gathered between May 2013 and June 2015. It was created for the purpose of classifying emotions, expressed on Twitter towards arts and cultural experiences in museums. It contains 3,085 tweets, with 5 emotions namely anger, disgust, happiness, surprise and sadness. Please see our paper "SMILE: Twitter Emotion Classification using Domain Adaptation" for more details of the dataset.License: The annotations are provided under a CC-BY license, while Twitter retains the ownership and rights of the content of the tweets.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
## Overview
Facial Emotion Recognition is a dataset for object detection tasks - it contains Emotions annotations for 4,540 images.
## Getting Started
You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
## License
This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Gilman-Adhered FilmClip Emotion Dataset (GAFED): Tailored Clips for Emotional Elicitation
Description:
Introducing the Gilman-Adhered FilmClip Emotion Dataset (GAFED) - a cutting-edge compilation of video clips curated explicitly based on the guidelines set by Gilman et al. (2017). This dataset is meticulously structured, leveraging both the realms of film and psychological research. The objective is clear: to induce specific emotional responses with utmost precision and reproducibility. Perfectly tuned for researchers, therapists, and educators, GAFED facilitates an in-depth exploration into the human emotional spectrum using the medium of film.
Dataset Highlights:
Gilman's Guidelines: GAFED's foundation is built upon the rigorous criteria and insights provided by Gilman et al., ensuring methodological accuracy and relevance in emotional elicitation.
Film Titles: Each selected film's title provides an immersive backdrop to the emotions sought to be evoked.
Emotion Label: A focused emotional response is designated for each clip, reinforcing the consistency in elicitation.
Clip Duration: Standardized duration of every clip ensures a uniform exposure, leading to consistent response measurements.
Curated with Precision: Every film clip in GAFED has been reviewed and handpicked, echoing Gilman et al.'s principles, thereby cementing their efficacy in triggering the intended emotion.
Emotion-Eliciting Video Clips within Dataset:
Film
Targeted Emotion
Duration (seconds)
The Lover
Baseline
43
American History X
Anger
106
Cry Freedom
Sadness
166
Alive
Happiness
310
Scream
Fear
395
The crowning feature of GAFED is its identification of "key moments". These crucial timestamps serve as a bridge between cinema and emotion, guiding researchers to intervals teeming with emotional potency.
Key Emotional Moments within Dataset:
Film
Targeted Emotion
Key moment timestamps (seconds)
American History X
Anger
36, 57, 68
Cry Freedom
Sadness
112, 132, 154
Alive
Happiness
227, 270, 289
Scream
Fear
23, 42, 79, 226, 279, 299, 334
Based on: Gilman, T. L., et al. (2017). A film set for the elicitation of emotion in research. Behavior Research Methods, 49(6).
GAFED isn't merely a dataset; it's an amalgamation of cinema and psychology, encapsulating the vastness of human emotion. Tailored to perfection and adhering to Gilman et al.'s insights, it stands as a beacon for researchers exploring the depths of human emotion through film.
BanglaEmotion is a manually annotated Bangla Emotion corpus, which incorporates the diversity of fine-grained emotion expressions in social-media text. More fine-grained emotion labels are considered such as Sadness, Happiness, Disgust, Surprise, Fear and Anger - which are, according to Paul Ekman (1999), the six basic emotion categories. For this task, a large amount of raw text data are collected from the user’s comments on two different Facebook groups (Ekattor TV and Airport Magistrates) and from the public post of a popular blogger and activist Dr. Imran H Sarker. These comments are mostly reactions to ongoing socio-political issues and towards the economic success and failure of Bangladesh. A total of 32923 comments are scraped from the three sources aforementioned above. Out of these, a total of 6314 comments were annotated into the six categories. The distribution of the annotated corpus is as follows:
sad = 1341 happy = 1908 disgust = 703 surprise = 562 fear = 384 angry = 1416
A balanced set is also provided from the above data and split the dataset into training and test set of equal ratio. A proportion of 5:1 is used for training and evaluation purposes. More information on the dataset and the experiments on it could be found in our paper (related links below).
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Here are a few use cases for this project:
Mental Health Monitoring: The emotion recognition model could be used in a mental health tracking app to analyze users' facial expressions during video diaries or calls, providing insights into their emotional state over time.
Customer Service Improvement: Businesses could use this model to monitor customer interactions in stores, analysing the facial expressions of customers to gauge their satisfaction level or immediate reaction to products or services.
Educational and Learning Enhancement: This model could be used in an interactive learning platform to observe students' emotional responses to different learning materials, enabling tailored educational experiences.
Online Content Testing: Marketing or content creation teams could utilize this model to test viewers' emotional reactions to different advertisements or content pieces, improving the impact of their messaging.
Social Robotics: The emotion recognition model could be incorporated in social robots or AI assistants to identify human emotions and respond accordingly, improving their overall user interaction and experience.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
## Overview
Facial Emotion is a dataset for object detection tasks - it contains Emotions annotations for 642 images.
## Getting Started
You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
## License
This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).
ABSTRACT: Mixed emotions have attracted increasing interest recently, but existing datasets rarely focus on mixed emotion recognition from multimodal signals, hindering the affective computing of mixed emotions. On this basis, we present a multimodal dataset with four kinds of signals recorded while watching mixed and non-mixed emotion videos. To ensure effective emotion induction, we first implemented a rule-based video filtering step to select the videos that could elicit stronger positive, negative, and mixed emotions. Then, an experiment with 80 participants was conducted, in which the data of EEG, GSR, PPG, and frontal face videos were recorded while they watched the selected video clips. We also recorded the subjective emotional rating on PANAS, VAD, and amusement-disgust dimensions. In total, the dataset consists of multimodal signal data and self-assessment data from 73 participants. We also present technical validations for emotion induction and mixed emotion classification from physiological signals and face videos. The average accuracy of the 3-class classification (i.e., positive, negative, and mixed) can reach 80.96\% when using SVM and features from all modalities, which indicates the possibility of identifying mixed emotional states.
https://cubig.ai/store/terms-of-servicehttps://cubig.ai/store/terms-of-service
1) Data introduction • Emotion-analysis dataset is data for analyzing the emotions of text.
2) Data utilization (1) Emotion-analysis data has characteristics that: • Contains a variety of texts that convey emotions ranging from happiness to anger to sadness. The goal is to build an efficient model for detecting emotions in text. (2) Emotion-analysis data can be used to: • Sentiment classification models: This dataset can be used to train machine learning models that classify text based on sentiment, which helps companies and researchers understand public opinion and sentiment trends. • Market research: Researchers can analyze sentiment data to understand consumer preferences and market trends and support data-driven decision making.
Video dataset capturing diverse facial expressions and emotions from 1000+ people, suitable for emotion recognition AI training
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Hi,KIA dataset is a shared short Wakeup Word database focusing on perceived emotion in speech The dataset contains 488 Wakeup Word speech.
For more detailed information about the dataset, please refer to our paper: Hi, KIA: A Speech Emotion Recognition Dataset for Wake-Up Words
File Description
wav/: wav files.
Filename f{gender}_{pid}_{scene}_{trial}_{emotion}.wav
The first letter was used to express emotion.
annotation/: Information related to annotation and human validation of the entire speech
split: 8fold data split with {train, valid, test}.csv
handcraft: Features used for data EDA and baseline performance
best_weights: wav2vec2.0 context network finetuning weights for re-implementation. Due to file size, we attach only fold M1, F5
Reference
Hi, KIA: A Speech Emotion Recognition Dataset for Wake-Up Words [ArXiv]
@inproceedings{kim2022hi,
title={Hi, KIA: A Speech Emotion Recognition Dataset for Wake-Up Words},
author={Taesu Kim, SeungHeon Doh, Gyunpyo Lee, Hyung seok Jun, Juhan Nam, Hyeon-Jeong Suk},
booktitle={Proceedings of the 14th Asia Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA)},
year={2022}
}
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
## Overview
Emotion Baby is a dataset for object detection tasks - it contains Emotion annotations for 451 images.
## Getting Started
You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
## License
This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
## Overview
Emotion Classification YOLO is a dataset for classification tasks - it contains Emotions annotations for 6,880 images.
## Getting Started
You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
## License
This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).
Masked Emotion FilmClip Dataset (MEFD): Emotion Elicitation with Facial Coverings
The Masked Emotion FilmClip Dataset (MEFD) stands as an avant-garde assembly of emotion-inducing video clips tailored for a unique niche - the elicitation of emotions in individuals wearing facial masks. This dataset emerges in response to the global need to understand emotional cues and expressions in the backdrop of widespread facial mask usage. Assembled by leveraging the synergies between cinematography and psychological research, MEFD serves as an invaluable trove for researchers, especially those in AI, seeking to decode emotions even when a significant portion of the face is concealed. Dataset Highlights:
Facial Masks: All subjects in the video clips are seen wearing facial masks, replicating real-world scenarios and augmenting the dataset's relevance. Film Titles: The title of each selected film enriching the context of the emotional narrative. Emotion Label: Clear emotion classification associated with every clip, ensuring replicability in emotional elicitation. Clip Duration: Precise duration details ensuring standardized exposure and consistent emotion elicitation. Curated with Expertise: Clips have undergone rigorous evaluation by seasoned psychologists and film critics, affirming their effectiveness in eliciting the designated emotion. Consent and Ethics: The dataset respects and upholds privacy and ethical standards. Every participant provided informed consent. This endeavor has received the green light from the Ethics Committee at the University of Granada, documented under the reference: 2100/CEIH/2021. Emotion-Eliciting Video Clips within Dataset: Film Targeted Emotion Duration (seconds) The Lover Baseline 43 American History X Anger 106 Cry Freedom Sadness 166 Alive Happiness 310 Scream Fear 395 A paramount feature of MEFD is its emphasis on "key moments". These timestamps, a product of collective expertise from psychologists and film critics, guide the researcher to intervals of heightened emotional resonance within the clips, especially challenging to discern with masked faces. Key Emotional Moments within Dataset: Film Targeted Emotion Key moment timestamps (seconds) American History X Anger 36, 57, 68 Cry Freedom Sadness 112, 132, 154 Alive Happiness 227, 270, 289 Scream Fear 23, 42, 79, 226, 279, 299, 334
DATA STRUCTURE----------------- SADNESS_XXX.CSVtimestamp emotion1625062890.938222 NEUTRAL --> Initial time start for the neutral video1625062932.567609 SADNESS --> Initial time start for the EMOTION video Notes:** Subject id 15: FEAR label started to fast; Neutral data very few-----------------
The ethical consent for this dataset was provided by La Comisión de Ética en Investigación de la Universidad de Granada, as documented in the approval titled: 'DETECCIÓN AUTOMÁTICA DE LAS EMOCIONES BÁSICAS Y SU INFLUENCIA EN LA TOMA DE DECISIONES MEDIANTE WEARABLES Y MACHINE LEARNING' registered under 2100/CEIH/2021. MEFD is more than just a dataset; it is a testament to human resilience and adaptability. As facial masks become ubiquitous, understanding the nuances of masked emotional expressions becomes imperative. MEFD rises to this challenge, bridging gaps and pioneering a new frontier in emotion research.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
The AFFEC (Advancing Face-to-Face Emotion Communication) dataset is a multimodal dataset designed for emotion recognition research. It captures dynamic human interactions through electroencephalography (EEG), eye-tracking, galvanic skin response (GSR), facial movements, and self-annotations, enabling the study of felt and perceived emotions in real-world face-to-face interactions. The dataset comprises 84 simulated emotional dialogues, 72 participants, and over 5,000 trials, annotated with more than 20,000 emotion labels.
The dataset follows the Brain Imaging Data Structure (BIDS) format and consists of the following components:
sub-*
: Individual subject folders (e.g., sub-aerj
, sub-mdl
, sub-xx2
)dataset_description.json
: General dataset metadataparticipants.json
and participants.tsv
: Participant demographics and attributestask-fer_events.json
: Event annotations for the FER taskREADME.md
: This documentation filesub-
):Each subject folder contains:
beh/
): Physiological recordings (eye tracking, GSR, facial analysis, cursor tracking) in JSON and TSV formats.eeg/
): EEG recordings in .edf
and corresponding metadata in .json
.*.tsv
): Trial event data for the emotion recognition task.*_channels.tsv
): EEG channel information.sub-
sub-
sub-
sub-
sub-
sub-
task-fer_events.json
Participants engaged in a Facial Emotion Recognition (FER) task, where they watched emotionally expressive video stimuli while their physiological and behavioral responses were recorded. Participants provided self-reported ratings for both perceived and felt emotions, differentiating between the emotions they believed the video conveyed and their internal affective experience.
The dataset enables the study of individual differences in emotional perception and expression by incorporating Big Five personality trait assessments and demographic variables.
AFFEC is well-suited for research in:
This dataset was collected with the support of brAIn lab, IT University of Copenhagen.
Special thanks to all participants and research staff involved in data collection.
This dataset is shared under the Creative Commons CC0 License.
For questions or collaboration inquiries, please contact [brainlab-staff@o365team.itu.dk].
Voice dataset featuring identical English texts spoken in four emotional tones for speech emotion recognition and AI training
Apache License, v2.0https://www.apache.org/licenses/LICENSE-2.0
License information was derived automatically
Leonardo6/memotion dataset hosted on Hugging Face and contributed by the HF Datasets community