This dataset was created by Jayesh sonawane
Attribution-NonCommercial-NoDerivs 4.0 (CC BY-NC-ND 4.0)https://creativecommons.org/licenses/by-nc-nd/4.0/
License information was derived automatically
Face Recognition, Face Detection, Male Photo Dataset šØ
If you are interested in biometric data - visit our website to learn more and buy the dataset :)
110,000+ photos of 74,000+ men from 141 countries. The dataset includes photos of people's faces. All people presented in the dataset are men. The dataset contains a variety of images capturing individuals from diverse backgrounds and age groups. Our dataset will diversify your data by adding more photos of menā¦ See the full description on the dataset page: https://huggingface.co/datasets/TrainingDataPro/male-selfie-image-dataset.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Biometric management and that to which uses face, is indeed a very challenging work and requires a dedicated dataset which imbibes in it variations in pose, emotion and even occlusions. The Current work aims at delivering a dataset for training and testing purposes.SJB Face dataset is one such Indian face image dataset, which can be used to recognize faces. SJB Face dataset contains face images which were collected from digital camera. The face dataset collected has certain conditions such as different pose, Expressions, face partially occluded and with a uniform attire. SJB Face Dataset was collected from 48 students in which each of them consisted of 13 face images. All the images have in it the students in white attire. This database shall be used for face recognition projects in academia and industry to develop attendance systems and other relevant areas, as attendance system requires to have a systematic images for training.
Flickr-Faces-HQ (FFHQ) consists of 70,000 high-quality PNG images at 1024Ć1024 resolution and contains considerable variation in terms of age, ethnicity and image background. It also has good coverage of accessories such as eyeglasses, sunglasses, hats, etc. The images were crawled from Flickr, thus inheriting all the biases of that website, and automatically aligned and cropped using dlib. Only images under permissive licenses were collected. Various automatic filters were used to prune the set, and finally Amazon Mechanical Turk was used to remove the occasional statues, paintings, or photos of photos.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
The dataset for this project is characterised by photos of individual human emotion expression and these photos are taken with the help of both digital camera and a mobile phone camera from different angles, posture, background, light exposure, and distances. This task might look and sound very easy but there were some challenges encountered along the process which are reviewed below: 1) People constraint One of the major challenges faced during this project is getting people to participate in the image capturing process as school was on vacation, and other individuals gotten around the environment were not willing to let their images be captured for personal and security reasons even after explaining the notion behind the project which is mainly for academic research purposes. Due to this challenge, we resorted to capturing the images of the researcher and just a few other willing individuals. 2) Time constraint As with all deep learning projects, the more data available the more accuracy and less error the result will produce. At the initial stage of the project, it was agreed to have 10 emotional expression photos each of at least 50 persons and we can increase the number of photos for more accurate results but due to the constraint in time of this project an agreement was later made to just capture the researcher and a few other people that are willing and available. These photos were taken for just two types of human emotion expression that is, āhappyā and āsadā faces due to time constraint too. To expand our work further on this project (as future works and recommendations), photos of other facial expression such as anger, contempt, disgust, fright, and surprise can be included if time permits. 3) The approved facial emotions capture. It was agreed to capture as many angles and posture of just two facial emotions for this project with at least 10 images emotional expression per individual, but due to time and people constraints few persons were captured with as many postures as possible for this project which is stated below: Ć Happy faces: 65 images Ć Sad faces: 62 images There are many other types of facial emotions and again to expand our project in the future, we can include all the other types of the facial emotions if time permits, and people are readily available. 4) Expand Further. This project can be improved furthermore with so many abilities, again due to the limitation of time given to this project, these improvements can be implemented later as future works. In simple words, this project is to detect/predict real-time human emotion which involves creating a model that can detect the percentage confidence of any happy or sad facial image. The higher the percentage confidence the more accurate the facial fed into the model. 5) Other Questions Can the model be reproducible? the supposed response to this question should be YES. If and only if the model will be fed with the proper data (images) such as images of other types of emotional expression.
CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
License information was derived automatically
The Tufts Face Database is a collection of images of human faces, often used in computer vision and facial recognition research. Tufts University's Face Database is known for its high-quality images and has been widely utilized in academic and industrial research projects..
CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
License information was derived automatically
The Animal Faces dataset is a collection of images featuring the faces of various animal species.
https://www.futurebeeai.com/data-license-agreementhttps://www.futurebeeai.com/data-license-agreement
Welcome to the Middle Eastern Human Face with Occlusion Dataset, meticulously curated to enhance face recognition models and support the development of advanced occlusion detection systems, biometric identification systems, KYC models, and other facial recognition technologies.
This dataset comprises over 3,000 human facial images, divided into participant-wise sets with each set including:
The dataset includes contributions from a diverse network of individuals across Middle Eastern countries:
To ensure high utility and robustness, all images are captured under varying conditions:
Each facial image set is accompanied by detailed metadata for each participant, including:
This metadata is essential for training models that can accurately recognize and identify human faces with occlusions across different demographics and conditions.
This facial image dataset is ideal for various applications in the field of computer vision, including but not limited to:
We understand the evolving nature of AI and machine
Flickr-Faces-HQ Dataset (FFHQ)
Flickr-Faces-HQ (FFHQ) is a high-quality image dataset of human faces, originally created as a benchmark for generative adversarial networks (GAN):
A Style-Based Generator Architecture for Generative Adversarial Networks Tero Karras (NVIDIA), Samuli Laine (NVIDIA), Timo Aila (NVIDIA) https://arxiv.org/abs/1812.04948
The dataset consists of 70,000 high-quality PNG images at 1024Ć1024 resolution and contains considerable variation in terms ofā¦ See the full description on the dataset page: https://huggingface.co/datasets/nuwandaa/ffhq128.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Similar face recognition has always been one of the most challenging research directions in face recognition.This project shared similar face images (SFD.zip) that we have collected so far. All images are labeld and collected from publicly available datasets such as LFW, CASIA-WebFace.We will continue to collect larger-scale data and continue to update this project.Because the data set is too large, we uploaded a compressed zip file (SFD.zip). Meanwhile here we upload a few examples for everyone to view.email: ileven@shu.edu.cn
Attribution-NonCommercial 3.0 (CC BY-NC 3.0)https://creativecommons.org/licenses/by-nc/3.0/
License information was derived automatically
Indonesian Muslim Student Face Dataset is a face image dataset designed to facilitate facial analysis or facial recognition research. The dataset encompasses a wide range of demographic attributes including diverse genders, ethnicities, and poses. Images were collected using various capture devices such as CCTV, Webcam, Front and Back Smartphone cameras, and DSLR camera, ensuring a diverse representation of facial characteristics and imaging conditions. The dataset underwent rigorous pre-processing, including frame extraction from video, accurate face detection, and precise cropping, resulting in standardized face images that effectively isolate and emphasize facial features.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
An index of face databases, their features, and how to access them has been unavailable. The āFace Image Meta-Databaseā (fIMDb) provides researchers with the tools to find the face images best suited to their research. The fIMDb is available from: https://cliffordworkman.com/resources/
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
This data is used in the second experimental evaluation of face smile detection in the paper titled "Smile detection using Hybrid Face Representaion" - O.A.Arigbabu et al. 2015.
Download the main images from LFWcrop website: http://conradsanderson.id.au/lfwcrop/ to select the samples we used for smile and non-smile, as in the list.
Kindly cite:
Arigbabu, Olasimbo Ayodeji, et al. "Smile detection using hybrid face representation." Journal of Ambient Intelligence and Humanized Computing (2016): 1-12.
C. Sanderson, B.C. Lovell. Multi-Region Probabilistic Histograms for Robust and Scalable Identity Inference. ICB 2009, LNCS 5558, pp. 199-208, 2009
Huang GB, Mattar M, Berg T, Learned-Miller E (2007) Labeled faces in the wild: a database for studying face recognition in unconstrained environments. University of Massachusetts, Amherst, Technical Report
FFHQ-Text is a small-scale face image dataset with large-scale facial attributes, designed for text-to-face generation & manipulation, text-guided facial image manipulation, and other vision-related tasks. This dataset is an extension of the NVIDIA Flickr-Faces-HQ Dataset (FFHQ), which is the selected top 760 female FFHQ images that only contain one complete human face.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
*** The paper published on Multimedia Tools and Applications (Springer) - 2024 ****** Title: "SPRITZ-PS: Vlaidation of Synthetic Face Images Using A Large Dataset of Printed Docuemnts"****** Authors: Ehsan Nowroozi, Yoosef Habibi, and Mauro Conti ***----------------------------------------------------------------------------------------------------------GAN-generated faces look challenging to distinguish from genuine human faces. As a result, because synthetic images are presently being used as profile photos for fake identities on social media, they may have serious social consequences. Iris pattern anomalies might expose GAN-generated facial photos. When photographs are printed and scanned, it becomes more difficult to distinguish between genuine and counterfeit since fraudulent images lose some of their qualities. We created a new collection of iris images from printed and scanned documents by segmenting pupils from face images to address these concerns. We employed Dlib, which provides 68 facial landmarks, and EyeCool to extract both left and right irises from a full face image to segment the iris portion. Nevertheless, due to eyelid occlusion, the extracted iris images are not entirely shaped. We have to fill missing pixels of extracted iris since there are no incomplete iris in the actual world and the incomplete image is not an adequate input to train deep neural networks. We employ the Hypergraph convolution-based image inpainting approach to do this. The detailed relationship between the iris images was determined using hypergraph convolution.
https://creativecommons.org/publicdomain/zero/1.0/https://creativecommons.org/publicdomain/zero/1.0/
All the images of faces here are generated using https://thispersondoesnotexist.com/
https://www.googleapis.com/download/storage/v1/b/kaggle-user-content/o/inbox%2F1842206%2F4c3d3569f4f9c12fc898d76390f68dab%2FBeFunky-collage.jpg?generation=1662079836729388&alt=media" alt="">
Under US copyright law, these images are technically not subject to copyright protection. Only "original works of authorship" are considered. "To qualify as a work of 'authorship' a work must be created by a human being," according to a US Copyright Office's report [PDF].
https://www.theregister.com/2022/08/14/ai_digital_artwork_copyright/
I manually tagged all images as best as I could and separated them between the two classes below
Some may pass either female or male, but I will leave it to you to do the reviewing. I included toddlers and babies under Male/ Female
Each of the faces are totally fake, created using an algorithm called Generative Adversarial Networks (GANs).
A generative adversarial network (GAN) is a class of machine learning frameworks designed by Ian Goodfellow and his colleagues in June 2014. Two neural networks contest with each other in a game (in the form of a zero-sum game, where one agent's gain is another agent's loss).
Given a training set, this technique learns to generate new data with the same statistics as the training set. For example, a GAN trained on photographs can generate new photographs that look at least superficially authentic to human observers, having many realistic characteristics. Though originally proposed as a form of generative model for unsupervised learning, GANs have also proved useful for semi-supervised learning, fully supervised learning,and reinforcement learning.
Just a simple Jupyter notebook that looped and invoked the website https://thispersondoesnotexist.com/ , saving all images locally
Data Collection - TagX can provides the dataset based on following scenarios to train a biasfree face analysis and detection dataset- Single and multiple faces images Monk skin-tones covered All Facial angles included
Metadata for Face Images- We can provide following metadata along with the collected images Age Range Distance from camera Emotion State Accessories present(Eyeglasses, hat etc.) pose with the values of pitch, roll, and yaw.
Annotation of Face Images- We can provides annotation for face detection applications like Bounding box annotation, Landmark annotation or polygon annotation. We have a dataset prepared with bounding box annotation around faces for 30000 images.
This dataset was created by Christos Antoniou
https://www.futurebeeai.com/data-license-agreementhttps://www.futurebeeai.com/data-license-agreement
Welcome to the Caucasian Human Facial Images Dataset, meticulously curated to enhance face recognition models and support the development of advanced biometric identification systems, KYC models, and other facial recognition technologies.
This dataset comprises over 1,000 Caucasian individual facial image sets, with each set including:
The dataset includes contributions from a diverse network of individuals across Caucasian countries.
To ensure high utility and robustness, all images are captured under varying conditions:
Each facial image set is accompanied by detailed metadata for each participant, including:
This metadata is essential for training models that can accurately recognize and identify faces across different demographics and conditions.
This facial image dataset is ideal for various applications in the field of computer vision, including but not limited to:
We understand the evolving nature of AI and machine learning requirements. Therefore, we continuously add more assets with diverse conditions to this off-the-shelf facial image dataset.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
This project investigates bias in automatic facial recognition (FR). Specifically, subjects are grouped into predefined subgroups based on gender, ethnicity, and age. We propose a novel image collection called Balanced Faces in the Wild (BFW), which is balanced across eight subgroups (i.e., 800 face images of 100 subjects, each with 25 face samples). Thus, along with the name (i.e., identification) labels and task protocols (e.g., list of pairs for face verification, pre-packaged data-table with additional metadata and labels, etc.), BFW groups into ethnicities (i.e., Asian (A), Black (B), Indian (I), and White (W)) and genders (i.e., Females (F) and Males (M)). Thus, the motivation and intent are that BFW will provide a proxy to characterize FR systems with demographic-specific analysis now possible. For instance, various confusion metrics and the predefined criteria (i.e., score threshold) are fundamental when characterizing performance ratings of FR systems. The following visualization summarizes the confusion metrics in a way that relates to the different measurements.
This dataset was created by Jayesh sonawane