CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
License information was derived automatically
Explore the Sokoto Coventry Fingerprint Dataset (SOCOFing), a robust biometric database featuring 6,000 fingerprint images from 600 African subjects. Ideal for academic research, SOCOFing includes detailed labels and synthetic alterations to enhance fingerprint recognition and security studies.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
A DATASET CONTAINING FINGERPRINT OF 100 FAMILY OF WEST BENGAL INDIA. THE DATASET CONTAINS FINGERPRINT IMAGE OF FIVE FINGER OF RIGHT HAND OF FATHER, MOTHER AND CHILD. THE DATASET CONTAINS TOTAL 1500 FINGERPRINT IMAGE IN PNG FORMAT. THE SIZE OF THE FINGERPRINT IMAGES ARE 512 x 512 PIXELS.
2D and 3D Fingerrpint dataset. It was created by former Ph.D. stduent Ms. Wei Zhou under the supervision of Prof. J. Hu who acts as the communication contact.
NIST, working with the FBI, has digitized 888 inked fingerprint arrest cards that were in various physical conditions, from pristine to badly damaged and faded, and were collected during law enforcement professionals' duties. This database contains images of the 10 rolled fingerprint impressions, the two four-finger slap impressions (finger positions 13 and 14), the two thumb slap impressions (finger positions 11 and 12) and the segmented impressions from the slap images (13,14). The database also includes the coordinates that were used to segment the impressions from the slap fingerprint images.The cards were scanned at three different resolutions: 500, 1,000, and 2,000 pixels per inch (PPI). All three resolutions were scanned in grayscale at a depth of 8 bits-per pixel.Data available as of July 2018 is Special Database 300a, in 500 ppi with PNG formatted impressions. Data at other resolutions, in other image formats, and in other record types may be forthcoming.
In April 2017, the Intelligence Advanced Research Projects Activity (IARPA) held a dry run for the data collection portion of its Nail to Nail (N2N) Fingerprint Challenge. This data collection event was designed to ensure that the real data collection event held in September 2017 would be successful. To this end, biometric data from unhabituated individuals needed to be collected. That data is now released by NIST as Special Database 301.In total, 14 fingerprint sensors were deployed during the data collection, amassing a series of rolled and plain images. The devices include rolled fingerprints captured by skilled experts from the Federal Bureau of Investigation (FBI) Biometric Training Team. Captures of slaps, palms, and other plain impression fingerprint impressions were additionally recorded. NIST also partnered with the FBI and Schwarz Forensic Enterprises to design activity scenarios in which subjects would likely leave fingerprints on different objects. The activities and associated objects were chosen in order to use a number of latent print development techniques and simulate the types of objects often found in real law enforcement case work. NIST also collected some mugshot-style face and iris images of the subjects who participated in the dry run. These data are also available for download.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
## Overview
Fingerprint Pattern Detection is a dataset for object detection tasks - it contains Fingerprints Fingerprints 4pZv annotations for 1,730 images.
## Getting Started
You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
## License
This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
and their augmented replicas.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
## Overview
Fingerprint is a dataset for object detection tasks - it contains Index Little Middle Ring annotations for 538 images.
## Getting Started
You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
## License
This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
The LUTBIO database provides a comprehensive resource for research in multimodal biometric authentication, featuring the following key aspects:
Extensive Biometric Modalities: The database contains data from nine biometric modalities: voice, face, fingerprint, contact-based palmprint, electrocardiogram (ECG), opisthenar (back of hand), ear, contactless palmprint, and periocular region.
Diverse Demographics: Data were collected from 306 individuals, with a balanced gender distribution of 164 males and 142 females, spanning an age range of 8 to 90 years. This diverse age representation enables analyses across a wide demographic spectrum.
Representative Population Sampling: Volunteers were recruited from naturally occurring communities, ensuring a large-scale, statistically representative population. The collected data encompass variations observed in real-world environments.
Support for Multimodal and Cross-Modality Research: LUTBIO provides both contact-based and contactless palmprint data, as well as fingerprint data (from optical images and scans), promoting advancements in multimodal biometric authentication. This resource is designed to guide the development of future multimodal databases.
Flexible, Decouplable Data: The biometric data in the LUTBIO database are designed to be highly decouplable, enabling independent processing of each modality without loss of information. This flexibility supports both single-modality and multimodal analysis, empowering researchers to optimize, combine, and customize biometric features for specific applications.
✅ Data Availability: If you wish to use the LUTBIO dataset, please download the attached Word document, fill in the information, and send it as an attachment to rykeryang AT 163.com. We will process your request as soon as possible!
🥸 Important Notice: Please read the data collection protocol of the LUTBIO dataset carefully before use, as it is essential for understanding and correctly interpreting the dataset. Thank you.
😎 Good news! Our paper has been accepted by Information Fusion, and the DOI is https://doi.org/10.1016/j.inffus.2025.102945. We appreciate the reviewers and the editor for their efforts.🥰🥰🥰
In September 2017, the Intelligence Advanced Research Projects Activity (IARPA) held a data collection as part of its Nail to Nail (N2N) Fingerprint Challenge. Participating Challengers deployed devices designed to collect an image of the full nail to nail surface area of a fingerprint equivalent to a rolled fingerprint from an unacclimated user, without assistance from a trained operator. Traditional operator-assisted live-scan rolled fingerprints were also captured, along with assorted other friction ridge live-scan and latent captures.In this data collection, study participants needed to have their fingerprints captured using traditional operator-assisted techniques in order to quantify the performance of the Challenger devices. IARPA invited members of the Federal Bureau of Investigation (FBI) Biometric Training Team to the data collection to perform this task. Each study participant had N2N fingerprint images captured twice, each by a different FBI expert, resulting in two N2N baseline datasets.To ensure the veracity of recorded N2N finger positions in the baseline datasets, Challenge test staff also captured plain fingerprint impressions in a 4-4-2 slap configuration. This capture method refers to simultaneously imaging the index, middle, ring, and little fingers on the right hand, then repeating the process on the left hand, and finishing with the simultaneous capture of the left and right thumbs. This technique is a best practice to ensure finger sequence order, since it is physically challenging for a study participant to change the ordering of fingers when imaging them simultaneously. There were four baseline (two rolled and two slap), eight challenger and ten auxiliary fingerprint sensors deployed during the data collection, amassing a series of rolled and plain images. It was required that the baseline devices achieve 100% acquisition rate, in order to verify the recorded friction ridge generalized positions (FRGPs) and study participant identifiers for other devices. There were no such requirements for Challenger devices. Not all devices were able to achieve 100% acquisition rate.Plain, rolled, and touch-free impression fingerprints were captured from a multitude of devices, as well as sets of plain palm impressions. NIST also partnered with the FBI and Schwarz Forensic Enterprises (SFE) to design activity scenarios in which subjects would likely leave fingerprints on different objects. The activities and associated objects were chosen in order to use a number of latent print development techniques and simulate the types of objects often found in real law enforcement case work.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
## Overview
Fingerprint Minutiae Detection is a dataset for object detection tasks - it contains BIfurcation annotations for 300 images.
## Getting Started
You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
## License
This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).
https://academictorrents.com/nolicensespecifiedhttps://academictorrents.com/nolicensespecified
The NIST database of fingerprint images contains 2000 8-bit gray scale fingerprint image pairs. Each image is 512-by-512 pixels with 32 rows of white space at the bottom and classified using one of the five following classes:
Attribution-NonCommercial-NoDerivs 4.0 (CC BY-NC-ND 4.0)https://creativecommons.org/licenses/by-nc-nd/4.0/
License information was derived automatically
Dataset Description:
The dataset comprises a collection of photos of people, organized into folders labeled "women" and "men." Each folder contains a significant number of images to facilitate training and testing of gender detection algorithms or models.
The dataset contains a variety of images capturing female and male individuals from diverse backgrounds, age groups, and ethnicities.
https://www.googleapis.com/download/storage/v1/b/kaggle-user-content/o/inbox%2F12421376%2F1c4708f0b856f7889e3c0eea434fe8e2%2FFrame%2045%20(1).png?generation=1698764294000412&alt=media" alt="">
This labeled dataset can be utilized as training data for machine learning models, computer vision applications, and gender detection algorithms.
The dataset is split into train and test folders, each folder includes: - folders women and men - folders with images of people with the corresponding gender, - .csv file - contains information about the images and people in the dataset
keywords: biometric system, biometric system attacks, biometric dataset, face recognition database, face recognition dataset, face detection dataset, facial analysis, gender detection, supervised learning dataset, gender classification dataset, gender recognition dataset
The BED dataset
Version 1.0.0
Please cite as: Arnau-González, P., Katsigiannis, S., Arevalillo-Herráez, M., Ramzan, N., "BED: A new dataset for EEG-based biometrics", IEEE Internet of Things Journal, vol. 8, no. 15, pp. 12219 - 12230, 2021.
Disclaimer
While every care has been taken to ensure the accuracy of the data included in the BED dataset, the authors and the University of the West of Scotland, Durham University, and Universitat de València do not provide any guaranties and disclaim all responsibility and all liability (including without limitation, liability in negligence) for all expenses, losses, damages (including indirect or consequential damage) and costs which you might incur as a result of the provided data being inaccurate or incomplete in any way and for any reason. 2020, University of the West of Scotland, Scotland, United Kingdom.
Contact
For inquiries regarding the BED dataset, please contact:
Dataset summary
BED (Biometric EEG Dataset) is a dataset specifically designed to test EEG-based biometric approaches that use relatively inexpensive consumer-grade devices, more specifically the Emotiv EPOC+ in this case. This dataset includes EEG responses from 21 subjects to 12 different stimuli, across 3 different chronologically disjointed sessions. We have also considered stimuli aimed to elicit different affective states, so as to facilitate future research on the influence of emotions on EEG-based biometric tasks. In addition, we provide a baseline performance analysis to outline the potential of consumer-grade EEG devices for subject identification and verification. It must be noted that, in this work, EEG data were acquired in a controlled environment in order to reduce the variability in the acquired data stemming from external conditions.
The stimuli include:
For more details regarding the experimental protocol and the design of the dataset, please refer to the associated publication: Arnau-González, P., Katsigiannis, S., Arevalillo-Herráez, M., Ramzan, N., "BED: A new dataset for EEG-based biometrics", IEEE Internet of Things Journal, 2021. (Under review)
Dataset structure and contents
The BED dataset contains EEG recordings from 21 subjects, acquired during 3 similar sessions for each subject. The sessions were spaced one week apart from each other.
The BED dataset includes:
The dataset is organised in 3 folders:
RAW/ Contains the RAW files
RAW/sN/ Contains the RAW files associated with subject N
Each folder sN is composed by the following files:
- sN_s1.csv, sN_s2.csv, sN_s3.csv -- Files containing the EEG recordings for subject N and session 1, 2, and 3, respectively. These files contain 39 columns:
COUNTER INTERPOLATED F3 FC5 AF3 F7 T7 P7 O1 O2 P8 T8 F8 AF4 FC6 F4 ...UNUSED DATA... UNIX_TIMESTAMP
- subject_N_session_1_time_X.log, subject_N_session_2_time_X.log, subject_N_session_3_time_X.log -- Log files containing the sequence of events for the subject N and the session 1,2, and 3 respectively.
RAW_PARSED/
Contains Matlab files named sN_sM.mat. The files contain the recordings for the subject N in the session M. These files are composed by two variables:
- recording: size (time@256Hz x 17), Columns: COUNTER INTERPOLATED F3 FC5 AF3 F7 T7 P7 O1 O2 P8 T8 F8 AF4 FC6 F4 UNIX_TIMESTAMP
- events: cell array with size (events x 3) START_UNIX END_UNIX ADDITIONAL_INFO
START_UNIX is the UNIX timestamp in which the event starts
END_UNIX is the UNIX timestamp in which the event ends
ADDITIONAL INFO contains a struct with additional information regarding the specific event, in the case of the images, the expected score, the voted score, in the case of the cognitive task the input, in the case of the VEP the pattern and the frequency, etc..
Features/
Features/Identification
Features/Identification/[ARRC|MFCC|SPEC]/: Each of these folders contain the extracted features ready for classification for each of the stimuli, each file is composed by two variables, "feat" the feature matrix and "Y" the label matrix.
- feat: N x number of features
- Y: N x 2 (the #subject and the #session)
- INFO: Contains details about the event same as the ADDITIONAL INFO
Features/Verification: This folder is composed by 3 different files each of them with one different set of features extracted. Each file is composed by one cstruct array composed by:
- data: the time-series features, as described in the paper
- y: the #subject
- stimuli: the stimuli by name
- session: the #session
- INFO: Contains details about the event
The features provided are in sequential order, so index 1 and index 2, etc. are sequential in time if they belong to the same stimulus.
Additional information
For additional information regarding the creation of the BED dataset, please refer to the associated publication: Arnau-González, P., Katsigiannis, S., Arevalillo-Herráez, M., Ramzan, N., "BED: A new dataset for EEG-based biometrics", IEEE Internet of Things Journal, vol. 8, no. 15, pp. 12219 - 12230, 2021.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Gait recognition is the characterization of unqiue biometric patterns associated with each inidvidual which can be utilized to identify a person without direct contact. A public gain database with relatively large number of subjects can provide a great oppportunity to future studies to build and validate gait authentication models. The goal of this study is to introduce a comprehensive gait database of 93 human subjects who walked between two end points (320 meters) during two different sessions and record their gait data using two smart phones, one was attached to right thigh and another one on left side of waist. This data is collected with intention to be utilized by deep learning-based method which requires enough time points. The meta data including age, gender, smoking, daily exercise time, height, and weight of an individual is recorded. this data set is publicly available.
Except 19 subjects who did not attend for second session, every subject is associated with 4 different log files (each session contains two log files). Every file name has one of the following patterns: · sub0-lw-s1.csv: subject number 0, left waist, session 1 · sub0-rp-s1.csv: subject number 0, right thigh, session 1 · sub0-lw-s2.csv: subject number 0, left waist, session 2 · sub0-rp-s2.csv: subject number 0, right thigh, session 2 Every log file contains 58 features that are internally captured and calculated using SensorLog app. Additionally, an Excel file contain the meta data is provided for each subject.
Attribution-NonCommercial-NoDerivs 4.0 (CC BY-NC-ND 4.0)https://creativecommons.org/licenses/by-nc-nd/4.0/
License information was derived automatically
The dataset consists of 98,000 videos and selfies from 170 countries, providing a foundation for developing robust security systems and facial recognition algorithms.
While the dataset itself doesn't contain spoofing attacks, it's a valuable resource for testing liveness detection system, allowing researchers to simulate attacks and evaluate how effectively their systems can distinguish between real faces and various forms of spoofing.
By utilizing this dataset, researchers can contribute to the development of advanced security solutions, enabling the safe and reliable use of biometric technologies for authentication and verification. - Get the data
https://www.googleapis.com/download/storage/v1/b/kaggle-user-content/o/inbox%2F22059654%2Fe46e401a5449bacce5f934aaea9bb06e%2FFrame%20155.png?generation=1730591437955112&alt=media" alt="">
The dataset offers a high-quality collection of videos and photos, including selfies taken with a range of popular smartphones, like iPhone, Xiaomi, Samsung, and more. The videos showcase individuals turning their heads in various directions, providing a natural range of movements for liveness detection training.
https://www.googleapis.com/download/storage/v1/b/kaggle-user-content/o/inbox%2F22059654%2F8350718e93ee92840995405815739c61%2FFrame%20136%20(1).png?generation=1730591760432249&alt=media" alt="">
Furthermore, the dataset provides detailed metadata for each set, including information like gender, age, ethnicity, video resolution, duration, and frames per second. This rich metadata provides crucial context for analysis and model development.
Researchers can develop more accurate liveness detection algorithms, which is crucial for achieving the iBeta Level 2 certification, a benchmark for robust and reliable biometric systems that prevent fraud.
Attribution-NonCommercial-NoDerivs 4.0 (CC BY-NC-ND 4.0)https://creativecommons.org/licenses/by-nc-nd/4.0/
License information was derived automatically
The dataset consists of more than 42,000 video attacks of 7 different types specifically curated for evaluating liveness detection algorithms. The dataset aims to incorporate different scenarios and challenges to enable robust assessment and comparison of liveness detection systems.
The iBeta Liveness Detection Level 1 dataset serves as a benchmark for the development and assessment of liveness detection systems and for evaluating and improving the performance of algorithms.
https://www.googleapis.com/download/storage/v1/b/kaggle-user-content/o/inbox%2F12421376%2F72c514066335ed6b6d1d273d1c4198ef%2FMASK_TYPES.png?generation=1678913343424357&alt=media" alt="">
Each attack was filmed on an Apple iPhone and Google Pixel. The videos were filmed on various backgrounds and using additional accessories such as faked facial hair, scarves, hats, and others.
The dataset comprises videos of genuine facial presentations using various methods, including 3D masks and photos, as well as real and spoof faces. It proposes a novel approach that learns and extracts facial features to prevent spoofing attacks, based on deep neural networks and advanced biometric techniques.
Our results show that this technology works effectively in securing most applications and prevents unauthorized access by distinguishing between genuine and spoofed inputs. Additionally, it addresses the challenging task of identifying unseen spoofing cues, making it one of the most effective techniques in the field of anti-spoofing research.
keywords: ibeta level 1, ibeta level 2, liveness detection systems, liveness detection dataset, biometric dataset, biometric data dataset, replay attack dataset, biometric system attacks, anti-spoofing dataset, face liveness detection, deep learning dataset, face spoofing database, face anti-spoofing, presentation attack detection, presentation attack dataset, 2D print attacks, 3D print attacks, phone attack dataset, face anti spoofing, large-scale face anti spoofing, rich annotations anti spoofing dataset, cut prints spoof attack
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Off line signature database. It contains data from 960 individuals: 24 genuine signatures for each individual, plus 30 forgeries of his/her signature. The 24 genuine specimens of each signer were collected in a single day writing sessions. The forgeries were produced from the static image of the genuine signature. Each forger was allowed to practice the signature for as long as s/he wishes. Each forger imitated 3 signatures of 5 signers in a single day writing session. The genuine signatures shown to each forger are chosen randomly from the 24 genuine ones. Therefore for each genuine signature there are 30 skilled forgeries made by 10 forgers from 10 different genuine specimens. The signatures are in "bmp" format, in black and white and 300 dpi. The files of the genuine signatures are named xxx\c-xxx-yy.bmp and the files of the forgeries are named xxx\cf-xxx-yy.bmp where xxx is the number of the signer and yy its repetition.
Download the license agreement, fill it out, sign it, and fax it to +34 928 451 243 (Attn. Miguel A. Ferrer) Send an email to gpds@gi.ulpgc.es, as follows: Subject DATABASE download Body: Your name, e-mail, phone number, organization, postal mail, Database you require, purpose for which you will use the database, time and date at which you sent the fax with the signed license agreement. Along with the email send a pdf file with the signed license agreement. Send by postal mail the original of your signed license agreement to the following address: Miguel Ángel Ferrer Ballester Departamento de Señales y Comunicaciones Universidad de Las Palmas de Gran Canaria Campus de Tafira s/n 35017 Las Palmas de Gran Canaria, SPAIN Once the fax and email of the license agreement have been received, you will receive an email with instructions to download the database. After you finish the download, please notify by email that you have successfully completed the transaction. For more information, please contact: gpds@gi.ulpgc.es
The 3D Mask Attack Database (3DMAD) is a biometric (face) spoofing database. It contains 76500 frames of 17 persons, recorded using Kinect for both real-access and spoofing attacks. Each frame consists of:
a depth image (640x480 pixels – 1x11 bits)
the corresponding RGB image (640x480 pixels – 3x8 bits)
manually annotated eye positions (with respect to the RGB image).
The data is collected in 3 different sessions for all subjects and for each session 5 videos of 300 frames are captured. The recordings are done under controlled conditions, with frontal-view and neutral expression. The first two sessions are dedicated to the real access samples, in which subjects are recorded with a time delay of ~2 weeks between the acquisitions. In the third session, 3D mask attacks are captured by a single operator (attacker).
In each video, the eye-positions are manually labelled for every 1st, 61st, 121st, 181st, 241st and 300th frames and they are linearly interpolated for the rest.
The real-size masks are obtained using "ThatsMyFace.com". The database additionally contains the face images used to generate these masks (1 frontal and 2 profiles) and paper-cut masks that are also produced by the same service and using the same images.
The satellite package which contains the Bob accessor methods to use this database directly from Python, with the certified protocols, is available in two different distribution formats:
You can download it from PyPI, or
You can download it in its source form from its git repository.
Acknowledgments
If you use this database, please cite the following publication:
Nesli Erdogmus and Sébastien Marcel, "Spoofing in 2D Face Recognition with 3D Masks and Anti-spoofing with Kinect", Biometrics: Theory, Applications and Systems, 2013. 10.1109/BTAS.2013.6712688 https://publications.idiap.ch/index.php/publications/show/2657
Attribution-NonCommercial-NoDerivs 4.0 (CC BY-NC-ND 4.0)https://creativecommons.org/licenses/by-nc-nd/4.0/
License information was derived automatically
The dataset consists of images of people for detection and segmentation of hairs within the oval region of the face. It primarily focuses on identifying the presence of hair strands within the facial area and accurately segmenting them for further analysis or applications.
The dataset contains a diverse collection of images depicting people with different hair styles, colors, lengths, and textures. Each image is annotated with annotations that indicate the boundaries and contours of the individual hair strands within the oval of the face.
The dataset can be utilized for various purposes, such as developing machine learning models or algorithms for hair detection and segmentation. It can also be used for research in facial recognition, virtual try-on applications, hairstyle recommendation systems, and other related areas.
https://www.googleapis.com/download/storage/v1/b/kaggle-user-content/o/inbox%2F12421376%2F81b5a9e6c755e04d97fc6b175a127432%2FMacBook%20Air%20-%201.png?generation=1691561622573906&alt=media" alt="">
Each image from images
folder is accompanied by an XML-annotation in the annotations.xml
file indicating the coordinates of the bounding boxes and labels for parking spaces. For each point, the x and y coordinates are provided.
https://www.googleapis.com/download/storage/v1/b/kaggle-user-content/o/inbox%2F12421376%2Fb634cd569d4bf7a253ac7a0e7a91ef7e%2Fcarbon.png?generation=1691562068420789&alt=media" alt="">
keywords: biometric dataset, biometric data dataset, face recognition database, face recognition dataset, face detection dataset, facial analysis, human images dataset, hair detection, hair segmentation,human hair segmentation, image segmentation, images dataset, computer vision, deep learning dataset, scalp, augmented reality, ar
CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
License information was derived automatically
Explore the Sokoto Coventry Fingerprint Dataset (SOCOFing), a robust biometric database featuring 6,000 fingerprint images from 600 African subjects. Ideal for academic research, SOCOFing includes detailed labels and synthetic alterations to enhance fingerprint recognition and security studies.