Attribution-NonCommercial-NoDerivs 4.0 (CC BY-NC-ND 4.0)https://creativecommons.org/licenses/by-nc-nd/4.0/
License information was derived automatically
Dataset of face images with different angles and head positions
Dataset contains 23,110 individuals, each contributing 28 images featuring various angles and head positions, diverse backgrounds, and attributes, along with 1 ID photo. In total, the dataset comprises over 670,000 images in formats such as JPG and PNG. It is designed to advance face recognition and facial recognition research, focusing on person re-identification and recognition systems. By utilizing this dataset… See the full description on the dataset page: https://huggingface.co/datasets/UniDataPro/face-re-identification-image-dataset.
https://www.datainsightsmarket.com/privacy-policyhttps://www.datainsightsmarket.com/privacy-policy
The Face Key Point Detection market is experiencing robust growth, driven by the increasing adoption of facial recognition technologies across diverse sectors. This surge is fueled by advancements in deep learning algorithms, leading to improved accuracy and efficiency in detecting key facial features. Applications span a wide range, from security and surveillance systems leveraging face recognition for authentication and identification, to the burgeoning field of emotion AI, using expression recognition for personalized user experiences. Head pose recognition further enhances the capabilities, enabling more natural and intuitive human-computer interactions. The market is segmented by application (Face Recognition, Expression Recognition, Head Pose Recognition, Others) and by the methods used (Holistic Approach, Constrained Local Model (CLM) Method, Regression-Based Methods). The holistic approach, offering a comprehensive analysis of the entire face, is currently dominant, although CLM and regression-based methods are gaining traction due to their computational efficiency. Major players like ULUCU, Roboflow, Oosto, and MathWorks are driving innovation and market penetration, while platforms like GitHub and Kaggle facilitate community development and resource sharing. Geographical growth is widespread, with North America and Europe currently leading the market due to higher technological adoption and infrastructure. However, the Asia-Pacific region is poised for significant expansion fueled by rapid technological advancements and growing demand in sectors like security and consumer electronics. The market's continued growth trajectory is projected to be influenced by several factors. The increasing availability of large, high-quality facial datasets for training advanced algorithms will enhance accuracy and reliability. Furthermore, the integration of face key point detection with other technologies like augmented reality (AR) and virtual reality (VR) will unlock new applications in entertainment, healthcare, and retail. However, challenges remain, including concerns surrounding data privacy and ethical considerations related to facial recognition technology. Addressing these concerns through robust regulatory frameworks and responsible development practices will be crucial for the market's sustainable and ethical growth. Considering a conservative CAGR of 15% (a reasonable estimate given the rapid technological advancements in this space), the market size, currently estimated around $1.5 billion in 2025, is likely to exceed $4 billion by 2033.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
The HFRD (v2020) dataset includes 4835 images of masked human faces. Originally, the faces were taken from three different publicly available datasets: MUCT, FASSEG, and AT&T datasets. The collected subsets of the HFRD dataset contain standard images that are previously used in the field of facial recognition, and then we added masks to cover the nose, mouth, and chin area, as for what remains visible from the face are the eyes, forehead, and hair. Thus, we have introduced new standard images for use in the field of partial facial recognition (PFR) domain. The new structure of the current HFRD store inherits the structure of each original dataset, which means there are three main datasets: MaskedMUCT contains five folders, MaskedFASSEG contains four folders, and MaskedAT&T has organized in 40 folders. This store of HFRD masked face datasets are not only inheriting the structure of the original datasets, but also inheriting the varsity of them. Finally, all folders consist of facial images surrounded and covered by various shapes of masks. Consequently, the proposed new enhanced HFRD datasets have been developed based on the impact of the pandemic COVID-2019 stage. Our enhanced improved Human Face Recognition Datasets (HFRD) datasets could be used to test studies of FR algorithms for human identification. The obtained outcomes from this enhanced HFRD data can also be useful in providing more knowledge for the Artificial Intelligence (AI) tools, and decision support system for predicting the spread of COVID-2019.
Attribution-NonCommercial-ShareAlike 4.0 (CC BY-NC-SA 4.0)https://creativecommons.org/licenses/by-nc-sa/4.0/
License information was derived automatically
Contact Information
If you would like further information about the RAVDESS Facial Landmark Tracking data set, or if you experience any issues downloading files, please contact us at ravdess@gmail.com.
Tracking Examples
Watch a sample of the facial tracking results.
Description
The RAVDESS Facial Landmark Tracking dataset set contains tracked facial landmark movements from the Ryerson Audio-Visual Database of Emotional Speech and Song (RAVDESS) [RAVDESS Zenodo page]. Motion tracking of actors' faces was produced by OpenFace 2.1.0 (Baltrusaitis, T., Zadeh, A., Lim, Y. C., & Morency, L. P., 2018). Tracked information includes: facial landmark detection, head pose estimation, facial action unit recognition, and eye-gaze estimation.
The Facial Landmark Tracking dataset was created in the Affective Data Science Lab.
This data set contains tracking for all 2452 RAVDESS trials. All tracking movement data are contained in "FacialTracking_Actors_01-24.zip", which contains 2452 .CSV files. Each actor has 104 tracked trials (60 speech, 44 song). Note, there are no song files for Actor 18.
Total Tracked Files = (24 Actors x 60 Speech trials) + (23 Actors x 44 Song trials) = 2452 files.
Tracking results for each trial are provided as individual comma separated value files (CSV format). File naming convention of tracked files is identical to that of the RAVDESS. For example, tracked file "01-01-01-01-01-01-01.csv" corresponds to RAVDESS audio-video file "01-01-01-01-01-01-01.mp4". For a complete description of the RAVDESS file naming convention and experimental manipulations, please see the RAVDESS Zenodo page.
Tracking overlay videos for all trials are also provided (720p Xvid, .avi), one zip file per Actor. As the RAVDESS does not contain "ground truth" facial landmark locations, the overlay videos provide a visual 'sanity check' for researchers to confirm the general accuracy of the tracking results. The file naming convention of tracking overlay videos also matches that of the RAVDESS. For example, tracking video "01-01-01-01-01-01-01.avi" corresponds to RAVDESS audio-video file "01-01-01-01-01-01-01.mp4".
Tracking File Output Format
This data set retained OpenFace's data output format, described here in detail. The resolution of all input videos was 1280x720. When tracking output units are in pixels, their range of values is (0,0) (top left corner) to (1280,720) (bottom right corner).
Columns 1-3 = Timing and Detection Confidence
Columns 4-291 = Eye Gaze Detection
Columns 292-297 = Head pose
Columns 298-433 = Facial Landmarks locations in 2D
Columns 434-637 = Facial Landmarks locations in 3D
Columns 638-677 = Rigid and non-rigid shape parameters
Parameters of a point distribution model (PDM) that describe the rigid face shape (location, scale and rotation) and non-rigid face shape (deformation due to expression and identity). For more details, please refer to chapter 4.2 of my Tadas Baltrusaitis's PhD thesis [download link].
Columns 687-712 = Facial Action Units
Facial Action Units (AUs) are a way to describe human facial movements (Ekman, Friesen, and Hager, 2002) [wiki link]. More information on OpenFace's implementation of AUs can be found here.
Note, OpenFace's columns 2 and 5 (face_id and success, respectively) were not included in this data set. These values were redundant as a single face was detected in all frames, in all 2452 trials.
Tracking Overlay Videos
Tracking overlay videos visualize most aspects of the tracking output described above.
Camera Parameters and 3D Calibration Procedure
This data set contains accurate estimates of actors' 3D head poses. To produce these, camera parameters at the time of recording were required (distance from camera to actor, and camera field of view). These values were used with OpenCV's camera calibration procedure, described here, to produce estimates of the camera's focal length and optical center at the time of actor recordings. The four values produced by the calibration procedure (fx,fy,cx,cy) were input to OpenFace as command line arguments during facial tracking, described here, to produce accurate estimates of 3D head pose.
Camera
CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
License information was derived automatically
Isi na olu Semantic Segmentation Dataset emebere maka e-azụmahịa & mkpọsa na ngalaba mgbasa ozi & ntụrụndụ, na-egosipụta mkpokọta eserese eserese AI mepụtara nwere mkpebi dị elu karịa 1024 x 1024 pikselụ. Nke a dataset na-elekwasị anya na semantic segmentation, kpọmkwem ezubere iche isi agwa, gụnyere ihu, ntutu isi, na ihe ọ bụla ngwa, nakwa dị ka n'olu ebe ruo n'olu, na-enye ohere maka obere, adịghị ekewa akụkụ n'akụkụ.
CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
License information was derived automatically
The Head and Neck Semantic Segmentation Dataset is designed for the e-commerce & retail and media & entertainment sectors, featuring a collection of AI-generated cartoon images with resolutions above 1024 x 1024 pixels. This dataset focuses on semantic segmentation, specifically targeting the main character's head, including face, hair, and any accessories, as well as the neck area up to the collarbone, with an allowance for small, unsegmented parts on the edges.
CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
License information was derived automatically
Ny Dataset Semantic Segmentation Head and Neck dia natao ho an'ny sehatry ny e-varotra & antsinjarany ary haino aman-jery & fialamboly, ahitana fitambarana sary sariitatra noforonin'ny AI miaraka amin'ny fanapahan-kevitra mihoatra ny 1024 x 1024 piksel. Ity tahirin-kevitra ity dia mifantoka amin'ny fizarana semantika, mikendry manokana ny lohan'ilay mpilalao fototra, ao anatin'izany ny tarehy, ny volo, ary ny kojakoja rehetra, ary koa ny faritry ny vozony ka hatreo amin'ny taolam-paty, miaraka amin'ny fanomezan-dàlana ho an'ny ampahany kely tsy voazara amin'ny sisiny.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Facial expression recognition faces great challenges due to factors such as face similarity, image quality, and age variation. Although various existing end-to-end Convolutional Neural Network (CNN) architectures have achieved good classification results in facial expression recognition tasks, these network architectures share a common drawback that the convolutional kernel can only compute the correlation between elements of a localized region when extracting expression features from an image. This leads to difficulties for the network to explore the relationship between all the elements that make up a complete expression. In response to this issue, this article proposes a facial expression recognition network called HFE-Net. In order to capture the subtle changes of expression features and the whole facial expression information at the same time, HFE-Net proposed a Hybrid Feature Extraction Block. Specifically, Hybrid Feature Extraction Block consists of parallel Feature Fusion Device and Multi-head Self-attention. Among them, Feature Fusion Device not only extracts the local information in expression features, but also measures the correlation between distant elements in expression features, which helps the network to focus more on the target region while realizing the information interaction between distant features. And Multi-head Self-attention can calculate the correlation between the overall elements in the feature map, which helps the network to extract the overall information of the expression features. We conducted a lot of experiments on four publicly available facial expression datasets and verified that the Hybrid Feature Extraction Block constructed in this paper can improve the network’s recognition ability for facial expressions.
CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
License information was derived automatically
I-Head and Neck Semantic Segmentation Dataset yakhelwe imikhakha ye-e-commerce & yokuthengisa nemidiya nezokuzijabulisa, ehlanganisa iqoqo lezithombe zopopayi ezikhiqizwe yi-AI ezinezinqumo ezingaphezu kwamaphikseli angu-1024 x 1024. Le dathasethi igxile ekuhlukaniseni kwe-semantic, eqondise ngokukhethekile ikhanda lomlingiswa oyinhloko, okuhlanganisa ubuso, izinwele, nanoma yiziphi izisetshenziswa, kanye nendawo yentamo kuze kufike kukholomu, nemvume yezingxenye ezincane, ezingahlukanisiwe emaphethelweni.
CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
License information was derived automatically
د سر او غاړې د سیمانټیک سیګمینټیشن ډیټا سیټ د ای کامرس او پرچون پلور او رسنیو او تفریحي سکتورونو لپاره ډیزاین شوی، چې د 1024 x 1024 پکسلز څخه پورته ریزولوشن سره د AI لخوا رامینځته شوي کارټون عکسونو ټولګه وړاندې کوي. دا ډیټا سیمانټیک سیګمینټیشن باندې تمرکز کوي، په ځانګړي توګه د اصلي کرکټر سر په نښه کوي، پشمول د مخ، ویښتو، او هر ډول لوازمو، او همدارنګه د غاړې ساحه تر کالر هډوکي پورې، په څنډو کې د کوچنیو، غیر سیګمینټ شوي برخو لپاره د تخصیص سره.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Comparison of experimental results on the FERPlus dataset.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Comparison of experimental results on the RAF-DB dataset.
CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
License information was derived automatically
هيڊ اينڊ نيڪ سيمينٽڪ سيگمينٽيشن ڊيٽا سيٽ اي-ڪامرس ۽ پرچون ۽ ميڊيا ۽ تفريحي شعبن لاءِ ٺاهيو ويو آهي، جنهن ۾ 1024 x 1024 پڪسل کان مٿي ريزوليوشن سان AI-جنريٽر ٿيل ڪارٽون تصويرن جو مجموعو شامل آهي. هي ڊيٽا سيٽ سيمينٽڪ سيگمينٽيشن تي ڌيان ڏئي ٿو، خاص طور تي مکيه ڪردار جي مٿي کي نشانو بڻائي ٿو، جنهن ۾ چهرو، وار، ۽ ڪنهن به لوازمات شامل آهن، انهي سان گڏ ڳچيءَ جو علائقو ڪالر بون تائين، ڪنارن تي ننڍڙن، غير سيگمينٽ ٿيل حصن لاءِ الاؤنس سان.
CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
License information was derived automatically
Dataset Segmentasi Semantik Kepala sareng Beuheung dirancang pikeun séktor e-commerce & ritel sareng média & hiburan, nampilkeun kumpulan gambar kartun anu dibangkitkeun AI kalayan résolusi di luhur 1024 x 1024 piksel. Dataset ieu museurkeun kana ségméntasi semantik, khususna nargétkeun sirah tokoh utama, kalebet raray, rambut, sareng asesoris naon waé, ogé daérah beuheung dugi ka tulang selangka, kalayan sangu pikeun bagian-bagian anu alit, henteu diségmentasi dina ujungna.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Multilayer perceptron, Multi-head Self-attention and Feature Fusion Device ablation experiments on four FER datasets.
CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
License information was derived automatically
A Head and Neck Semantic Segmentation Dataset az e-kereskedelem és a kiskereskedelem, valamint a média és a szórakoztatóipar számára készült, és mesterséges intelligencia által generált, 1024 x 1024 pixel feletti felbontású rajzfilmek gyűjteményét tartalmazza. Ez az adatkészlet a szemantikai szegmentációra összpontosít, különös tekintettel a főszereplő fejére, beleértve az arcát, a haját és az esetleges kiegészítőket, valamint a nyak területére a kulcscsontig, a széleken lévő apró, szegmentálatlan részek figyelembevételével.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Comparison of experimental results on the Affectnet dataset.
CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
License information was derived automatically
Bosh va boʻyinning semantik segmentatsiyasi maʼlumotlar toʻplami elektron tijorat, chakana savdo hamda media va koʻngilochar sektorlar uchun moʻljallangan boʻlib, AI tomonidan yaratilgan 1024 x 1024 pikseldan yuqori oʻlchamdagi multfilm tasvirlari toʻplamini oʻz ichiga oladi. Ushbu ma'lumotlar to'plami semantik segmentatsiyaga qaratilgan, xususan, bosh qahramonning boshi, shu jumladan yuzi, sochlari va har qanday aksessuarlari, shuningdek, bo'yinbog'igacha bo'yin sohasi, chekkalarda kichik, bo'linmagan qismlarga ruxsat berilgan.
CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
License information was derived automatically
Series datorum Segmentationis Semanticae Capitis et Colli destinata est sectoribus commercii electronici et venditionis minutae, instrumentorum communicationis socialis et oblectationis, collectionem imaginum animatarum ab intellegentia artificiali generatarum cum resolutionibus supra 1024 x 1024 elementa punctorum exhibens. Haec series datorum in segmentatione semantica intendit, praecipue caput personae principalis spectans, incluso facie, capillis, et quibuslibet ornamentis, necnon regione colli usque ad claviculam, cum spatio parvorum partium non segmentatarum in marginibus.
CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
License information was derived automatically
O conxunto de datos de segmentación semántica de cabeza e pescozo está deseñado para os sectores do comercio electrónico e a venda polo miúdo, así como dos medios de comunicación e o entretemento, e inclúe unha colección de imaxes de debuxos animados xeradas por IA con resolucións superiores a 1024 x 1024 píxeles. Este conxunto de datos céntrase na segmentación semántica, dirixíndose especificamente á cabeza do personaxe principal, incluíndo a cara, o cabelo e calquera accesorio, así como a zona do pescozo ata a clavícula, cunha marxe para as partes pequenas non segmentadas nos bordos.
Attribution-NonCommercial-NoDerivs 4.0 (CC BY-NC-ND 4.0)https://creativecommons.org/licenses/by-nc-nd/4.0/
License information was derived automatically
Dataset of face images with different angles and head positions
Dataset contains 23,110 individuals, each contributing 28 images featuring various angles and head positions, diverse backgrounds, and attributes, along with 1 ID photo. In total, the dataset comprises over 670,000 images in formats such as JPG and PNG. It is designed to advance face recognition and facial recognition research, focusing on person re-identification and recognition systems. By utilizing this dataset… See the full description on the dataset page: https://huggingface.co/datasets/UniDataPro/face-re-identification-image-dataset.