100+ datasets found
  1. m

    THVD (Talking Head Video Dataset)

    • data.mendeley.com
    Updated Apr 2, 2025
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Mario Peedor (2025). THVD (Talking Head Video Dataset) [Dataset]. http://doi.org/10.17632/ykhw8r7bfx.1
    Explore at:
    Dataset updated
    Apr 2, 2025
    Authors
    Mario Peedor
    License

    Attribution-NonCommercial 3.0 (CC BY-NC 3.0)https://creativecommons.org/licenses/by-nc/3.0/
    License information was derived automatically

    Description

    About

    We provide a comprehensive talking-head video dataset with over 50,000 videos, totaling more than 500 hours of footage and featuring 23,841 unique identities from around the world.

    Distribution

    Detailing the format, size, and structure of the dataset: Data Volume: -Total Size: 2.5TB

    -Total Videos: 47,200

    -Identities Covered: 23,000

    -Resolution: 60% 4k(1980), 33% fullHD(1080)

    -Formats: MP4

    -Full-length videos with visible mouth movements in every frame.

    -Minimum face size of 400 pixels.

    -Video durations range from 20 seconds to 5 minutes.

    -Faces have not been cut out, full screen videos including backgrounds.

    Usage

    This dataset is ideal for a variety of applications:

    Face Recognition & Verification: Training and benchmarking facial recognition models.

    Action Recognition: Identifying human activities and behaviors.

    Re-Identification (Re-ID): Tracking identities across different videos and environments.

    Deepfake Detection: Developing methods to detect manipulated videos.

    Generative AI: Training high-resolution video generation models.

    Lip Syncing Applications: Enhancing AI-driven lip-syncing models for dubbing and virtual avatars.

    Background AI Applications: Developing AI models for automated background replacement, segmentation, and enhancement.

    Coverage

    Explaining the scope and coverage of the dataset:

    Geographic Coverage: Worldwide

    Time Range: Time range and size of the videos have been noted in the CSV file.

    Demographics: Includes information about age, gender, ethnicity, format, resolution, and file size.

    Languages Covered (Videos):

    English: 23,038 videos

    Portuguese: 1,346 videos

    Spanish: 677 videos

    Norwegian: 1,266 videos

    Swedish: 1,056 videos

    Korean: 848 videos

    Polish: 1,807 videos

    Indonesian: 1,163 videos

    French: 1,102 videos

    German: 1,276 videos

    Japanese: 1,433 videos

    Dutch: 1,666 videos

    Indian: 1,163 videos

    Czech: 590 videos

    Chinese: 685 videos

    Italian: 975 videos

    Who Can Use It

    List examples of intended users and their use cases:

    Data Scientists: Training machine learning models for video-based AI applications.

    Researchers: Studying human behavior, facial analysis, or video AI advancements.

    Businesses: Developing facial recognition systems, video analytics, or AI-driven media applications.

    Additional Notes

    Ensure ethical usage and compliance with privacy regulations. The dataset’s quality and scale make it valuable for high-performance AI training. Potential preprocessing (cropping, down sampling) may be needed for different use cases. Dataset has not been completed yet and expands daily, please contact for most up to date CSV file. The dataset has been divided into 100GB zipped files and is hosted on a private server (with the option to upload to the cloud if needed). To verify the dataset's quality, please contact me for the full CSV file. I’d be happy to provide example videos selected by the potential buyer.

  2. Youtube video statistics for 1 million videos

    • kaggle.com
    zip
    Updated Jun 29, 2020
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Mattia Zeni (2020). Youtube video statistics for 1 million videos [Dataset]. https://www.kaggle.com/mattiazeni/youtube-video-statistics-1million-videos
    Explore at:
    zip(6696303511 bytes)Available download formats
    Dataset updated
    Jun 29, 2020
    Authors
    Mattia Zeni
    License

    Open Data Commons Attribution License (ODC-By) v1.0https://www.opendatacommons.org/licenses/by/1.0/
    License information was derived automatically

    Description

    Motivation

    Study how YouTube videos become viral or, more in general, how they evolve in terms of views, likes and subscriptions is a topic of interest in many disciplines. With this dataset you can study such phenomena, with statistics about 1 million YouTube videos. The information was collected in 2013 when YouTube was exposing the data publicly: they removed this functionality in the years and now it's possible to have such statistics only to the owner of the video. This makes this dataset unique.

    Context

    This Dataset has been generated with YOUStatAnalyzer, a tool developed by myself (Mattia Zeni) when I was working for CREATE-NET (www.create-net.org) within the framework of the CONGAS FP7 project (http://www.congas-project.eu). For the project we needed to collect and analyse the dynamics of YouTube videos popularity. The dataset contains statistics of more than 1 million Youtube videos, chosen accordingly to random keywords extracted from the WordNet library (http://wordnet.princeton.edu).

    The motivation that led us to the development of the YOUStatAnalyser data collection tool and the creation of this dataset is that there's an active research community working on the interplay among user individual preferences, social dynamics, advertising mechanisms and a common problem is the lack of open large-scale datasets. At the same time, no tool was present at that time. Today, YouTube removed the possibility to visualize these data on each video's page, making this dataset unique.

    When using our dataset for research purposes, please cite it as:

    @INPROCEEDINGS{YOUStatAnalyzer, author={Mattia Zeni and Daniele Miorandi and Francesco {De Pellegrini}}, title = {{YOUStatAnalyzer}: a Tool for Analysing the Dynamics of {YouTube} Content Popularity}, booktitle = {Proc.\ 7th International Conference on Performance Evaluation Methodologies and Tools (Valuetools, Torino, Italy, December 2013)}, address = {Torino, Italy}, year = {2013} }

    Content

    The dataset contains statistics and metadata of 1 million YouTube videos, collected in 2013. The videos have been chosen accordingly to random keywords extracted from the WordNet library (http://wordnet.princeton.edu).

    Dataset structure

    The structure of a dataset is the following: { u'_id': u'9eToPjUnwmU', u'title': u'Traitor Compilation # 1 (Trouble ...', u'description': u'A traitor compilation by one are ...', u'category': u'Games', u'commentsNumber': u'6', u'publishedDate': u'2012-10-09T23:42:12.000Z', u'author': u'ServilityGaming', u'duration': u'208', u'type': u'video/3gpp', u'relatedVideos': [u'acjHy7oPmls', u'EhW2LbCjm7c', u'UUKigFAQLMA', ...], u'accessControl': { u'comment': {u'permission': u'allowed'}, u'list': {u'permission': u'allowed'}, u'videoRespond': {u'permission': u'moderated'}, u'rate': {u'permission': u'allowed'}, u'syndicate': {u'permission': u'allowed'}, u'embed': {u'permission': u'allowed'}, u'commentVote': {u'permission': u'allowed'}, u'autoPlay': {u'permission': u'allowed'} }, u'views': { u'cumulative': { u'data': [15.0, 25.0, 26.0, 26.0, ...] }, u'daily': { u'data': [15.0, 10.0, 1.0, 0.0, ..] } }, u'shares': { u'cumulative': { u'data': [0.0, 0.0, 0.0, 0.0, ...] }, u'daily': { u'data': [0.0, 0.0, 0.0, 0.0, ...] } }, u'watchtime': { u'cumulative': { u'data': [22.5666666667, 36.5166666667, 36.7, 36.7, ...] }, u'daily': { u'data': [22.5666666667, 13.95, 0.166666666667, 0.0, ...] } }, u'subscribers': { u'cumulative': { u'data': [0.0, 0.0, 0.0, 0.0, ...] }, u'daily': { u'data': [-1.0, 0.0, 0.0, 0.0, ...] } }, u'day': { u'data': [1349740800000.0, 1349827200000.0, 1349913600000.0, 1350000000000.0, ...] } }

    From the structure above is possible to see which fields an entry in the dataset has. It is possible to divide them into 2 sections:

    1) Video Information.

    _id -> Corresponding to the video ID and to the unique identifier of an entry in the database. title -> Te video's title. description -> The video's description. category -> The YouTube category the video is inserted in. commentsNumber -> The number of comments posted by users. publishedDate -> The date the video has been published. author -> The author of the video. duration -> The video duration in seconds. type -> The encoding type of the video. relatedVideos -> A list of related videos. accessControl -> A list of access policies for different aspects related to the video.

    2) Video Statistics.

    Each video can have 4 different statistics variables: views, shares, subscribers and watchtime. Recent videos have all of them while older video can have only the 'views' variable. Each variable has 2 dimensions, daily and cumulative.

    views -> number of views collected by the video. shares -> number of sharing operations performed by users. watchtime -> the time spent by users watching the video, in minute. subscribers -> number of subscriptions to the channel the video is inserted in, caused by the selected video. day -> a list of days indicating the analysed period for the statistic.

    Data usage

    In the case you are using mongoDB as database system, you can import our dataset using the command:

    mongoimport --db [MONGODB_NAME] --collection [MONGODB_COLLECTION] --file dataset.json

    Once you imported the Dataset in your DB, you can access the data performing queries. Let's present some example code in python in order to perform queries.

    The following code will perform a query without research parameters, returning all the entries in the database, each one saved into the variable entry:

    client = MongoClient('localhost', 27017) db = client[MONGODB_NAME] collection = db[MONGODB_COLLECTION] for entry in db.collection.find(): print entry["day"]["data"]

    If you want to restrict the results to some entries that answer to a specified query you can use:

    client = MongoClient('localhost', 27017) db = client[MONGODB_NAME] collection = db[MONGODB_COLLECTION] for entry in (db.collection.find({"watchtime":{ "$exists": True }})) and (db.collection.find({"category":"Music"})): print entry["day"]["data"]

  3. m

    Honeybee video tracking data

    • bridges.monash.edu
    • researchdata.edu.au
    bin
    Updated May 31, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Malika Nisal Ratnayake; Adrian Dyer; Alan Dorin (2023). Honeybee video tracking data [Dataset]. http://doi.org/10.26180/5f4c8d5815940
    Explore at:
    binAvailable download formats
    Dataset updated
    May 31, 2023
    Dataset provided by
    Monash University
    Authors
    Malika Nisal Ratnayake; Adrian Dyer; Alan Dorin
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Monitoring animals in their natural habitat is essential for the advancement of animal behavioural studies, especially in pollination studies. We present a novel hybrid detection and tracking algorithm "HyDaT" to monitor unmarked insects outdoors. Our software can detect an insect, identify when a tracked insect becomes occluded from view and when it re-emerges, determine when an insect exits the camera field of view, and our software assembles a series of insect locations into a coherent trajectory. The insect detecting component of the software uses background subtraction and deep learning-based detection together to accurately and efficiently locate the insect.This dataset includes videos of honeybees foraging in two ground-covers Scaevola and Lamb's-ear, comprising of complex background detail, wind-blown foliage, and honeybees moving into and out of occlusion beneath leaves and among three-dimensional plant structures. Honeybee tracks and associated outputs of experiments extracted using HyDaT algorithm are included in the dataset. The dataset also contains annotated images and pre-trained YOLOv2 object detection models of honeybees.

  4. YouTube Trending Video Dataset (updated daily)

    • kaggle.com
    zip
    Updated Apr 15, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Rishav Sharma (2024). YouTube Trending Video Dataset (updated daily) [Dataset]. https://www.kaggle.com/rsrishav/youtube-trending-video-dataset
    Explore at:
    zip(0 bytes)Available download formats
    Dataset updated
    Apr 15, 2024
    Authors
    Rishav Sharma
    License

    https://creativecommons.org/publicdomain/zero/1.0/https://creativecommons.org/publicdomain/zero/1.0/

    Area covered
    YouTube
    Description

    This dataset is a daily record of the top trending YouTube videos and it will be updated daily.

    Context

    YouTube maintains a list of the top trending videos on the platform. According to Variety magazine, “To determine the year’s top-trending videos, YouTube uses a combination of factors including measuring users interactions (number of views, shares, comments and likes). Note that they’re not the most-viewed videos overall for the calendar year”.

    Note that this dataset is a structurally improved version of this dataset.

    Content

    This dataset includes several months (and counting) of data on daily trending YouTube videos. Data is included for the IN, US, GB, DE, CA, FR, RU, BR, MX, KR, and JP regions (India, USA, Great Britain, Germany, Canada, France, Russia, Brazil, Mexico, South Korea, and, Japan respectively), with up to 200 listed trending videos per day.

    Each region’s data is in a separate file. Data includes the video title, channel title, publish time, tags, views, likes and dislikes, description, and comment count.

    The data also includes a category_id field, which varies between regions. To retrieve the categories for a specific video, find it in the associated JSON. One such file is included for each of the 11 regions in the dataset.

    For more information on specific columns in the dataset refer to the column metadata.

    Acknowledgements

    This dataset was collected using the YouTube API. This dataset is the updated version of Trending YouTube Video Statistics.

    Inspiration

    Possible uses for this dataset could include: - Sentiment analysis in a variety of forms - Categorizing YouTube videos based on their comments and statistics. - Training ML algorithms like RNNs to generate their own YouTube comments. - Analyzing what factors affect how popular a YouTube video will be. - Statistical analysis over time .

    For further inspiration, see the kernels on this dataset!

  5. d

    Selfie Video Dataset | 3K+ videos | Global Coverage | Face & Voice...

    • datarade.ai
    .mp4, .mov
    Updated Aug 1, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    FileMarket (2025). Selfie Video Dataset | 3K+ videos | Global Coverage | Face & Voice Biometrics | Computer-Vision Data [Dataset]. https://datarade.ai/data-products/selfie-video-dataset-3k-videos-global-coverage-face-filemarket
    Explore at:
    .mp4, .movAvailable download formats
    Dataset updated
    Aug 1, 2025
    Dataset authored and provided by
    FileMarket
    Area covered
    Costa Rica, Cuba, Jersey, Iceland, Holy See, Lesotho, San Marino, State of, Estonia, Suriname
    Description

    Every recording captures a single subject performing slow head sweeps (left ↔ right ↕ up ↕ down) while counting “one … ten” in English—yielding synchronized face, lips, and voice data.

    1. Scope & Volume
    2. Videos: 3 076 MP4 / MOV clips
    3. Typical length: 18-22 seconds (target 20 s)
    4. Participants: 2403 (≈1.28 clip per person)

    5. Capture Protocol

    6. The dataset was assembled through a GDPR-compliant crowdsourcing task focused on secure-transaction AI. Contributors followed a strict brief:

    7. Environment – indoor, even lighting, plain or uncluttered background, no back-lighting or shadows.

    8. Appearance – full face visible; no glasses, hats, masks, filters, or overlays.

    9. Action – look straight at the camera, then slowly rotate head left, right, up, down while maintaining gaze; finish by speaking the ten-count.

    10. Duration – ~20 s continuous take, 30 fps or higher.

    11. Framing – single person, shoulders-up composition; no other people, pets, or distractions.

    12. All submissions passed automated and manual QC for framing, focus, lighting, and audio intelligibility.

    13. Demographic Breakdown

    Gender: Male - 84.2 % Female - 15.8 %

    Ethnicity: African - 69.3% South Asian - 10.3% South-East Asian - 9.7% European - 4.0% Middle East - 3.4% Arab - 1.9% Latino - 1.2% East Asian - 0.2%

    Age distribution: <18 - 5.42% 18-25 - 48.88% 25-30 - 20.92% 30-40 - 18.17% 40-50 - 5.35% 60+ - 1.26%

  6. DeepFake Videos Dataset

    • kaggle.com
    Updated Jun 16, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Unidata (2025). DeepFake Videos Dataset [Dataset]. https://www.kaggle.com/datasets/unidpro/deepfake-videos-dataset
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Jun 16, 2025
    Dataset provided by
    Kagglehttp://kaggle.com/
    Authors
    Unidata
    License

    Attribution-NonCommercial-NoDerivs 4.0 (CC BY-NC-ND 4.0)https://creativecommons.org/licenses/by-nc-nd/4.0/
    License information was derived automatically

    Description

    DeepFake Videos for detection tasks

    Dataset consists of 10,000+ files featuring 7,000+ people, providing a comprehensive resource for research in deepfake detection and deepfake technology. It includes real videos of individuals with AI-generated faces overlaid, specifically designed to enhance liveness detection systems.

    By utilizing this dataset, researchers can advance their understanding of deepfake generation and improve the performance of detection methods. - Get the data

    Metadata for the dataset

    https://www.googleapis.com/download/storage/v1/b/kaggle-user-content/o/inbox%2F22059654%2F7f47885f0afdca5c22f9f47e81307b95%2FFrame%201%20(8).png?generation=1742726304761567&alt=media" alt=""> Dataset was created by generating fake faces and overlaying them onto authentic video clips sourced from platforms such as aisaver.io, faceswapvideo.ai, and magichour.ai.Videos featuring different individuals, backgrounds, and scenarios, making it suitable for various research applications.

    💵 Buy the Dataset: This is a limited preview of the data. To access the full dataset, please contact us at https://unidata.pro to discuss your requirements and pricing options.

    Researchers can leverage this dataset to enhance their understanding of deepfake detection and contribute to the development of more robust detection methods that can effectively combat the challenges posed by deepfake technology.

    🌐 UniData provides high-quality datasets, content moderation, data collection and annotation for your AI/ML projects

  7. f

    Comparisons between different video camera outputs as well as the background...

    • plos.figshare.com
    • datasetcatalog.nlm.nih.gov
    xls
    Updated May 31, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Paul D. Meek; Guy-Anthony Ballard; Peter J. S. Fleming; Michael Schaefer; Warwick Williams; Greg Falzon (2023). Comparisons between different video camera outputs as well as the background (*: denotes statistical significance below the p = 0.05 level). [Dataset]. http://doi.org/10.1371/journal.pone.0110832.t004
    Explore at:
    xlsAvailable download formats
    Dataset updated
    May 31, 2023
    Dataset provided by
    PLOS ONE
    Authors
    Paul D. Meek; Guy-Anthony Ballard; Peter J. S. Fleming; Michael Schaefer; Warwick Williams; Greg Falzon
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Comparisons between different video camera outputs as well as the background (*: denotes statistical significance below the p = 0.05 level).

  8. Habits of video podcast consumption in the U.S. 2024

    • statista.com
    • thefarmdosupply.com
    Updated Jul 10, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Statista (2025). Habits of video podcast consumption in the U.S. 2024 [Dataset]. https://www.statista.com/statistics/1369071/video-podcast-consumption-habits-us/
    Explore at:
    Dataset updated
    Jul 10, 2025
    Dataset authored and provided by
    Statistahttp://statista.com/
    Time period covered
    2024
    Area covered
    United States
    Description

    According to a study on podcast listening in the United States in 2024, ** percent of weekly podcast listeners stated to have consumed video podcasts. While ** percent of the respondents indicated to watch video podcasts actively, ** percent also stated to let the video content play in the background while they were listening to the audio.

  9. HWID12 (Highway Incidents Detection Dataset)

    • kaggle.com
    Updated May 25, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Landry KEZEBOU (2022). HWID12 (Highway Incidents Detection Dataset) [Dataset]. https://www.kaggle.com/datasets/landrykezebou/hwid12-highway-incidents-detection-dataset
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    May 25, 2022
    Dataset provided by
    Kagglehttp://kaggle.com/
    Authors
    Landry KEZEBOU
    Description

    Context

    Action Recognition in video is known to be more challenging than image recognition problems. Unlike image recognition models which use 2D convolutional neural blocks, action classification models require additional dimensionality to capture the spatio-temporal information in video sequences. This intrinsically makes video action recognition models computationally intensive and significantly more data-hungry than image recognition counterparts. Unequivocally, existing video datasets such as Kinetics, AVA, Charades, Something-Something, HMDB51, and UFC101 have had tremendous impact on the recently evolving video recognition technologies. Artificial Intelligence models trained on these datasets have largely benefited applications such as behavior monitoring in elderly people, video summarization, and content-based retrieval. However, this growing concept of action recognition has yet to be explored in Intelligent Transportation System (ITS), particularly in vital applications such as incidents detection. This is partly due to the lack of availability of annotated dataset adequate for training models suitable for such direct ITS use cases. In this paper, the concept of video action recognition is explored to tackle the problem of highway incident detection and classification from live surveillance footage. First, a novel dataset - HWID12 (Highway Incidents Detection) dataset is introduced. The HWAD12 consists of 11 distinct highway incidents categories, and one additional category for negative samples representing normal traffic. The proposed dataset also includes 2780+ video segments of 3 to 8 seconds on average each, and 500k+ temporal frames. Next, the baseline for highway accident detection and classification is established with a state-of-the-art action recognition model trained on the proposed HWID12 dataset. Performance benchmarking for 12-class (normal traffic vs 11 accident categories), and 2-class (incident vs normal traffic) settings is performed. This benchmarking reveals a recognition accuracy of up to 88% and 98% for 12-class and 2-class recognition setting, respectively.

    Data Acquisition

    The Proposed Highway Incidents Detection Dataset (HWID12) is the first of its kind dataset aimed at fostering experimentation of video action recognition technologies to solve the practical problem of real-time highway incident detections which currently challenges intelligent transportation systems. The lack of such dataset has limited the expansion of the recent breakthroughs in video action classification for practical uses cases in intelligent transportation systems.. The proposed dataset contains more than 2780 video clips of length varying between 3 to 8 seconds. These video clips capture moments leading to, up until right after an incident occurred. The clips were manually segmented from accident compilations videos sourced from YouTube and other videos data platforms.

    Content

    There is one main zip file available for download. The zip file contains 2780+ video clips. 1) 12 folders
    2) each folder represents an incident category. One of the classes represent the negative sample class which simulates normal traffic.

    Terms and Conditions

    • Videos provided in this dataset are freely available for research and education purposes only. Please be sure to properly credit the authors by citing the article below.
    • Be sure to upvote this dataset if you find it useful by scrolling up and clicking the up-Arrow ^ sign at the top banner of the page, next to "New Notebook" button.
    • Be sure to blur out all plate numbers before publishing any of the contents available in this dataset.

    Acknowledgements

    Any publication using this database must reference to the following journal manuscript:

    • Landry Kezebou, Victor Oludare, Karen Panetta, James Intriligator, and Sos Agaian "Highway accident detection and classification from live traffic surveillance cameras: a comprehensive dataset and video action recognition benchmarking", Proc. SPIE 12100, Multimodal Image Exploitation and Learning 2022, 121000M (27 May 2022); https://doi.org/10.1117/12.2618943

    Note: if the link is broken, please use http instead of https.

    In Chrome, use the steps recommended in the following website to view the webpage if it appears to be broken https://www.technipages.com/chrome-enabledisable-not-secure-warning

    Other relevant datasets VCoR dataset: https://www.kaggle.com/landrykezebou/vcor-vehicle-color-recognition-dataset VRiV dataset: https://www.kaggle.com/landrykezebou/vriv-vehicle-recognition-in-videos-dataset

    For any enquires regarding the HWID12 dataset, contact: landrykezebou@gmail.com

  10. N

    Replication Data for: Multi-Class Video Co-Segmentation with a Generative...

    • dataverse.lib.nycu.edu.tw
    Updated Jun 14, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    NYCU Dataverse (2022). Replication Data for: Multi-Class Video Co-Segmentation with a Generative Multi-Video Model [Dataset]. http://doi.org/10.57770/AG50YQ
    Explore at:
    png(131370), png(107290), png(171566), png(135422), png(188519), png(130541), png(108340), png(3451), png(347788), png(171671), application/matlab-mat(3457), png(346130), png(268371), png(134621), png(392249), png(125680), png(2006), png(94691), png(138779), png(177536), png(2672), png(417861), png(111726), png(294884), png(138135), application/matlab-mat(1131), png(139891), png(2227), png(206530), png(124006), png(1834), application/matlab-mat(1107), application/matlab-mat(3527), png(213157), png(1339), png(110891), png(109130), png(326585), png(313010), png(141322), png(282849), png(140015), png(177786), png(144791), png(127279), png(206011), text/x-matlab(1044), application/matlab-mat(3813), png(318812), png(323740), png(108862), png(1182), png(206871), png(153370), png(157236), text/x-matlab(17992), png(199087), application/matlab-mat(418942), png(193325), png(1026), text/x-matlab(334), png(415364), application/matlab-mat(2740), png(158902), png(419569), text/x-matlab(959), png(137203), text/x-matlab(519), png(162560), png(196099), png(190176), png(133263), png(129252), png(127604), application/matlab-mat(1590), png(194179), png(213655), application/matlab-mat(2492), png(299446), png(415359), text/x-matlab(603), png(319493), application/matlab-mat(2772), png(1389), png(141796), application/matlab-mat(1612), png(134929), application/matlab-mat(430524), application/matlab-mat(1705), png(393682), png(350362), text/x-matlab(798), text/x-matlab(5092), png(171493), application/matlab-mat(2143), png(379952), png(372880), png(137367), png(192282), png(354031), png(107872), png(305159), png(340675), application/matlab-mat(964), png(120625), application/matlab-mat(424901), png(2607), png(403799), png(339179), png(137843), png(134677), png(311898), png(344293), application/matlab-mat(1750), png(126883), png(208322), png(237256), png(344510), png(132888), application/matlab-mat(431742), png(109676), application/matlab-mat(2823), png(211799), text/x-matlab(369), png(170812), png(2909), png(394819), png(286566), application/matlab-mat(1070), png(367772), png(220763), png(105347), png(318010), png(164240), png(88287), application/matlab-mat(3178), application/matlab-mat(1034), png(141317), png(421902), png(194634), png(208072), png(190019), png(141670), png(111433), text/x-matlab(605), png(139637), application/matlab-mat(2162), png(106779), png(136606), png(293088), application/matlab-mat(408579), png(409113), png(132822), png(167235), png(201920), png(416335), png(208829), png(366889), png(325462), png(105548), png(196677), png(418723), text/x-matlab(8820), application/matlab-mat(1094), png(197250), png(90810), png(144250), png(4054), application/matlab-mat(425225), png(212481), png(102643), png(319859), application/matlab-mat(977), png(155541), text/x-matlab(231), png(393618), png(322623), text/x-matlab(481), png(340400), png(102231), png(388190), png(321264), png(368967), application/matlab-mat(2279), png(1231), png(415105), png(404892), png(306968), png(376883), application/matlab-mat(1844), png(366211), png(143067), png(319741), png(153442), png(134280), png(113170), png(324760), application/matlab-mat(1715), png(140612), text/x-matlab(528), png(110588), png(207723), text/x-matlab(1615), png(317481), png(139990), png(1512), png(340438), png(1128), png(107547), png(213448), text/x-matlab(434), png(420414), png(202196), png(400831), text/x-matlab(622), png(355889), png(109724), application/matlab-mat(4031), png(153511), png(1266), application/matlab-mat(850), png(164947), png(148847), text/x-matlab(891), png(397267), png(132715), png(405141), text/x-matlab(552), png(146114), png(295751), png(147730), png(408145), png(2033), png(326837), png(3650), text/x-matlab(3006), application/matlab-mat(2738), png(323233), png(214140), png(106351), png(394534), png(223681), png(408719), png(188233), application/matlab-mat(3167), application/matlab-mat(924), png(129865), png(357971), png(398099), png(130881), png(1328), application/matlab-mat(421051), png(162353), png(154494), text/x-matlab(2353), png(141395), png(1118), png(118869), application/matlab-mat(1431), png(105170), application/matlab-mat(2078), png(139460), application/matlab-mat(2331), png(117533), text/plain; charset=us-ascii(1264), png(1318), png(365505), png(222727), png(332125), png(4030), png(209954), png(144307), png(193850), png(148336), png(163083), text/x-matlab(382), application/matlab-mat(2240), png(146720), application/matlab-mat(422019), png(341026), png(194120), png(192279), text/x-matlab(4355), png(130862), png(208591), application/matlab-mat(423255), application/matlab-mat(2215), png(391578), png(168506), application/matlab-mat(428870), png(94312), png(356718), application/matlab-mat(2922), application/matlab-mat(1827), png(390593), application/matlab-mat(415617), png(110755), application/matlab-mat(422027), png(160626), png(175648), png(213693), png(2161), png(2702), png(140557), application/matlab-mat(2571), png(136349), application/matlab-mat(3091), png(298249), png(405984), png(184679), application/matlab-mat(1466), png(408270), png(159641), png(405677), application/matlab-mat(3276), png(395406), png(409342), application/matlab-mat(1113), png(285677), png(1254), png(147166), application/matlab-mat(3003), text/x-matlab(1584), png(158907), application/matlab-mat(1294), png(204314), png(421597), png(148848), png(213293), png(339889), png(210667), png(2800), png(92334), png(142724), text/x-matlab(2769), png(234041), png(325939), png(2025), png(121973), png(408404), png(130088), application/matlab-mat(3664), png(127008), application/matlab-mat(424962), png(215780), png(1369), png(200318), png(208684), png(112228), png(220026), png(374091), png(187308), png(418303), text/x-matlab(322), png(272181), png(1131), png(140087), png(211075), png(221660), png(2764), png(313713), png(127093), png(363709), png(146675), png(108762), png(156860), png(390592), png(166566), png(111698), png(400813), png(417168), png(344011), png(191034), png(394358), application/matlab-mat(3855), png(339084), application/matlab-mat(2088), png(130979), png(163337), png(210018), png(164602), png(189181), png(321306), application/matlab-mat(420460), png(194624), png(148231), png(139157), png(414185), png(167904), png(309077), application/matlab-mat(2610), text/x-matlab(1013), png(137661), png(155939), png(170411), png(130405), png(927), png(175585), png(317209), application/matlab-mat(930), png(337561), png(1914), png(108238), text/x-matlab(545), application/matlab-mat(422822), png(194037), png(2171), application/matlab-mat(422316), png(416503), png(204440), png(324096), png(166597), text/x-matlab(468), png(110308), png(135824), application/matlab-mat(933), application/matlab-mat(2781), application/matlab-mat(1075), text/x-matlab(4550), application/matlab-mat(2570), application/matlab-mat(1423), png(406925), png(139253), png(139576), png(349628), application/matlab-mat(1128), application/matlab-mat(3353), png(406437), application/matlab-mat(2949), png(124934), png(298781), png(155911), png(298017), application/matlab-mat(3132), application/matlab-mat(2439), png(3015), application/matlab-mat(2952), png(198080), png(275247), png(171609), png(2956), png(399999), png(158447), png(198671), png(161632), png(127089), png(193212), png(164436), png(175289), png(142861), application/matlab-mat(3580), png(102476), png(111228), png(106771), png(364104), application/matlab-mat(419649), png(326249), application/matlab-mat(3267), png(1105), png(260313), png(194535), png(101678), png(419964), png(179465), png(323389), application/matlab-mat(429008), png(391484), png(208705), png(1276), png(2356), png(102206), application/matlab-mat(2605), application/matlab-mat(2370), png(291289), png(275482), png(417550), png(111493), png(306338), png(307676), text/plain; charset=us-ascii(2190), png(105312), png(196659), png(413788), png(1285), text/x-matlab(548), png(330578), png(406386), png(314774), png(154376), png(2728), png(161305), png(215834), png(99041), png(179248), png(143371), png(181693), png(164713), png(313636), png(108740), png(409075), png(212644), png(315429), png(390050), png(179032), application/matlab-mat(1252), text/x-matlab(717), png(416923), png(312031), png(141717), png(168196), png(1902), application/matlab-mat(4212), png(348628), application/matlab-mat(2144), png(313744), application/matlab-mat(1129), png(417926), png(94782), text/x-matlab(288), application/matlab-mat(1899), png(327779), text/x-matlab(5753), png(119903), application/matlab-mat(1216), png(297718), png(349673), png(165133), application/matlab-mat(936), application/matlab-mat(3720), png(321843), png(186541), png(2923), application/matlab-mat(2121), png(163230), application/matlab-mat(3563), text/x-matlab(286), application/matlab-mat(3764), png(190992), png(212446), png(108576), png(394890), png(394333), png(111055), application/matlab-mat(1816), png(99037), application/matlab-mat(2760), png(147020), text/x-matlab(224), application/matlab-mat(2524), png(280094), png(153546), application/matlab-mat(2789), png(137859), png(141598), png(192358), png(1408), png(141567), application/matlab-mat(6917), png(137063), png(149103), png(125715), png(111546), png(166993), png(394553), png(185802), png(149394), text/x-matlab(482), png(193715), application/matlab-mat(4007), application/matlab-mat(1016), application/matlab-mat(3459), png(158423), png(306787), png(110580), png(2458), png(341106), png(324454), png(104628), png(333983), text/x-matlab(324), text/x-matlab(4454), png(226018), png(414532), png(214071), png(417369), application/matlab-mat(2399), png(138043), png(363432), png(124157), png(189964), application/matlab-mat(431490), application/matlab-mat(2684), png(136768), png(315883), png(247996), png(291958), png(315058), png(110571), png(138800), png(142408), png(356579), text/plain; charset=us-ascii(1743), application/matlab-mat(2126), png(111806), png(398365), png(141658), png(96259), png(208666), application/matlab-mat(424450), png(3243), application/matlab-mat(3656), png(381663), png(208003), png(167107), png(302794), png(320708), application/matlab-mat(1923), png(414831), png(216123), png(303672), png(188453), png(418987), png(133661), text/x-matlab(3241), png(315079), png(125598), png(108949), png(2589), application/matlab-mat(3008), png(169080), png(330168), png(113415), png(2694), png(97614), png(138483), png(110612), png(133299), png(109822), png(167170), png(329461), png(138129), png(2345), mp4(798871), text/x-matlab(1885), png(117812), png(191451), png(121245), text/x-matlab(1025), png(349296), png(386901), png(388647), png(132769), application/matlab-mat(3547), png(190704), application/matlab-mat(2392), application/matlab-mat(412872), application/matlab-mat(3250), png(139180), application/matlab-mat(2970), png(322537), png(108279), png(115116), png(211305), application/matlab-mat(429301), png(323561), png(341326), png(202155), png(397653), application/matlab-mat(3589), png(318566), png(138190), png(190173), png(353017), application/matlab-mat(3304), png(343714), png(1044), png(320057), application/matlab-mat(2910), text/x-matlab(3020), png(2655), png(97876), png(193108), png(108552), png(407800), png(163357), txt(3585), png(107157), png(2512), png(217060), application/matlab-mat(2441), png(323547), png(3234), png(282539), application/matlab-mat(1739), application/matlab-mat(2190), png(341186), text/x-matlab(202), png(211545), png(407597), png(342667), text/x-matlab(6244), png(345763), png(1732), png(416824), png(117050), application/matlab-mat(3428), png(319258), application/matlab-mat(3016), png(344760), application/matlab-mat(417609), text/x-matlab(254), png(138657), png(415908), application/matlab-mat(2582), png(142018), png(199110), text/x-matlab(196), application/matlab-mat(3596), png(158293), png(106234), application/matlab-mat(3346), png(146000), png(323782), text/plain; charset=us-ascii(651), png(85480), png(141804), text/x-matlab(669), png(337530), text/x-matlab(479), application/matlab-mat(1081), png(303189), png(328947), application/matlab-mat(1525), png(103704), png(214052), png(324542), png(391921), application/matlab-mat(420489), png(2976), png(214867), png(417105), png(343204), application/matlab-mat(3471), png(400370), png(151107), mp4(3155138), png(130131), application/matlab-mat(1651), png(194262), png(138854), png(219016), application/matlab-mat(2992), png(363335), png(144552), png(413807), text/x-matlab(598), png(209056), png(156331), png(212117), application/matlab-mat(3307), png(169264), png(170745), text/x-matlab(487), png(414723), png(369367), png(393532), png(143239), png(416674), png(416606)Available download formats
    Dataset updated
    Jun 14, 2022
    Dataset provided by
    NYCU Dataverse
    License

    CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
    License information was derived automatically

    Description

    Video data provides a rich source of information that is available to us today in large quantities e.g. from online resources. Tasks like segmentation benefit greatly from the analysis of spatio-temporal motion patterns in videos and recent advances in video segmentation has shown great progress in exploiting these addition cues. However, observing a single video is often not enough to predict meaningful segmentations and inference across videos becomes necessary in order to predict segmentations that are consistent with objects classes. Therefore the task of video co-segmentation is being proposed, that aims at inferring segmentation from multiple videos. But current approaches are limited to only considering binary foreground/background segmentation and multiple videos of the same object. This is a clear mismatch to the challenges that we are facing with videos from online resources or consumer videos. We propose to study multi-class video co-segmentation where the number of object classes is unknown as well as the number of instances in each frame and video. We achieve this by formulating a non-parametric bayesian model across videos sequences that is based on a new videos segmentation prior as well as a global appearance model that links segments of the same class. We present the first multi-class video co-segmentation evaluation. We show that our method is applicable to real video data from online resources and outperforms state-of-the-art video segmentation and image co-segmentation baselines.

  11. WormSwin: C. elegans Video Datasets

    • zenodo.org
    • data.niaid.nih.gov
    zip
    Updated Jan 31, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Maurice Deserno; Maurice Deserno; Katarzyna Bozek; Katarzyna Bozek (2024). WormSwin: C. elegans Video Datasets [Dataset]. http://doi.org/10.5281/zenodo.7456803
    Explore at:
    zipAvailable download formats
    Dataset updated
    Jan 31, 2024
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Maurice Deserno; Maurice Deserno; Katarzyna Bozek; Katarzyna Bozek
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Data used for our paper "WormSwin: Instance Segmentation of C. elegans using Vision Transformer".
    This publication is divided into three parts:

    1. CSB-1 Dataset
    2. Synthetic Images Dataset
    3. MD Dataset

    The CSB-1 Dataset consists of frames extracted from videos of Caenorhabditis elegans (C. elegans) annotated with binary masks. Each C. elegans is separately annotated, providing accurate annotations even for overlapping instances. All annotations are provided in binary mask format and as COCO Annotation JSON files (see COCO website).

    The videos are named after the following pattern:

    <"worm age in hours"_"mutation"_"irradiated (binary)"_"video index (zero based)">

    For mutation the following values are possible:

    1. wild type
    2. csb-1 mutant
    3. csb-1 with rescue mutation

    An example video name would be 24_1_1_2 meaning it shows C. elegans with csb-1 mutation, being 24h old which got irradiated.

    Video data was provided by M. Rieckher; Instance Segmentation Annotations were created under supervision of K. Bozek and M. Deserno.

    The Synthetic Images Dataset was created by cutting out C. elegans (foreground objects) from the CSB-1 Dataset and placing them randomly on background images also taken from the CSB-1 Dataset. Foreground objects were flipped, rotated and slightly blurred before placed on the background images.
    The same was done with the binary mask annotations taken from CSB-1 Dataset so that they match the foreground objects in the synthetic images. Additionally, we added rings of random color, size, thickness and position to the background images to simulate petri-dish edges.

    This synthetic dataset was generated by M. Deserno.

    The Mating Dataset (MD) consists of 450 grayscale image patches of 1,012 x 1,012 px showing C. elegans with high overlap, crawling on a petri-dish.
    We took the patches from a 10 min. long video of size 3,036 x 3,036 px. The video was downsampled from 25 fps to 5 fps before selecting 50 random frames for annotating and patching.
    Like the other datasets, worms were annotated with binary masks and annotations are provided as COCO Annotation JSON files.

    The video data was provided by X.-L. Chu; Instance Segmentation Annotations were created under supervision of K. Bozek and M. Deserno.


    Further details about the datasets can be found in our paper.

  12. m

    SDFVD: Small-scale Deepfake Forgery Video Dataset

    • data.mendeley.com
    Updated Apr 23, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Shilpa Kaman (2024). SDFVD: Small-scale Deepfake Forgery Video Dataset [Dataset]. http://doi.org/10.17632/bcmkfgct2s.1
    Explore at:
    Dataset updated
    Apr 23, 2024
    Authors
    Shilpa Kaman
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Small-scale Deepfake Forgery Video Dataset (SDFVD) is a custom dataset consisting of real and deepfake videos with diverse contexts designed to study and benchmark deepfake detection algorithms. The dataset comprising of a total of 106 videos, with 53 original and 53 deepfake videos. Equal number of real and deepfake videos, ensures balance for machine learning model training and evaluation. The original videos were collected from Pexels: a well- known provider of stock photography and stock footage(video). These videos include a variety of backgrounds, and the subjects represent different genders and ages, reflecting a diverse range of scenarios. The input videos have been pre-processed by cropping them to a length of approximately 4 to 5 seconds and resizing them to 720p resolution, ensuring a consistent and uniform format across the dataset. Deepfake videos were generated using Remaker AI employing face-swapping techniques. Remaker AI is an AI-powered platform that can generate images, swap faces in photos and videos, and edit content. The source face photos for these swaps were taken from Freepik: is an image bank website provides contents such as photographs, illustrations and vector images. SDFVD was created due to the lack of availability of any such comparable small-scale deepfake video datasets. Key benefits of such datasets are: • In educational settings or smaller research labs, smaller datasets can be particularly useful as they require fewer resources, allowing students and researchers to conduct experiments with limited budgets and computational resources. • Researchers can use small-scale datasets to quickly prototype new ideas, test concepts, and refine algorithms before scaling up to larger datasets. Overall, SDFVD offers a compact but diverse collection of real and deepfake videos, suitable for a variety of applications, including research, security, and education. It serves as a valuable resource for exploring the rapidly evolving field of deepfake technology and its impact on society.

  13. d

    2004-2018 Video Services Franchise Revenue History

    • catalog.data.gov
    • data.bloomington.in.gov
    Updated Sep 29, 2023
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    data.bloomington.in.gov (2023). 2004-2018 Video Services Franchise Revenue History [Dataset]. https://catalog.data.gov/dataset/2004-2018-video-services-franchise-revenue-history
    Explore at:
    Dataset updated
    Sep 29, 2023
    Dataset provided by
    data.bloomington.in.gov
    Description

    Video service providers (cable) are required to compensate municipalities for the use of public rights-of-way. This compensation is used by the City of Bloomington for a number of communications and information technology projects. This data reflects the payments of wireline video service providers in the City of Bloomington. Attached is an Excel report using this dataset.

  14. f

    Dataset-I-drinking-related-object-detection (in both YoloV8 and COCO format)...

    • kcl.figshare.com
    Updated Feb 27, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Xin Chen; Xinqi Bao; Ernest Kamavuako (2025). Dataset-I-drinking-related-object-detection (in both YoloV8 and COCO format) [Dataset]. http://doi.org/10.18742/26337085.v1
    Explore at:
    Dataset updated
    Feb 27, 2025
    Dataset provided by
    King's College London
    Authors
    Xin Chen; Xinqi Bao; Ernest Kamavuako
    License

    https://www.kcl.ac.uk/researchsupport/assets/DataAccessAgreement-Description.pdfhttps://www.kcl.ac.uk/researchsupport/assets/DataAccessAgreement-Description.pdf

    Description

    This dataset contains annotated images for object detection for containers and hands in a first-person view (egocentric view) during drinking activities. Both YOLOV8 format and COCO format are provided.Please refer to our paper for more details.Purpose: Training and testing the object detection model.Content: Videos from Session 1 of Subjects 1-20.Images: Extracted from the videos of Subjects 1-20 Session 1.Additional Images:~500 hand/container images from Roboflow Open Source data.~1500 null (background) images from VOC Dataset and MIT Indoor Scene Recognition Dataset:1000 indoor scenes from 'MIT Indoor Scene Recognition'400 other unrelated objects from VOC DatasetData Augmentation:Horizontal flipping±15% brightness change±10° rotationFormats Provided:COCO formatPyTorch YOLOV8 formatImage Size: 416x416 pixelsTotal Images: 16,834Training: 13,862Validation: 1,975Testing: 997Instance Numbers:Containers: Over 10,000Hands: Over 8,000

  15. I

    Global Video Background Remover Market Strategic Planning Insights 2025-2032...

    • statsndata.org
    excel, pdf
    Updated Sep 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Stats N Data (2025). Global Video Background Remover Market Strategic Planning Insights 2025-2032 [Dataset]. https://www.statsndata.org/report/video-background-remover-market-103310
    Explore at:
    pdf, excelAvailable download formats
    Dataset updated
    Sep 2025
    Dataset authored and provided by
    Stats N Data
    License

    https://www.statsndata.org/how-to-orderhttps://www.statsndata.org/how-to-order

    Area covered
    Global
    Description

    The Video Background Remover market is rapidly evolving, driven by the growing demand for high-quality content in digital media production, social media marketing, and virtual communications. This technology enables users to seamlessly remove or alter backgrounds in videos without extensive editing skills, providing

  16. Global Stock Video Market Size By User Type (Individual Creators, Small And...

    • verifiedmarketresearch.com
    pdf,excel,csv,ppt
    Updated May 2, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Verified Market Research (2025). Global Stock Video Market Size By User Type (Individual Creators, Small And Medium Enterprises, Large Enterprises, Educational Institutions), By Platform Type (Stock Video Agencies, User-Generated Content Platforms), By Vertical Industries (Media and Entertainment, Corporate and Business, Education, Healthcare and Medical) And Region For 2026-2032 [Dataset]. https://www.verifiedmarketresearch.com/product/stock-video-market/
    Explore at:
    pdf,excel,csv,pptAvailable download formats
    Dataset updated
    May 2, 2025
    Dataset authored and provided by
    Verified Market Researchhttps://www.verifiedmarketresearch.com/
    License

    https://www.verifiedmarketresearch.com/privacy-policy/https://www.verifiedmarketresearch.com/privacy-policy/

    Time period covered
    2026 - 2032
    Area covered
    Global
    Description

    Stock Video Market size was valued at USD 5.99 Billion in 2024 and is projected to reach USD 9.98 Billion by 2032, growing at a CAGR of 8.75% during the forecast period 2026-2032.

    Stock Video Market: Definition/ Overview

    Stock video is pre-recorded material available for license to filmmakers, video producers and content developers. These movies include a wide range of subjects and scenarios, from natural scenes to urban landscapes and are utilized to supplement video projects without requiring original filming. Stock videos save time and resources by providing high-quality visuals quickly.

    Stock video assets are adaptable and can be utilized in a variety of media projects. They improve marketing campaigns, social media postings and advertising by providing professional quality without the cost of specialized shoots. Filmmakers and video developers use them for B-roll, background scenes and visual storytelling. They can also be used in educational videos, presentations and website designs to interest and inform viewers.

    Stock video offers the potential to transform content development by allowing for quick, cost-effective production in marketing, education and entertainment. It benefits a wide range of industries, including advertising and movies by strengthening storytelling with high-quality images. As AI progresses, personalized and dynamic stock footage will enhance user experiences making it a useful tool for both creators and corporations.

  17. ComplexVAD Video Anomaly Detection Dataset

    • zenodo.org
    • data.niaid.nih.gov
    zip
    Updated Jun 12, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Furkan Mumcu; Furkan Mumcu; Mike Jones; Mike Jones; Anoop Cherian; Anoop Cherian; Yasin Yilmaz; Yasin Yilmaz (2024). ComplexVAD Video Anomaly Detection Dataset [Dataset]. http://doi.org/10.5281/zenodo.11475281
    Explore at:
    zipAvailable download formats
    Dataset updated
    Jun 12, 2024
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Furkan Mumcu; Furkan Mumcu; Mike Jones; Mike Jones; Anoop Cherian; Anoop Cherian; Yasin Yilmaz; Yasin Yilmaz
    License

    Attribution-ShareAlike 4.0 (CC BY-SA 4.0)https://creativecommons.org/licenses/by-sa/4.0/
    License information was derived automatically

    Description

    Introduction

    The ComplexVAD dataset consists of 104 training and 113 testing video sequences taken from a static camera looking at a scene of a two-lane street with sidewalks on either side of the street and another sidewalk going across the street at a crosswalk. The videos were collected over a period of a few months on the campus of the University of South Florida using a camcorder with 1920 x 1080 pixel resolution. Videos were collected at various times during the day and on each day of the week. Videos vary in duration with most being about 12 minutes long. The total duration of all training and testing videos is a little over 34 hours. The scene includes cars, buses and golf carts driving in two directions on the street, pedestrians walking and jogging on the sidewalks and crossing the street, people on scooters, skateboards and bicycles on the street and sidewalks, and cars moving in the parking lot in the background. Branches of a tree also move at the top of many frames.

    The 113 testing videos have a total of 118 anomalous events consisting of 40 different anomaly types.

    Ground truth annotations are provided for each testing video in the form of bounding boxes around each anomalous event in each frame. Each bounding box is also labeled with a track number, meaning each anomalous event is labeled as a track of bounding boxes. A single frame can have more than one anomaly labeled.

    At a Glance

    • The size of the unzipped dataset is ~39GB
    • The dataset consists of Train sequences (containing only videos with normal activity), Test sequences (containing some anomalous activity), a ground truth annotation file for each Test sequence, and a README.md file describing the data organization and ground truth annotation format.
    • The zip files contain a Train directory, a Test directory, an annotations directory, and a README.md file.

    License

    The ComplexVAD dataset is released under CC-BY-SA-4.0 license.

    All data:

    Created by Mitsubishi Electric Research Laboratories (MERL), 2024
    
    SPDX-License-Identifier: CC-BY-SA-4.0
  18. i

    Grant Giving Statistics for The Video Game History Foundation Inc

    • instrumentl.com
    Updated Mar 31, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    (2024). Grant Giving Statistics for The Video Game History Foundation Inc [Dataset]. https://www.instrumentl.com/990-report/video-game-history-foundation-inc
    Explore at:
    Dataset updated
    Mar 31, 2024
    Variables measured
    Total Assets, Total Giving
    Description

    Financial overview and grant giving statistics of The Video Game History Foundation Inc

  19. Most viewed YouTube videos of all time 2025

    • statista.com
    • tokrwards.com
    Updated Feb 17, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Statista (2025). Most viewed YouTube videos of all time 2025 [Dataset]. https://www.statista.com/statistics/249396/top-youtube-videos-views/
    Explore at:
    Dataset updated
    Feb 17, 2025
    Dataset authored and provided by
    Statistahttp://statista.com/
    Time period covered
    Feb 2025
    Area covered
    Worldwide, YouTube
    Description

    On June 17, 2016, Korean education brand Pinkfong released their video "Baby Shark Dance", and the rest is history. In January 2021, Baby Shark Dance became the first YouTube video to surpass 10 billion views, after snatching the crown of most-viewed YouTube video of all time from the former record holder "Despacito" one year before. "Baby Shark Dance" currently has over 15 billion lifetime views on YouTube. Music videos on YouTube “Baby Shark Dance” might be the current record-holder in terms of total views, but Korean artist Psy’s “Gangnam Style” video remained on the top spot for longest (1,689 days or 4.6 years) before ceding its spot to its successor. With figures like these, it comes as little surprise that the majority of the most popular videos on YouTube are music videos. Since 2010, all but one the most-viewed videos on YouTube have been music videos, signifying the platform’s shift in focus from funny, viral videos to professionally produced content. As of 2022, about 40 percent of the U.S. digital music audience uses YouTube Music. Popular video content on YouTube Music fans are also highly engaged audiences and it is not uncommon for music videos to garner significant amounts of traffic within the first 24 hours of release. Other popular types of videos that generate lots of views after their first release are movie trailers, especially superhero movies related to the MCU (Marvel Cinematic Universe). The first official trailer for the upcoming film “Avengers: Endgame” generated 289 million views within the first 24 hours of release, while the movie trailer for Spider-Man: No Way Home generated over 355 views on the first day from release, making it the most viral movie trailer.

  20. Scene Classification: Images and Audio

    • kaggle.com
    Updated Feb 1, 2020
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Jordan J. Bird (2020). Scene Classification: Images and Audio [Dataset]. https://www.kaggle.com/datasets/birdy654/scene-classification-images-and-audio
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Feb 1, 2020
    Dataset provided by
    Kagglehttp://kaggle.com/
    Authors
    Jordan J. Bird
    Description

    Do images and audio complement one another in scene classification?

    These dataset is made up of images from 8 different environments. 37 video sources have been processed, every 1 second an image is extracted (frame at 0.5s, 1.5s, 2.5s ... and so on) and to accompany that image, the MFCC audio statistics are also extracted from the relevant second of video.

    In this dataset, you will notice some common errors from single classifiers. For example, in the video of London, the image classifier confuses the environment with "FOREST" when a lady walks past with flowing hair. Likewise, the audio classifier gets confused by "RIVER" when we walk past a large fountain in Las Vegas due to the sounds of flowing water. Both of these errors can be fixed by a multi-modal approach, where fusion allows for the correction of errors. In our study, both of these issues were classified as "CITY" since multimodality can provide a solution for single-modal errors due to anomalous data occurring.

    Please cite this study if you use the dataset

    Look and Listen: A Multi-Modal Late Fusion Approach to Scene Classification for Autonomous Machines Jordan J. Bird, Diego R. Faria, Cristiano Premebida, Aniko Ekart, and George Vogiatzis

    Context

    In this challenge, we can learn environments ("Where am I?") from either images, audio, or take a multimodal approach to fuse the data.

    Multi-modal fusion often requires far fewer computing resources than temporal models, but sometimes at the cost of classification ability. Can a method of fusion overcome this? Let's find out!

    Content

    Class data are given as strings in dataset.csv

    Each row of the dataset contains a path to the image, as well as the MFCC data extracted from the second of video that accompany the frame.

    MFCC Extraction

    (copied and pasted from the paper) we extract the the Mel-Frequency Cepstral Coefficients (MFCC) of the audio clips through a set of sliding windows 0.25s in length (ie frame size of 4K sampling points) and an additional set of overlapping windows, thus producing 8 sliding windows, 8 frames/sec. From each audio-frame, we extract 13 MFCC attributes, producing 104 attributes per 1 second clip.

    These are numbered in sequence from MFCC_1

    Two Classes?

    The original study deals with Class 2 (the actual environment, 8 classes) but we have included Class 1 also. Class 1 is a much easier binary classification problem of "Outdoors" and "Indoors"

Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
Mario Peedor (2025). THVD (Talking Head Video Dataset) [Dataset]. http://doi.org/10.17632/ykhw8r7bfx.1

THVD (Talking Head Video Dataset)

Explore at:
Dataset updated
Apr 2, 2025
Authors
Mario Peedor
License

Attribution-NonCommercial 3.0 (CC BY-NC 3.0)https://creativecommons.org/licenses/by-nc/3.0/
License information was derived automatically

Description

About

We provide a comprehensive talking-head video dataset with over 50,000 videos, totaling more than 500 hours of footage and featuring 23,841 unique identities from around the world.

Distribution

Detailing the format, size, and structure of the dataset: Data Volume: -Total Size: 2.5TB

-Total Videos: 47,200

-Identities Covered: 23,000

-Resolution: 60% 4k(1980), 33% fullHD(1080)

-Formats: MP4

-Full-length videos with visible mouth movements in every frame.

-Minimum face size of 400 pixels.

-Video durations range from 20 seconds to 5 minutes.

-Faces have not been cut out, full screen videos including backgrounds.

Usage

This dataset is ideal for a variety of applications:

Face Recognition & Verification: Training and benchmarking facial recognition models.

Action Recognition: Identifying human activities and behaviors.

Re-Identification (Re-ID): Tracking identities across different videos and environments.

Deepfake Detection: Developing methods to detect manipulated videos.

Generative AI: Training high-resolution video generation models.

Lip Syncing Applications: Enhancing AI-driven lip-syncing models for dubbing and virtual avatars.

Background AI Applications: Developing AI models for automated background replacement, segmentation, and enhancement.

Coverage

Explaining the scope and coverage of the dataset:

Geographic Coverage: Worldwide

Time Range: Time range and size of the videos have been noted in the CSV file.

Demographics: Includes information about age, gender, ethnicity, format, resolution, and file size.

Languages Covered (Videos):

English: 23,038 videos

Portuguese: 1,346 videos

Spanish: 677 videos

Norwegian: 1,266 videos

Swedish: 1,056 videos

Korean: 848 videos

Polish: 1,807 videos

Indonesian: 1,163 videos

French: 1,102 videos

German: 1,276 videos

Japanese: 1,433 videos

Dutch: 1,666 videos

Indian: 1,163 videos

Czech: 590 videos

Chinese: 685 videos

Italian: 975 videos

Who Can Use It

List examples of intended users and their use cases:

Data Scientists: Training machine learning models for video-based AI applications.

Researchers: Studying human behavior, facial analysis, or video AI advancements.

Businesses: Developing facial recognition systems, video analytics, or AI-driven media applications.

Additional Notes

Ensure ethical usage and compliance with privacy regulations. The dataset’s quality and scale make it valuable for high-performance AI training. Potential preprocessing (cropping, down sampling) may be needed for different use cases. Dataset has not been completed yet and expands daily, please contact for most up to date CSV file. The dataset has been divided into 100GB zipped files and is hosted on a private server (with the option to upload to the cloud if needed). To verify the dataset's quality, please contact me for the full CSV file. I’d be happy to provide example videos selected by the potential buyer.

Search
Clear search
Close search
Google apps
Main menu