19 datasets found
  1. R

    Yolo Find Text Dataset

    • universe.roboflow.com
    zip
    Updated Jul 17, 2022
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    YoloV5 Train (2022). Yolo Find Text Dataset [Dataset]. https://universe.roboflow.com/yolov5-train/yolo-find-text
    Explore at:
    zipAvailable download formats
    Dataset updated
    Jul 17, 2022
    Dataset authored and provided by
    YoloV5 Train
    License

    MIT Licensehttps://opensource.org/licenses/MIT
    License information was derived automatically

    Variables measured
    Text Bounding Boxes
    Description

    Yolo Find Text

    ## Overview
    
    Yolo Find Text is a dataset for object detection tasks - it contains Text annotations for 290 images.
    
    ## Getting Started
    
    You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
    
      ## License
    
      This dataset is available under the [MIT license](https://creativecommons.org/licenses/MIT).
    
  2. R

    Testdigit Dataset

    • universe.roboflow.com
    zip
    Updated Jul 18, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Christoph Ponath (2023). Testdigit Dataset [Dataset]. https://universe.roboflow.com/christoph-ponath/testdigit/model/2
    Explore at:
    zipAvailable download formats
    Dataset updated
    Jul 18, 2023
    Dataset authored and provided by
    Christoph Ponath
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Variables measured
    Testdigit Bounding Boxes
    Description

    Here are a few use cases for this project:

    1. Educational Tools: This computer vision model can be integrated into educational software, assisting students in learning numerical and symbolic comprehension. For instance, recognizing and categorizing handwritten digits and symbols to facilitate learning mathematics.

    2. Optical Character Recognition (OCR): This model can be used to recognize and categorize digits and symbols in scanned documents or photos, aiding in digitization and data extraction purposes.

    3. Handwriting Recognition Systems: It can be applied in handwriting recognition systems to identify and categorize unique handwritten digits or characters, supporting automated evaluation or data entry.

    4. Accessibility Applications: It can support the creation of tools for visually impaired individuals, by recognizing text and symbols in physical documents and producing spoken output.

    5. Automated Testing Applications: In a testing or examination scenario, the model can automatically grade multiple-choice tests or quizzes by recognizing and categorizing filled answer bubbles or handwritten digits/symbols.

  3. Saudi Arabian Number Plates Annotations

    • kaggle.com
    Updated Dec 27, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Patrick Muthii (2024). Saudi Arabian Number Plates Annotations [Dataset]. https://www.kaggle.com/datasets/patrickmuthii/saudi-arabian-number-plates-annotations
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Dec 27, 2024
    Dataset provided by
    Kaggle
    Authors
    Patrick Muthii
    License

    MIT Licensehttps://opensource.org/licenses/MIT
    License information was derived automatically

    Area covered
    Saudi Arabia
    Description

    This dataset comprises a total of 190 images used to create a license plate recognition model for Saudi Arabian number plates. First, we create bounding box annotations for Saudi Arabian license plate images to train an object detection model using YOLO Bounding Box Annotation Tool (YBAT). The annotations are saved as .xml files. https://youtu.be/k-d1OFHeikg Then, we implement a Faster R-CNN model with a ResNet-50 backbone using PyTorch. The model is trained to detect and localize various components of the license plate, including Arabic and Latin characters, numbers, and the KSA logo.

  4. R

    Game Dataset

    • universe.roboflow.com
    zip
    Updated May 20, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    lxue Xue (2023). Game Dataset [Dataset]. https://universe.roboflow.com/lxue-xue-cxrjf/game-kfdts/model/1
    Explore at:
    zipAvailable download formats
    Dataset updated
    May 20, 2023
    Dataset authored and provided by
    lxue Xue
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Variables measured
    Game Bounding Boxes
    Description

    Here are a few use cases for this project:

    1. Game Character Customization: Utilize the "game" computer vision model to recognize and differentiate between player characters in video games for customization purposes. Users can select various outfits, skins or colors for their Hero, Hero Red, and Hero Blue characters based on the identification of game classes.

    2. In-Game Advertising and Sponsorships: The model can assist game developers and marketers in identifying specific game characters for dynamic in-game advertising or targeted sponsorships, by determining whether the character on screen is Hero, Hero Red, or Hero Blue.

    3. eSports Analytics and Insights: Leverage the "game" model for real-time analytics and insights in the eSports industry by tracking each class of Hero on the screen during a live-streamed or recorded gaming session. This can help teams and coaches with monitoring character performance, gameplay strategies, and time management.

    4. Accessibility Enhancements: Develop assistive technologies that utilize the model to narrate or describe scenes and characters to visually impaired gamers, by recognizing the Hero, Hero Red, and Hero Blue characters on screen during gameplay.

    5. Content Filtering and Parental Controls: Implement content filtering and parental control mechanisms that can identify specific game classes and characters. Parents can use these features to filter or block games based on the presence of certain character classes like Hero, Hero Red, or Hero Blue to maintain age-appropriate gaming experiences.

  5. R

    Data from: Rmai Dataset

    • universe.roboflow.com
    zip
    Updated Jun 27, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Ayush Ranjan (2022). Rmai Dataset [Dataset]. https://universe.roboflow.com/ayush-ranjan-y2h6m/rmai
    Explore at:
    zipAvailable download formats
    Dataset updated
    Jun 27, 2022
    Dataset authored and provided by
    Ayush Ranjan
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Variables measured
    Letters Bounding Boxes
    Description

    Here are a few use cases for this project:

    1. Educational Tools: The RMAI model can be used to develop applications or tools to teach children or adults about letters and numbers. By scanning real-life objects or text, it can identify the mentioned classes and further enhance the learning experience.

    2. Identification of License Plate Numbers: The model can be employed in surveillance software to identify vehicle license plates. Despite the model not being explicitly trained for this purpose, the ability to recognize the mentioned numeral and letter classes may be sufficient for basic applications.

    3. Robot Navigation: The reference image suggests potential for robot navigation use. Robots could use this model to read numbers and letters in their environment, which could be used in synchronizing tasks or following specified routes in a warehouse or factory setting.

    4. Accessibility Tools: The model can be used to develop applications for visually impaired people to read and comprehend written material. This can range from reading books, recognizing signs, or identifying different objects that have numbers or letters on them.

    5. Data Sorting: In an office or warehouse setting, this model could be used to sort packages, files or items based on numbers and letters. This will help in increasing efficiency and reducing potential errors in the process.

  6. R

    Ocr Multilingual Dataset

    • universe.roboflow.com
    zip
    Updated Apr 11, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    workspace (2025). Ocr Multilingual Dataset [Dataset]. https://universe.roboflow.com/workspace-8hc0w/ocr-multilingual
    Explore at:
    zipAvailable download formats
    Dataset updated
    Apr 11, 2025
    Dataset authored and provided by
    workspace
    License

    MIT Licensehttps://opensource.org/licenses/MIT
    License information was derived automatically

    Variables measured
    Characters Bounding Boxes
    Description

    To train an Optical Character Recognition (OCR) model, a comprehensive dataset is essential. This dataset serves as the foundation for the model's learning process, enabling it to recognize and decipher various fonts, styles, and languages.

  7. R

    Captcha Images Dataset

    • universe.roboflow.com
    zip
    Updated Jul 15, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    CAPTCHA solver (2022). Captcha Images Dataset [Dataset]. https://universe.roboflow.com/captcha-solver/captcha-images/dataset/4
    Explore at:
    zipAvailable download formats
    Dataset updated
    Jul 15, 2022
    Dataset authored and provided by
    CAPTCHA solver
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Variables measured
    Letter Bounding Boxes
    Description

    Here are a few use cases for this project:

    1. Improved CAPTCHA Technologies: Use the model to implement more sophisticated and advanced CAPTCHA for online platforms. It would help in distinguishing humans and bots, leading to better system security.

    2. Data Entry Automation: Deploying this model can help automate data entry processes by decoding CAPTCHA-like handwritten or printed text to digital text.

    3. Digital Archive Transcription: The model can be used to transcribe historical documents or books, which often feature letterings that resemble the jumbled nature of CAPTCHA images, into a digital format.

    4. Improving Optical Character Recognition (OCR) Systems: This model could serve as a training tool for improving OCR systems in recognizing unconventional or distorted characters that often appear in CAPTCHA images.

    5. Assisting Visually Impaired: Develop assistive technologies that could help visually impaired users navigate by converting printed text (that might be in CAPTCHA form due to print deformities) into speech or braille.

  8. R

    Face Features Test Dataset

    • universe.roboflow.com
    zip
    Updated Dec 6, 2021
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Peter Lin (2021). Face Features Test Dataset [Dataset]. https://universe.roboflow.com/peter-lin/face-features-test/dataset/1
    Explore at:
    zipAvailable download formats
    Dataset updated
    Dec 6, 2021
    Dataset authored and provided by
    Peter Lin
    License

    MIT Licensehttps://opensource.org/licenses/MIT
    License information was derived automatically

    Variables measured
    Face Features Bounding Boxes
    Description

    A simple dataset for benchmarking CreateML object detection models. The images are sampled from COCO dataset with eyes and nose bounding boxes added. It’s not meant to be serious or useful in a real application. The purpose is to look at how long it takes to train CreateML models with varying dataset and batch sizes.

    Training performance is affected by model configuration, dataset size and batch configuration. Larger models and batches require more memory. I used CreateML object detection project to compare the performance.

    Hardware

    M1 Macbook Air * 8 GPU * 4/4 CPU * 16G memory * 512G SSD

    M1 Max Macbook Pro * 24 GPU * 2/8 CPU * 32G memory * 2T SSD

    Small Dataset Train: 144 Valid: 16 Test: 8

    Results |batch | M1 ET | M1Max ET | peak mem G | |--------|:------|:---------|:-----------| |16 | 16 | 11 | 1.5 | |32 | 29 | 17 | 2.8 | |64 | 56 | 30 | 5.4 | |128 | 170 | 57 | 12 |

    Larger Dataset Train: 301 Valid: 29 Test: 18

    Results |batch | M1 ET | M1Max ET | peak mem G | |--------|:------|:---------|:-----------| |16 | 21 | 10 | 1.5 | |32 | 42 | 17 | 3.5 | |64 | 85 | 30 | 8.4 | |128 | 281 | 54 | 16.5 |

    CreateML Settings

    For all tests, training was set to Full Network. I closed CreateML between each run to make sure memory issues didn't cause a slow down. There is a bug with Monterey as of 11/2021 that leads to memory leak. I kept an eye on the memory usage. If it looked like there was a memory leak, I restarted MacOS.

    Observations

    In general, more GPU and memory with MBP reduces the training time. Having more memory lets you train with larger datasets. On M1 Macbook Air, the practical limit is 12G before memory pressure impacts performance. On M1 Max MBP, the practical limit is 26G before memory pressure impacts performance. To work around memory pressure, use smaller batch sizes.

    On the larger dataset with batch size 128, the M1Max is 5x faster than Macbook Air. Keep in mind a real dataset should have thousands of samples like Coco or Pascal. Ideally, you want a dataset with 100K images for experimentation and millions for the real training. The new M1 Max Macbooks is a cost effective alternative to building a Windows/Linux workstation with RTX 3090 24G. For most of 2021, the price of RTX 3090 with 24G is around $3,000.00. That means an equivalent windows workstation would cost the same as the M1Max Macbook pro I used to run the benchmarks.

    Full Network vs Transfer Learning

    As of CreateML 3, training with full network doesn't fully utilize the GPU. I don't know why it works that way. You have to select transfer learning to fully use the GPU. The results of transfer learning with the larger dataset. In general, the training time is faster and loss is better.

    batchET minTrain AccVal AccTest AccTop IU TrainTop IU ValidTop IU TestPeak mem Gloss
    1647519127823131.50.41
    3287521107826112.760.02
    641375238782495.30.017
    128257522137825148.40.012

    Github Project

    The source code and full results are up on Github https://github.com/woolfel/createmlbench

  9. R

    Hollow Knight Dataset

    • universe.roboflow.com
    zip
    Updated Jan 23, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Hollow Knight Dataset (2025). Hollow Knight Dataset [Dataset]. https://universe.roboflow.com/hollow-knight-dataset/hollow-knight/dataset/2
    Explore at:
    zipAvailable download formats
    Dataset updated
    Jan 23, 2025
    Dataset authored and provided by
    Hollow Knight Dataset
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Variables measured
    Mana Health_bar Player... Bounding Boxes
    Description

    Project Description for Roboflow

    Project Title: Hollow Knight Object Detection for Reinforcement Learning Agent

    Description:
    This project focuses on developing an object detection model tailored to the popular game Hollow Knight. The goal is to detect and classify various in-game elements in real-time to create a dataset that powers a reinforcement learning (RL) agent. This agent will use the detected objects as inputs to interact with the game environment, make decisions, and achieve specific objectives such as defeating enemies, collecting items, and progressing through the game.

    The object detection model will classify key elements in the game into the following 10 classes:

    1. Mana: Represents the player's magic reserve, displayed as a circular indicator that fills with white. Additionally, smaller circles may appear, representing extra magic capacity.
    2. Health Bar: Displays the player’s health as a series of masks that fill with white. Each mask corresponds to one unit of health.
    3. HK (Hollow Knight): The main character's position, representing the player’s sprite on the screen.
    4. Enemy: Any hostile character or entity that can attack or be attacked by the player. Includes grounded, flying, and stationary enemies.
    5. Collectible Item: Objects that can be picked up or interacted with to provide rewards such as Geo (currency), life fountains, or station stops.
    6. Bench: Resting spots where the player can save progress and heal.
    7. Upgrade Item: Rare collectibles that permanently enhance abilities or stats, such as health or mana upgrades.
    8. Key Item: Special objects necessary for game progression, such as keys or crests.
    9. Exit: Doorways, breakable barriers, or hidden passages that transition the player to new areas.
    10. NPC: Non-hostile characters that the player can interact with for trading, quests, or story progression.

    The object detection system will enable the RL agent to process and interpret the game environment, enabling intelligent decision-making.

    Purpose and Objectives:

    1. Object Detection:
      Develop a robust YOLO-based object detection model to identify and classify game elements from video frames.

    2. Reinforcement Learning (RL):
      Utilize the outputs of the object detection system (e.g., bounding boxes and class predictions) as the state inputs for an RL algorithm. The RL agent will learn to perform tasks such as:

      • Navigating through the game world.
      • Interacting with NPCs.
      • Avoiding or defeating enemies.
      • Collecting items and managing resources like health and mana.
    3. Dynamic Adaptation:
      Begin training the RL agent with a limited dataset of annotated images, gradually expanding the dataset to improve model performance and adaptability as more scenarios are introduced.

    4. Automation:
      The ultimate goal is to automate the gameplay of Hollow Knight, enabling the agent to mimic human-like decision-making.

    Integration Plan:

    1. Object Detection Training:
      Use Roboflow for data preprocessing, annotation, augmentation, and model training. Generate a YOLO-compatible dataset and fine-tune the model for detecting the 10 classes.

    2. Reinforcement Learning Agent:
      Implement a deep RL algorithm (e.g., Deep Q-Networks (DQN) or Proximal Policy Optimization (PPO)).

      • State Input: The bounding boxes and class probabilities from the object detection model.
      • Actions: Movement (e.g., left, right, jump), interactions (e.g., attack, heal, cast spells), and resource management.
      • Rewards: Positive rewards for objectives like collecting items or defeating enemies, and negative rewards for losing health or failing objectives.
    3. Feedback Loop:
      The RL agent's actions will be fed back into the game, generating new frames that the object detection model processes, creating a closed loop for training and evaluation.

  10. R

    Training Dataset

    • universe.roboflow.com
    zip
    Updated Mar 14, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    MyDepressionProj (2023). Training Dataset [Dataset]. https://universe.roboflow.com/mydepressionproj/training-5mx9q/model/3
    Explore at:
    zipAvailable download formats
    Dataset updated
    Mar 14, 2023
    Dataset authored and provided by
    MyDepressionProj
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Variables measured
    Box Letters Number Shapes Bounding Boxes
    Description

    Here are a few use cases for this project:

    1. Object Tracking in Robotics: This model can be used in a robotic picking and sorting system. The robot, equipped with a camera, can identify box-letters-number-shapes and appropriately sort them.

    2. Intelligent Traffic Management: The system can recognize traffic signs (e.g., "STOP", "bullseye", number signs) for real-time traffic monitoring and control. It would especially be useful in autonomous vehicle navigation.

    3. Education and Learning: The model can be used in educational apps or tools for helping children learn about numbers, letters, and shapes, by identifying and naming them within images.

    4. Manufacturing Quality Control: In a manufacturing settings, the model can help in identifying and sorting parts based on shapes, letters, and numbers, increasing efficiency and reducing error rate.

    5. Interactive Gaming: In games, the model can be used to identify specified shapes, numbers, or obstacles, creating an interactive and immersive experience for the players.

  11. R

    F_text Dataset

    • universe.roboflow.com
    zip
    Updated Apr 24, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    new-workspace-rsztl (2022). F_text Dataset [Dataset]. https://universe.roboflow.com/new-workspace-rsztl/f_text/dataset/1
    Explore at:
    zipAvailable download formats
    Dataset updated
    Apr 24, 2022
    Dataset authored and provided by
    new-workspace-rsztl
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Variables measured
    Text Bounding Boxes
    Description

    Here are a few use cases for this project:

    1. Text Translation Application: For apps that are set to translate text from one language to another and are in need of identifying English definite and indefinite articles, operation keywords, names of places, institutes or even certain popular business terms, the "F_Text" computer vision model would be beneficial.

    2. Education and Learning Tools: This model could be used to develop an educational application for kids, where they are trained to identify different classes of English words, symbols, and numbers. The model could pull out words from various contexts and ask the students to categorize them, enhancing their language skills.

    3. Urban Navigation Applications: Applications meant to help people navigate through areas (like Essex or Colchester mentioned within the classes) could utilize this model to recognize those names in various context like in a road sign or a building name.

    4. Access Control Systems: The model could be employed in a scenario where access to certain physical or digital spaces is outfitted with text-based barriers or captcha. Here, only a distinct sequence of these words or symbols would allow access.

    5. Contextual Advertising: In the field of e-commerce or digital advertising, the model could be used to scan digital or physical text sources (like a book in the dataset), identify specific keywords or phrases related to a product or brand and trigger related advertisements. This would make advertising more contextual and personalized.

  12. R

    Paper Parts Fsod Okht Dataset

    • universe.roboflow.com
    zip
    Updated Feb 25, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    roboflow20temp (2025). Paper Parts Fsod Okht Dataset [Dataset]. https://universe.roboflow.com/roboflow20temp/paper-parts-fsod-okht
    Explore at:
    zipAvailable download formats
    Dataset updated
    Feb 25, 2025
    Dataset authored and provided by
    roboflow20temp
    License

    MIT Licensehttps://opensource.org/licenses/MIT
    License information was derived automatically

    Variables measured
    Paper Parts Fsod Okht Okht Bounding Boxes
    Description

    Overview

    Introduction

    This dataset is designed to annotate the structural elements of academic papers. It aims to train models to recognize different parts of a paper. Each class corresponds to a text or graphical element commonly found in papers.

    • Author: The name(s) of the person(s) who wrote the document.
    • Chapter: The major divisions within the paper, usually denoted by a number and a title.
    • Equation: Mathematical formulas or expressions.
    • Equation Number: The numeral identifiers for equations.
    • Figure: Visual representations like graphs or charts.
    • Figure Caption: Text descriptions associated with figures.
    • Footnote: Additional information at the bottom of the page.
    • List of Content Heading: The titles of content sections in a list.
    • List of Content Text: Descriptions or details within a list of content.
    • Page Number: The numeral indicating the page's position.
    • Paragraph: Blocks of text conveying an idea or point.
    • Reference Text: Citations or bibliographic information.
    • Section: Main headings within a chapter.
    • Subsection: Subheadings under a section.
    • Subsubsection: Further subdivisions under a subsection.
    • Table: Data or information arranged in rows and columns.
    • Table Caption: Text descriptions associated with tables.
    • Table of Contents Text: Entries listing sections and page numbers.
    • Title: The main heading or name of the paper.

    Object Classes

    Author

    Description

    Text indicating the name(s) of the author(s), typically found near the beginning of a document.

    Instructions

    Identify the text block containing the author names. It usually follows the title and may include affiliations. Do not include titles or titles of sections adjacent to author names.

    Chapter

    Description

    Indicates a major division of the document, often labeled with a number and title.

    Instructions

    Locate text labeled with "Chapter" followed by a number and title. Capture the entire heading, ensuring no unrelated text is included.

    Equation

    Description

    Symbols and numbers arranged to represent a mathematical concept.

    Instructions

    Draw boxes around all mathematical expressions, excluding any accompanying text or numbers identifying the equations.

    Equation Number

    Description

    Numerals used to uniquely identify equations.

    Instructions

    Identify numbers in parentheses next to equations. Do not include equation text or variables.

    Figure

    Description

    Visual content such as graphs, diagrams, or images.

    Instructions

    Outline the entire graphical representation. Do not include captions or any surrounding text.

    Figure Caption

    Description

    Text providing a description or explanation of a figure.

    Instructions

    Identify the text directly associated with a figure below it. Ensure no unrelated figures or text are included.

    Footnote

    Description

    Clarifications or additional details located at the bottom of a page.

    Instructions

    Locate text at the page's bottom that refers back to a mark or reference in the main text. Exclude any unrelated content.

    List of Content Heading

    Description

    Headings at the start of a list, identifying its purpose or content.

    Instructions

    Identify and label only the heading for lists in content sections. Do not include subsequent list items.

    List of Content Text

    Description

    The detailed entries or points in a list.

    Instructions

    Identify each item in a content list. Exclude list headings and any non-list content.

    Page Number

    Description

    Numerical indication of the current page.

    Instructions

    Locate numbers typically positioned at the top or bottom margins. Do not include text or symbols beside the numbers.

    Paragraph

    Description

    Blocks of text separated by spacing or indentation.

    Instructions

    Enclose individual text blocks that form coherent sections. Ensure each paragraph is distinguished separately.

    Reference Text

    Description

    Bibliographic information found typically in a reference section.

    Instructions

    Identify the full reference entries. Ensure each citation is clearly distinguished without over

  13. Clashroyalechardetector Xus94 Giyri Fsod Voeq Dataset

    • universe.roboflow.com
    zip
    Updated Apr 3, 2025
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Roboflow 100-VL FSOD (2025). Clashroyalechardetector Xus94 Giyri Fsod Voeq Dataset [Dataset]. https://universe.roboflow.com/rf100-vl-fsod/clashroyalechardetector-xus94-giyri-fsod-voeq/dataset/2
    Explore at:
    zipAvailable download formats
    Dataset updated
    Apr 3, 2025
    Dataset provided by
    Roboflowhttps://roboflow.com/
    Authors
    Roboflow 100-VL FSOD
    License

    MIT Licensehttps://opensource.org/licenses/MIT
    License information was derived automatically

    Variables measured
    Clashroyalechardetector Xus94 Giyri Fsod Voeq Bounding Boxes
    Description

    Overview

    Introduction

    The Clash Royale Character Detector dataset is used to train object detection models to identify and distinguish various characters and objects from the game Clash Royale. . These classes are separated into two categories: "Ally" and "Enemy," each with specific characters and objects. The annotation involves drawing bounding boxes around each character or object to train models for automatic detection. All the Allies wear blue whereas the Enemies wear red. The top half of the field belongs to the enemy, whereas the bottom belongs to the ally.

    The classes in the dataset include:

    • Ally Barbarian: A muscular, blonde, shirtless video game character wearing blue.
    • Ally Battle Ram: A wooden battering ram with metal rings around each end wearing blue.
    • Ally Bomber: A small unit with a blue hood holding a bomb with a wick wearing blue.
    • Ally Executioner: A large unit with an dual-ended ax and a hood covering his head wearing blue.
    • Ally Firecracker: A character holding a brass tube to launch rockets, with a purple hood wearing blue.
    • Ally Goblin: A small green-skinned creature wearing blue.
    • Ally Goblin Brawler: A larger and more armored goblin wearing blue.
    • Ally Goblin Cage: A wooden box with accents being blue.
    • Ally Knight: A knight wearing chain-mail and with a rapier, wearing a blue banner.
    • Ally Mini Pekka: A small robot-like unit with blue.
    • Ally Minion: A tiny flying creature with wings whose skin is blue.
    • Ally Pekka: A large robot-like unit with two pairs of eyes and horned helmet with blue.
    • Ally Ram Rider: A female warrior riding a ram, carrying a whip, with tall black hair. She wears blue.
    • Ally Skeleton: A white skeletal figure holding a bone used as a club, wearing blue.
    • Ally Spear Goblin: A goblin carrying a spear with a blue headband.
    • Ally Tombstone: A grey, stone-like tombstone with a skull on it and a blue ribbon going down the middle..
    • Enemy Archer: A character wielding a bow, usually seen in groups, with red attire.
    • Enemy Bat: Small flying bat-like creatures with wings with red accents.∂
    • Enemy Cannon: A stationary cannon with wheels with red accents.
    • Enemy Closed Tesla: A closed metalic trap door structure on the upper half of the field.
    • Enemy Evo Archers: A purple archer with a number in a diamond with a red background above them.
    • Enemy Evo Barbarians: Muscled character with blond hair and a purple head band, with a red health bar near them.
    • Enemy Executioner: Same as ally executioner, but wearing red.
    • Enemy Fire Spirit: Small purple creatures appearing to have a fiery appearance, with red health bars.
    • Enemy Flying Machine: A wooden and metallic structure with rotating wooden blades and a cannon. Has red accents.
    • Enemy Furnace: Oven-like structure emitting fire in the top half of the field.
    • Enemy Goblin: Same as allied goblin but wearing red.
    • Enemy Knight: Same as ally knight but wearing red.
    • Enemy Log: A rolling log with spikes. You can tell it's enemy due to the sparkles on the top side of the log.
    • Enemy Minion: Same as allied minion but it has a red health bar above it.
    • Enemy Mortar: A stationary weapon used for launching projectiles. It looks like a cannon aimed into the air. Usually on the top half of the field with a red healthbar.
    • Enemy Musketeer: A character wielding a musket with a metal helmet and purple hair, having a red health bar above it.
    • Enemy Open Tesla: A tower coming out of an open trap door with a red health bar above it.
    • Enemy Skeleton: Same as ally skeleton, but wearing red.

    Object Classes

    Ally Barbarian

    Description

    A muscular, blonde, shirtless video game character wearing blue.

    Instructions

    Draw bounding boxes around the ally barbarian.

    Ally Battle Ram

    Description

    A wooden battering ram with metal rings around each end wearing blue.

    Instructions

    Draw bounding boxes around the battle ram.

    Ally Bomber

    Description

    A small unit wi

  14. R

    Break Captcha Life Dataset

    • universe.roboflow.com
    zip
    Updated Oct 30, 2021
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Navid (2021). Break Captcha Life Dataset [Dataset]. https://universe.roboflow.com/navid-p4wgh/break-captcha-life/dataset/1
    Explore at:
    zipAvailable download formats
    Dataset updated
    Oct 30, 2021
    Dataset authored and provided by
    Navid
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Variables measured
    Captcha Bounding Boxes
    Description

    Here are a few use cases for this project:

    1. Automated Accessibility: Improve accessibility for visually impaired or disabled users who struggle with solving Captchas by providing alternative methods for user verification, such as audio or text-based questions.

    2. CAPTCHA Improvement Research: Study the effectiveness of the current Captcha classes and develop more secure and user-friendly Captcha systems to prevent malicious automated bots from bypassing the security.

    3. Educational Tool: Teach machine learning and computer vision students and enthusiasts how to train and test models in recognizing and classifying Captcha characters, with the emphasis on proper use and ethical considerations.

    4. Data Entry Quality Assurance: Incorporate the "Break Captcha Life" model into data entry platforms to automatically verify Captcha solutions entered by human users, ensuring they have correctly solved the Captcha before submitting the form.

    5. Security Stress Testing: Test the robustness of various online platforms' Captcha systems by using the "Break Captcha Life" model to evaluate their susceptibility to automated attacks while adhering to responsible disclosure and ethical hacking practices.

  15. R

    Orange_bots Dataset

    • universe.roboflow.com
    zip
    Updated Jul 13, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    DJW (2022). Orange_bots Dataset [Dataset]. https://universe.roboflow.com/djw/orange_bots
    Explore at:
    zipAvailable download formats
    Dataset updated
    Jul 13, 2022
    Dataset authored and provided by
    DJW
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Variables measured
    Human Models Bounding Boxes
    Description

    Here are a few use cases for this project:

    1. Video Game Development: Developers can use the "orange_bots" model to create more immersive and interactive games, especially for games that include 'bots' as part of their character roster. The model could help developers more easily create non-player characters (NPCs) and categorize them properly.

    2. eSports Analysis: This model could be used to study and analyze gameplay in eSports, particularly in recognizing bot strategies and player interactions with bots. This data could then be used to improve game design, player training, or competitive strategies.

    3. Content Moderation: For platforms hosting user-generated gaming content, the model can help identify the portions of the game that include bots. This can assist in moderating content, ensuring the fair play principles are adhered to, and identifying any bot-related cheating.

    4. User-Generated Content Curation: The model can be used as a tool for curating user-generated content, like videos or streamed content featuring gameplay. By recognizing bots, videos could be correctly labeled and categorized for easier discovery.

    5. Interactive Entertainment: This model could be employed in theme parks or virtual reality experiences for user interaction with bots. As users engage with virtual bots, their behaviors and responses can be analyzed to enhance the user experience.

  16. Paper Parts Fsod Rmrg Dataset

    • universe.roboflow.com
    zip
    Updated Jun 4, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Roboflow 20-VL FSOD (2025). Paper Parts Fsod Rmrg Dataset [Dataset]. https://universe.roboflow.com/rf20-vl-fsod/paper-parts-fsod-rmrg/dataset/1
    Explore at:
    zipAvailable download formats
    Dataset updated
    Jun 4, 2025
    Dataset provided by
    Roboflowhttps://roboflow.com/
    Authors
    Roboflow 20-VL FSOD
    License

    MIT Licensehttps://opensource.org/licenses/MIT
    License information was derived automatically

    Variables measured
    Paper Parts Fsod Rmrg Rmrg Bounding Boxes
    Description

    Overview

    Introduction

    This dataset is designed to annotate the structural elements of academic papers. It aims to train models to recognize different parts of a paper. Each class corresponds to a text or graphical element commonly found in papers.

    • Author: The name(s) of the person(s) who wrote the document.
    • Chapter: The major divisions within the paper, usually denoted by a number and a title.
    • Equation: Mathematical formulas or expressions.
    • Equation Number: The numeral identifiers for equations.
    • Figure: Visual representations like graphs or charts.
    • Figure Caption: Text descriptions associated with figures.
    • Footnote: Additional information at the bottom of the page.
    • List of Content Heading: The titles of content sections in a list.
    • List of Content Text: Descriptions or details within a list of content.
    • Page Number: The numeral indicating the page's position.
    • Paragraph: Blocks of text conveying an idea or point.
    • Reference Text: Citations or bibliographic information.
    • Section: Main headings within a chapter.
    • Subsection: Subheadings under a section.
    • Subsubsection: Further subdivisions under a subsection.
    • Table: Data or information arranged in rows and columns.
    • Table Caption: Text descriptions associated with tables.
    • Table of Contents Text: Entries listing sections and page numbers.
    • Title: The main heading or name of the paper.

    Object Classes

    Author

    Description

    Text indicating the name(s) of the author(s), typically found near the beginning of a document.

    Instructions

    Identify the text block containing the author names. It usually follows the title and may include affiliations. Do not include titles, affiliations or titles of sections adjacent to author names.

    Chapter

    Description

    Indicates a major division of the document, often labeled with a number and title.

    Instructions

    Locate text labeled with "Chapter" followed by a number and title. Capture the entire heading, ensuring no unrelated text is included.

    Equation

    Description

    Symbols and numbers arranged to represent a mathematical concept.

    Instructions

    Draw boxes around all mathematical expressions, excluding any accompanying text or numbers identifying the equations.

    Equation Number

    Description

    Numerals used to uniquely identify equations.

    Instructions

    Identify numbers in parentheses next to equations. Do not include equation text or variables.

    Figure

    Description

    Visual content such as graphs, diagrams, code or images.

    Instructions

    Outline the entire graphical representation. Do not include captions or any surrounding text.

    Figure Caption

    Description

    Text providing a description or explanation above or below a figure.

    Instructions

    Identify the text directly associated with a figure. Ensure no unrelated figures or text are included.

    Footnote

    Description

    Clarifications or additional details located at the bottom of a page.

    Instructions

    Locate text at the page's bottom that refers back to a mark or reference in the main text. Exclude any unrelated content.

    List of Content Heading

    Description

    Headings at the list of context text, identifying its purpose or content. This may also be called a list of figures.

    Instructions

    Identify and label only the heading for lists in content sections. Do not include subsequent list items.

    List of Content Text

    Description

    The detailed entries or points in a list. These often summarize all figures in the paper.

    Instructions

    Identify each item in a content list. Exclude list headings and any non-list content.

    Page Number

    Description

    Numerical indication of the current page.

    Instructions

    Locate numbers typically positioned at the top or bottom margins. Do not include text or symbols beside the numbers.

    Paragraph

    Description

    Blocks of text separated by spacing or indentation.

    Instructions

    Enclose individual text blocks that form coherent sections. Ensure each paragraph is distinguished separately.

    Reference Text

    Description

    Bibliographic information found typically in a reference sect

  17. R

    Container Character Codes Dataset

    • universe.roboflow.com
    zip
    Updated Oct 14, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Public Workspace (2022). Container Character Codes Dataset [Dataset]. https://universe.roboflow.com/public-workspace-n6wxn/container-character-codes
    Explore at:
    zipAvailable download formats
    Dataset updated
    Oct 14, 2022
    Dataset authored and provided by
    Public Workspace
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Variables measured
    Container Codes Bounding Boxes
    Description

    This project labels container codes on trucks. It can be used with optical character recognition (OCR) software to identify vehicles entering and exiting facilities or passing a checkpoint via a security camera feed or traffic cam.

    The project includes several exported versions, and fine-tuned models that can be used in the cloud or on an edge device.

  18. R

    Emergent Object Dataset

    • universe.roboflow.com
    zip
    Updated Apr 1, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    HKU UAS (2023). Emergent Object Dataset [Dataset]. https://universe.roboflow.com/hku-uas/emergent-object/dataset/3
    Explore at:
    zipAvailable download formats
    Dataset updated
    Apr 1, 2023
    Dataset authored and provided by
    HKU UAS
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Variables measured
    Manikin Bounding Boxes
    Description

    Here are a few use cases for this project:

    1. Retail Store Analytics: The "Emergent Object" model can be used in observing consumer behavior in retail environments by distinguishing between a human shopper and manikins. This could help glean insights into shopping patterns, time spent near certain displays, or it can be used to enhance security measures.

    2. Virtual Reality: This model could be particularly useful in VR simulations. By being able to distinguish between human players and simulated characters (manikins), it can create a more immersive and interactive experience.

    3. Film and Television Production: The model could be used in the production phase to differentiate between actors and manikins or props, helping in tracking shots, scene comprehension, and subsequent CGI implementation.

    4. Advanced Driver Assistance Systems: The model can be used in vehicular technologies to identify pedestrians crossing the street. This can enhance the accuracy of driver assistance systems and contribute to safer driving.

    5. Manikin-Based Training: In medical or emergency training scenarios where manikins are used to mimic real-life situations, the model can differentiate between learners/participants and manikins, helping in the evaluation of the training session.

  19. R

    Taiwan License Plate Char Recognition Research Dataset

    • universe.roboflow.com
    zip
    Updated Nov 28, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    JackResearch0 (2023). Taiwan License Plate Char Recognition Research Dataset [Dataset]. https://universe.roboflow.com/jackresearch0/taiwan-license-plate-char-recognition-research
    Explore at:
    zipAvailable download formats
    Dataset updated
    Nov 28, 2023
    Dataset authored and provided by
    JackResearch0
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Area covered
    台灣
    Variables measured
    License Plate Bounding Boxes
    Description

    Notice

    This project just for archive usage, you can training you own model

    English

    The "Taiwan License Plate Character Recognition Research" project focuses on identifying characters primarily based on Taiwan license plate fonts, coupled with license plate detection technology. Through our simple yet practical code, users can assemble a full license plate number according to the X-coordinate of the characters. The aim of this project is to optimize the license plate recognition process, enabling a faster, more accurate capture of license plate numbers.

    Generate by GPT4 Here are a few use cases for this project:

    1. Automated Parking System: Utilize the "taiwan-license-plate-char-recognition-research" model to read and recognize license plates in parking lots, allowing for streamlined and automated entry/exit management and billing.

    2. Traffic Surveillance and Enforcement: Integrate the model into traffic monitoring systems to identify traffic violators, such as speeding or running red lights, by capturing and recognizing their license plates, and assist law enforcement in issuing fines or citations.

    3. Stolen Vehicle Detection: Leverage the model within police and security systems to identify stolen or flagged vehicles by matching their license plates in real-time with a database of reported stolen or wanted vehicles.

    4. Intelligent Transportation System: Incorporate the model into smart city infrastructure for monitoring and predicting traffic flow, analyzing road conditions, and managing traffic signals, based on real-time vehicle count and license-plate identification.

    5. Access Control and Security: Implement the model in gated communities, corporate campuses, or sensitive facilities to provide automated access control to authorized vehicles, enhancing security and convenience for residents, employees, and visitors.

    Additional Explanation: The images in this project come from multiple different authors' projects. Prior to the creation of this dataset, we performed the following steps on the images:

    1. Detected the presence of license plates.
    2. Centered and cropped the images around the license plates to reduce training time and maintain the proportion of license plate fonts.
    3. Manually inspected the clarity of the license plate letters.
    4. Completed preliminary annotations.
    5. Conducted semi-automatic annotations using a smaller initial model.
    6. Reviewed each image individually.
    However, due to the complexity of annotating license plate fonts and the high requirements for image quality, we have only annotated over 700 images with our limited manpower. Notably, each image requires at least seven letters to be marked.

    If you have other questions or want to discuss this data set, you can contact: https://t.me/jtx257

    注意事項

    這個項目僅用於存檔使用,您可以訓練自己的模型

    台灣車牌字元識別研究

    英文

    台灣車牌字元識別研究專案主要聚焦於識別基於台灣車牌字體的字元,結合車牌檢測技術。通過我們簡潔實用的程式碼,用戶可以根據字元的X坐標組合出完整的車牌號碼。此項目旨在優化車牌識別過程,使其更快速、準確地捕捉車牌號碼。

    由GPT4生成 以下是此項目的幾個**應用案例**:

    1. 自動停車系統:利用“台灣車牌字元識別研究”模型,在停車場讀取和識別車牌,從而實現出入口管理和計費的自動化。

    2. 交通監控與執法:將模型整合到交通監控系統中,識別違反交通規則的行為,如超速或闖紅燈,通過捕捉並識別其車牌,協助執法部門開出罰單或傳票。

    3. 被盜車輛檢測:在警方和安全系統中利用該模型,通過與報告中被盜或通緝車輛的數據庫即時匹配其車牌,識別被盜或被標記的車輛。

    4. 智能交通系統:將模型納入智慧城市基礎設施,基於實時車輛計數和車牌識別,用於監測和預測交通流量,分析道路條件,並管理交通信號。

    5. 出入控制與安全:在封閉社區、企業園區或敏感設施中實施該模型,為授權車輛提供自動出入控制,提升居民、員工和訪客的安全性和便利性。

    額外說明: 該專案的圖片來自多個不同作者的專案。在製作這個資料集之前,我們已經對照片進行了以下幾個步驟:

    1. 檢測車牌存在。
    2. 以車牌為中心點,切割圖片,旨在降低訓練時間,並保留車牌字型的比例。
    3. 進行人眼檢查,確認車牌字母是否清晰。
    4. 進行初步的標註。
    5. 利用較小的初始模型進行半自動標註。
    6. 對每一張照片進行逐一檢查。
    然而,由於車牌字型的標註難度較大,並且對照片品質要求嚴格,我們在有限的人力下,僅標註了700多張照片。值得注意的是,每張照片至少有7個字母需要進行標記。

    如果對此資料集有其他疑問或想討論的,可聯繫: https://t.me/jtx257

  20. Not seeing a result you expected?
    Learn how you can add new datasets to our index.

Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
YoloV5 Train (2022). Yolo Find Text Dataset [Dataset]. https://universe.roboflow.com/yolov5-train/yolo-find-text

Yolo Find Text Dataset

yolo-find-text

yolo-find-text-dataset

Explore at:
zipAvailable download formats
Dataset updated
Jul 17, 2022
Dataset authored and provided by
YoloV5 Train
License

MIT Licensehttps://opensource.org/licenses/MIT
License information was derived automatically

Variables measured
Text Bounding Boxes
Description

Yolo Find Text

## Overview

Yolo Find Text is a dataset for object detection tasks - it contains Text annotations for 290 images.

## Getting Started

You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.

  ## License

  This dataset is available under the [MIT license](https://creativecommons.org/licenses/MIT).
Search
Clear search
Close search
Google apps
Main menu