Facebook
TwitterApache License, v2.0https://www.apache.org/licenses/LICENSE-2.0
License information was derived automatically
This dataset is specifically curated for object detection tasks aimed at identifying and classifying road damage and potholes. The original dataset on which this augmented dataset is based, included images labeled with four distinct classes: - Pothole - Alligator Crack - Long Crack - Lat Crack However, for training the model for detecting road damages, it has been placed into 1 class, namely the "Pothole" class, which now also includes the alligator, longitudinal, and lateral cracks.
To enhance the robustness and generalization capability of models trained on this dataset, extensive data augmentation techniques have been applied. The augmentation pipeline includes:
These augmentations ensure that models can learn to recognize road damages under various conditions and viewpoints, improving their detection performance.
Bounding boxes are provided in the YOLO format, ensuring easy integration with popular object detection frameworks. The bounding boxes are adjusted to correspond with the augmented images to maintain annotation accuracy.
The dataset includes the following class:
Class ID Class Name 0 Pothole
The dataset is divided into training, validation, and testing sets with the following proportions:
This split ensures a sufficient amount of data for training the model while maintaining enough data for validation and testing to assess model performance accurately.
This dataset aims to aid researchers and developers in building and fine-tuning models for road damage detection, contributing to safer and more efficient road maintenance systems.
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
## Overview
6 7pm Augmented Dataset is a dataset for object detection tasks - it contains Motorcycle Helmet annotations for 5,125 images.
## Getting Started
You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
## License
This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).
Facebook
TwitterThis dataset was created by Swadesh Jana
Facebook
TwitterThis dataset contains 810 images of 12 different classes of food types. The dataset contains food that is generically found across the globe like Pizzas, Burgers, Fries, etc., and some food items that are geographically specific to India. Those include Idli, Vada, Chapathi, etc. In order for the Yolo model to recognize extremely generic items like fruits and common ingredients, the dataset was trained on Apples, Bananas, Rice, Tomatoes, etc. This dataset was created using roboflow's dataset creator present on the roboflow website. The data was augmented using roboflow's dataset augmentation methods like Flip 90 degrees and different ranges of saturation. The dataset can be used with YoloV5 and YoloV8 as well.
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
## Overview
Bccd_YOLOv5_augmented is a dataset for object detection tasks - it contains Cells_group annotations for 364 images.
## Getting Started
You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
## License
This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).
Facebook
Twitterhttps://public.roboflow.ai/object-detection/chess-full
Provided by Roboflow License: Public Domain
This is a dataset of Chess board photos and various pieces. All photos were captured from a constant angle, a tripod to the left of the board. The bounding boxes of all pieces are annotated as follows: white-king, white-queen, white-bishop, white-knight, white-rook, white-pawn, black-king, black-queen, black-bishop, black-knight, black-rook, black-pawn. There are 2894 labels across 292 images.
https://i.imgur.com/nkjobw1.png" alt="Chess Example">
Follow this tutorial to see an example of training an object detection model using this dataset or jump straight to the Colab notebook.
At Roboflow, we built a chess piece object detection model using this dataset.
https://blog.roboflow.ai/content/images/2020/01/chess-detection-longer.gif" alt="ChessBoss">
You can see a video demo of that here. (We did struggle with pieces that were occluded, i.e. the state of the board at the very beginning of a game has many pieces obscured - let us know how your results fare!)
We're releasing the data free on a public license.
Roboflow makes managing, preprocessing, augmenting, and versioning datasets for computer vision seamless.
Developers reduce 50% of their boilerplate code when using Roboflow's workflow, save training time, and increase model reproducibility.

Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
## Overview
YOLOv5 Only Large Images is a dataset for object detection tasks - it contains MIC annotations for 211 images.
## Getting Started
You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
## License
This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Here are a few use cases for this project:
Horror Movie Content Filtering: Thesis YoloV5 can be used by streaming platforms and content providers to identify and categorize horror movies based on the type of horror imagery present, offering tailored content recommendations to users based on their preferred horror subgenres.
Video Game Scene Classification: Game developers can use Thesis YoloV5 to analyze video game scenes in horror games, enabling more immersive and dynamic gameplay experiences by adapting game environments, NPC interactions, or difficulty levels according to the detected horror elements.
Content Moderation for Online Communities: Online forums, social media platforms, and image sharing sites can utilize Thesis YoloV5 to ensure that users adhere to content policies, automatically moderating and flagging inappropriate horror imagery to maintain a safe and inclusive online community.
Augmented Reality Experiences: Thesis YoloV5 can be integrated into AR applications to generate interactive and engaging horror-themed experiences for entertainment, education, or marketing purposes. Users could interact with AI-generated horror characters, solve puzzles based on detected horror elements, or enjoy immersive and personalized storytelling experiences.
Horror-centric Art and Design: Artists, graphic designers, and filmmakers can use Thesis YoloV5 to analyze and reference horror imagery for creating unique visual styles, mood boards, and thematic concepts for art, design projects, or marketing campaigns centered around horror themes.
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Accurate identification of small tea buds is a key technology for tea harvesting robots, which directly affects tea quality and yield. However, due to the complexity of the tea plantation environment and the diversity of tea buds, accurate identification remains an enormous challenge. Current methods based on traditional image processing and machine learning fail to effectively extract subtle features and morphology of small tea buds, resulting in low accuracy and robustness. To achieve accurate identification, this paper proposes a small object detection algorithm called STF-YOLO (Small Target Detection with Swin Transformer and Focused YOLO), which integrates the Swin Transformer module and the YOLOv8 network to improve the detection ability of small objects. The Swin Transformer module extracts visual features based on a self-attention mechanism, which captures global and local context information of small objects to enhance feature representation. The YOLOv8 network is an object detector based on deep convolutional neural networks, offering high speed and precision. Based on the YOLOv8 network, modules including Focus and Depthwise Convolution are introduced to reduce computation and parameters, increase receptive field and feature channels, and improve feature fusion and transmission. Additionally, the Wise Intersection over Union loss is utilized to optimize the network. Experiments conducted on a self-created dataset of tea buds demonstrate that the STF-YOLO model achieves outstanding results, with an accuracy of 91.5% and a mean Average Precision of 89.4%. These results are significantly better than other detectors. Results show that, compared to mainstream algorithms (YOLOv8, YOLOv7, YOLOv5, and YOLOx), the model improves accuracy and F1 score by 5-20.22 percentage points and 0.03-0.13, respectively, proving its effectiveness in enhancing small object detection performance. This research provides technical means for the accurate identification of small tea buds in complex environments and offers insights into small object detection. Future research can further optimize model structures and parameters for more scenarios and tasks, as well as explore data augmentation and model fusion methods to improve generalization ability and robustness.
Facebook
TwitterCC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
License information was derived automatically
Here are a few use cases for this project:
Culinary Education Apps: The model can be integrated into culinary education software or apps to help students and enthusiasts learn the names and appearances of various Indian dishes. When used in combination with augmented reality, users can view the image of a dish and get information about it instantly.
Healthy Eating and Diet Planning: Nutritionists and dieticians can use this model to categorize and identify Indian food items. Using this information, they can develop personalized meal plans for their clients and provide detailed nutritional breakdowns of traditional Indian meals based on their components.
Restaurant Automation: The model can be used in an automated ordering system at Indian restaurants. By identifying the dishes being served, the system can correctly account for what's been served to each table and add it to their bill, reducing human error in bill generation.
Multicultural Cooking Shows or Competitions: In multi-cuisine cooking shows, judges can use this model to verify if participants have correctly prepared a specified Indian dish. This can help ensure fair judging and provide a more informed analysis of the dishes.
Food Delivery & Recognition Apps: This model can be used in food delivery apps where customers take a photo of a dish and the app recognizes the dish for them. The feature can recommend restaurants where they can order similar dishes or suggest recipes they can try.
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
This dataset was originally created by Team Roboflow and Augmented Startups. To see the current project, which may have been updated since this version, please go here: https://universe.roboflow.com/roboflow-100/poker-cards-cxcvz.
This dataset is part of RF100, an Intel-sponsored initiative to create a new object detection benchmark for model generalizability.
Access the RF100 Github repo: https://github.com/roboflow-ai/roboflow-100-benchmark
Facebook
Twitterhttps://creativecommons.org/publicdomain/zero/1.0/https://creativecommons.org/publicdomain/zero/1.0/
This dataset is a balanced and augmented version of the original Scoliosis Detection Dataset designed for deep learning and computer vision tasks, particularly spinal curvature classification using YOLOv5.
It contains dermatoscopic-style spine X-ray images categorized into four classes based on the severity of scoliosis:
1-derece → Mild scoliosis
2-derece → Moderate scoliosis
3-derece → Severe scoliosis
saglikli → Healthy (no scoliosis)
⚙️ Data Details
Train set: ../train/images
Validation set: ../valid/images
Test set: ../test/images
Total Classes: 4
Balanced Samples: Each class contains approximately 1259 images and labels
Augmentations Applied:
Rotation
Brightness and contrast adjustment
Horizontal flip
Random zoom and cropping
Gaussian noise
These augmentations were used to improve model robustness and reduce class imbalance.
🎯 Use Cases
This dataset is ideal for:
Scoliosis detection and classification research
Object detection experiments (YOLOv5, YOLOv8, EfficientDet)
Transfer learning on medical image datasets
Model comparison and explainability studies
📊 Source
Originally sourced and preprocessed using Roboflow, then restructured and balanced manually for research and experimentation.
Roboflow Project Link: 🔗 View on Roboflow
🧠 License
CC BY 4.0 — Free to use and share with attribution.
Facebook
TwitterMIT Licensehttps://opensource.org/licenses/MIT
License information was derived automatically
This dataset is a curated and preprocessed version of a Scoliosis Spine X-ray dataset, designed specifically for deep learning–based object detection and classification tasks using frameworks like YOLOv5, YOLOv8, and TensorFlow Object Detection API.
It contains annotated spinal X-ray images categorized into three classes, representing different spinal conditions.
train: scoliosis2.v16i.tensorflow/images/train
val: scoliosis2.v16i.tensorflow/images/valid
test: scoliosis2.v16i.tensorflow/images/test
nc: 3
names: ['Vertebra', 'scoliosis spine', 'normal spine']
/images/train/images/valid/images/test.txt files with class, x_center, y_center, width, height).jpg / .pngClasses Description:
To enhance diversity and model robustness, the dataset was augmented using:
This dataset is ideal for:
The dataset was preprocessed and labeled using Roboflow, then manually refined and balanced for research use. Originally derived from a spinal X-ray dataset and adapted for deep learning object detection.
Roboflow Project Link: 🔗 View on Roboflow (add your Roboflow link here)
CC BY 4.0 — Free to use, modify, and share with attribution.
Would you like me to make a short summary version (under 1000 characters) for the “Short Description” field on Kaggle too? It’s required for the dataset card.
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
With the development of integrated circuit packaging technology, the layout of printed circuit boards has become complicated. Moreover, the traditional defect detection methods have been difficult to meet the requirements of high precision. Therefore, in order to solve the problem of low efficiency in defect detection of printed circuit boards, a defect detection method based on pseudo-inverse transform and improved YOLOv5 is proposed. Firstly, a defect image restoration model is constructed to improve image clarity. Secondly, Transformer is introduced to improve YOLOv5, and the batch normalization and network loss function are optimized. These methods improve the speed and accuracy of PCB defect detection. Experimental verification showed that the restoration speed of the image restoration model was 37.60%-42.38% higher than other methods. Compared with other models, the proposed PCB defect detection model had an average increase of 10.90% in recall and 12.87% in average detection accuracy. The average detection accuracy of six types of defects in the self-made PCB data set was over 98.52%, and the average detection accuracy was as high as 99.1%. The results demonstrate that the proposed method can enhance the quality of image processing and optimize YOLOv5 to improve the accuracy of detecting defects in printed circuit boards. This method is demonstrably more effective than existing technology, offering significant value and potential for application in industrial contexts. Its promotion could facilitate the advancement of industrial automation manufacturing.
Facebook
TwitterMIT Licensehttps://opensource.org/licenses/MIT
License information was derived automatically
Here are a few use cases for this project:
"Sports Analysis": The "tennis-tracker" model could be used for detailed game analysis in broadcasting. Being able to differentiate between player-front, player-back, and ball are crucial elements for sports analysts to study player movements, strategies, and game patterns.
"Player Performance Evaluation": Coaches and trainers could use this model to assess players' performance during training or matches. The model's ability to identify players and tennis balls can be used for tracking player movement, speed, consistency, and accuracy, contributing to better training strategies.
"Automated Replay System": This model can be utilized for managing replays in live or recorded matches. It can quickly identify key moments or points of interest (like when a player hits the ball) to create automated highlights or checks for foul play.
"Augmented Reality Tennis Game": Game developers could use this model in the development of AR-based tennis games. The model could identify player and ball positions to create a realistic and interactive gaming experience.
"Crowd Control & Safety Management": During major tournaments, security staff can use this model to monitor crowd behavior. Distinguishing between players, balls, and spectators can help identify potential disruptions or emergencies. It can also ensure player safety, tracking unauthorized individuals entering the court.
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Here are a few use cases for this project:
Workplace Compliance Monitoring: This model can be used in industrial or construction environments where compliance with safety protocol is crucial. The system can automatically identify if workers wear the required safety shoes or not, ensuring immediate action can be taken to reduce safety risks.
Retail Assistance: In a retail setting, the model could help in identifying and categorizing various types of footwear on the shelves. It could specifically highlight safety shoes for customers searching for them, enhancing the shopping experience.
Smart CCTV Surveillance: The model can be leveraged in CCTV footage analysis where it identifies individuals with or without safety footwear in restricted or hazardous areas. It can enable instant notifications for improper attire.
Automated Sorting in Warehouses: In logistics and supply chain warehouses that deal with various kinds of shoes, this model could speed up the packing and sorting processes by correctly identifying and categorizing the safety shoes.
Remote Safety Training: In a virtual or augmented reality training environment, the model could be used for real-time verification if the trainees are wearing their safety shoes correctly, especially in occupations where proper safety training is required.
Facebook
TwitterThis dataset enables real-time object detection of sheep, wolves, dogs, wild dog, redfox, fox, coyote, cow and humans for robotic shepherding applications. Built from raw YOLOv5-style sources, it integrates class balancing, video-based diversity, and strong augmentations to enhance robustness. A recycling strategy is used for rare classes. Compatible with YOLOv5 to YOLOv12, RT-DETR, and ROS 2 deployments on legged robots, the dataset includes labels, images, statistics, and visualizations, ready for direct use in training detection models for autonomous livestock protection.
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Comparison of detection results of different models in PCB dataset.
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
NTS-YOLO:a nocturnal traffic sign detection method based on improved YOLOv5In this paper, a nighttime traffic sign recognition method "NTS-YOLO" is proposed, which consists of three main parts. Firstly, this paper adopts the unsupervised nighttime image enhancement technique proposed by Ye-Young Kim et al. Secondly, the Convolutional Block Attention Module (CBAM) attentional mechanism is introduced on the basis of the YOLOv5 network structure, and lastly, the Optimal Transmission Allocation (OTA) loss function is used to optimize the model's performance in the target detection task. With this approach, the accuracy of predicting the bounding box can be effectively optimized so that the model can predict the location of the target and the bounding box more accurately, thus improving the robustness and stability of the model in the target detection task.Other datasIn this paper, 599 nighttime images from the CCTSDB2021 dataset are referenced, of which 80% of the images (479 images) are used as the training set and 20% of the images (120 images) are used as the validation set. In view of the relatively small number of road sign types at night, 9170 daytime road scene images from the TT100K dataset are also referenced to increase the diversity of the data, which are divided into a training set (7208 images) and a validation set (1962 images) at a ratio of 8:2.Links to other publicly accessible locations of the data:CCTSDB2021:GitHub - csust7zhangjm/CCTSDB2021TT100K:http://cg.cs.tsinghua.edu.cn/traffic-sign/data_model_code/data.zipEnvironmentThe experimental environment consists of a high-performance computer configured with an Intel Core i7 processor, 32GB RAM, and an NVIDIA GeForce RTX 4060 graphics card. PyTorch 2.0.1 was chosen as the main deep learning framework, and CUDA technology was utilized to accelerate model training and inference to ensure computational efficiency and data processing power during the experiment.
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Here are a few use cases for this project:
Advanced Driver-Assistance Systems (ADAS): This model can be used as part of an ADAS, enhancing the vehicle's perception capabilities. It could read and interpret signs, improving navigation accuracy and ensuring better road safety by offering real-time information on traffic conditions and road signs to the driver.
Navigational Apps: This model can be incorporated into navigational apps. It can identify road signs and provide real-time directives to users, potentially offering more dynamic, contextually aware directions than current GPS systems.
Autonomous Vehicles: The NavigateIt model may serve as a critical component for Autonomous vehicles. It can provide essential information on road signs and traffic conditions, helping the autonomous vehicle's decision-making process concerning speed adjustment, direction, and adherence to traffic rules.
Advanced Mapping Services: Mapping services could use this model to automatically update and correct their map databases, ensuring they include up-to-date traffic sign information.
Augmented Reality (AR) Applications: In an AR setting, the model can be used to provide context-based information to users. For instance, it could identify signs and inform walkers or cyclists about routes, potential hazards, or guidance based on the perceived traffic signs.
Facebook
TwitterApache License, v2.0https://www.apache.org/licenses/LICENSE-2.0
License information was derived automatically
This dataset is specifically curated for object detection tasks aimed at identifying and classifying road damage and potholes. The original dataset on which this augmented dataset is based, included images labeled with four distinct classes: - Pothole - Alligator Crack - Long Crack - Lat Crack However, for training the model for detecting road damages, it has been placed into 1 class, namely the "Pothole" class, which now also includes the alligator, longitudinal, and lateral cracks.
To enhance the robustness and generalization capability of models trained on this dataset, extensive data augmentation techniques have been applied. The augmentation pipeline includes:
These augmentations ensure that models can learn to recognize road damages under various conditions and viewpoints, improving their detection performance.
Bounding boxes are provided in the YOLO format, ensuring easy integration with popular object detection frameworks. The bounding boxes are adjusted to correspond with the augmented images to maintain annotation accuracy.
The dataset includes the following class:
Class ID Class Name 0 Pothole
The dataset is divided into training, validation, and testing sets with the following proportions:
This split ensures a sufficient amount of data for training the model while maintaining enough data for validation and testing to assess model performance accurately.
This dataset aims to aid researchers and developers in building and fine-tuning models for road damage detection, contributing to safer and more efficient road maintenance systems.