Facebook
TwitterFENCING SCOREBOARD DATASET (YOLOv8 FORMAT)
Project: CMU Fencing Classification Project Author: Michael Stefanov (Carnegie Mellon University) License: MIT Date: 2025
Description:
Labeled images of fencing scoreboards in lit and unlit states, used to train the YOLOv8 detection model. Includes augmented samples and negatives for robust learning.
Dataset Summary:
Total Images: ~2000 Splits: train (1600), valid (400) Classes: 1 ("scoreboard") Format: YOLOv8… See the full description on the dataset page: https://huggingface.co/datasets/mastefan/fencing-scoreboard-yolov8.
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Comparison between the improved model and the original YOLOv8 classification model.
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Manual differentiation of benign hematogones vs. B-ALL malignant sub-types (Pre-B, Pro-B) on bone-marrow smears is time-consuming and error-prone. We built an end-to-end computer-vision workflow that segments single cells, balances class counts, and benchmarks seven modern CNN / YOLO classifiers.
| Step | Purpose | Key details |
|---|---|---|
| Exploration | quantify imbalance | 3 242 images, 4 original classes (benign 512, pre-B 955, pro-B 796, early-pre-B 979) |
| Segmentation | isolate blast regions | LAB-A-channel → K-means (k = 2) → morphology → size ≥ 500 px |
| Standardisation | model-ready tensors | 224 × 224 resize → float [0–1] normalisation |
| Pairing & Splitting | dual-input & leakage-free split | original + mask pairs, 85 / 10 / 5 % (train/val/test) |
| Augmentation | balance classes | flips to 811 pairs per class (training only) |
All CNN backbones are frozen; heads trained for 30 epochs, LR = 1 e-3, batch = 32.
| Model | Top-1 Acc | Params | Input |
|---|---|---|---|
| YOLOv8-n | 100 % | 1.82 M | mask |
| YOLOv11-n | 100 % | 1.63 M | mask |
| YOLOv12-n | 100 % | 1.82 M | mask |
| MobileNetV2 | 99.1 % | 2.59 M (0.33 M trainable) | mask |
| Dual-channel MobileNetV2 | 99.1 % | 3.70 M (1.44 M trainable) | ori + mask |
| NASNet-Mobile | 96.4 % | 4.54 M (0.27 M trainable) | mask |
| EfficientNet-B0 | 35.1 % | 4.38 M (0.33 M trainable) | mask |
Best performer: YOLOv8-n – 100 % accuracy with lightweight 1.8 M parameters.
Bottom line: A compact YOLO classifier, fed only segmented cell masks, can achieve perfect subtype recognition on our internal dataset—promising for real-time, point-of-care B-ALL decision support.
Facebook
TwitterMIT Licensehttps://opensource.org/licenses/MIT
License information was derived automatically
This dataset is a curated and preprocessed version of a Scoliosis Spine X-ray dataset, designed specifically for deep learning–based object detection and classification tasks using frameworks like YOLOv5, YOLOv8, and TensorFlow Object Detection API.
It contains annotated spinal X-ray images categorized into three classes, representing different spinal conditions.
train: scoliosis2.v16i.tensorflow/images/train
val: scoliosis2.v16i.tensorflow/images/valid
test: scoliosis2.v16i.tensorflow/images/test
nc: 3
names: ['Vertebra', 'scoliosis spine', 'normal spine']
/images/train/images/valid/images/test.txt files with class, x_center, y_center, width, height).jpg / .pngClasses Description:
To enhance diversity and model robustness, the dataset was augmented using:
This dataset is ideal for:
The dataset was preprocessed and labeled using Roboflow, then manually refined and balanced for research use. Originally derived from a spinal X-ray dataset and adapted for deep learning object detection.
Roboflow Project Link: 🔗 View on Roboflow (add your Roboflow link here)
CC BY 4.0 — Free to use, modify, and share with attribution.
Would you like me to make a short summary version (under 1000 characters) for the “Short Description” field on Kaggle too? It’s required for the dataset card.
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Book covers typically contain a wealth of information. With the annual increase in the number of books published, deep learning has been utilised to achieve automatic identification and classification of book covers. This approach overcomes the inefficiency of traditional manual classification operations and enhances the management efficiency of modern book retrieval systems. In the realm of computer vision, the YOLO algorithm has garnered significant attention owing to its excellent performance across various visual tasks. Therefore, this study introduces the CPPDE-YOLO model, a novel dual-convolution adaptive focus neural network that integrates the PConv and PWConv operators, alongside dynamic sampling technology and efficient multi-scale attention. By incorporating specific enhancement features, the original YOLOv8 framework has been optimised to yield superior performance in book cover classification. The aim of this model is to significantly enhance the accuracy of image classification by refining the algorithm. For effective book cover classification, it is imperative to consider complex global feature information to capture intricate features while managing computational costs. To address this, we propose a hybrid model that integrates parallel convolution and point-by-point convolution within the backbone network, integrating it into the DualConv framework to capture complex feature information. Moreover, we integrate the efficient multi-scale attention mechanism into each cross stage partial network fusion residual block in the head section to focus on learning key features for more precise classification. The dynamic sampling method is employed instead of the traditional UPsample method to overcome its inherent limitations. Finally, experimental results on real datasets validate the performance enhancement of our proposed CPPDE-YOLO network structure compared to the original YOLOv8 classification structure, achieving Top_1 Accuracy and Top_5 Accuracy improvement of 1.1% and 1.0%, respectively. This underscores the effectiveness of our proposed algorithm in enhancing book genre classification.
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
The EyeOnWater app is designed to assess the ocean's water quality using images captured by regular citizens. In order to have an extra helping hand in determining whether an image meets the criteria for inclusion in the app, the YOLOv8 model for image classification is employed. With the help of this model all uploaded pictures are assessed. If the model deems a water image unsuitable, it is excluded from the app's online database. In order to train this model a training dataset containing a large pool of different images is required. The training dataset includes 12,357 'good' and 10,019 'bad' water quality images that were submitted to the EyeOnWater app.
In order to create a larger training dataset the set of original images (containing a total of 1700 images) are augmented, by rotating, displacing and resizing them. Using the following settings:
The training dataset is 80% used for training, 10% for validation and 10% for prediction.
The training dataset contains 2 classes with 2 labels 'good' and 'bad'. The 'good' images contain water images that are suited to determine the water quality using the Forel-Ule scale. The 'bad' images can include for example too much water reflection, a visible bottom surface, objects or not even include water at all.
From the images the water quality can be obtained by comparing the water color to the 21 colors in the Forel-Ule scale.
Parameter: http://vocab.nerc.ac.uk/collection/P01/current/CLFORULE/
The images are taken by citizen scientists, often with a smartphone.
As the images are taken by smartphones, the image quality can be low. Next to this, the images are taken outside, in a non-confined space, meaning that there can be bad lightning, reflections and other problems occurring. Therefore, the images need first to be checked before they can be included in the app.
Larger images are resized to 256px by 256px, smaller images are excluded from the training dataset.
Images are taken on a global scale.
For more information on the training dataset and/or the app, you can contact tjerk@maris.nl.
Facebook
Twitterhttps://github.com/siromermer/CS2-CSGO-Yolov8-Yolov7-ObjectDetection
in .yaml file there are 5 classes , but actual class number is 4 . When annotating images i mistakenly give blank space in classes.txt file and because of that there is empty class which is 0 in this case , but it wont create any problem , i just wanted to inform kaggle users . Dataset is little bit small for now , but as soos as possible i will increase image number for sure
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Compare different datasets using the same algorithm.
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Accuracy values of 30 categories according to CPPDE-YOLO model classification.
Facebook
TwitterMIT Licensehttps://opensource.org/licenses/MIT
License information was derived automatically
Precious Gemstone Identification
Description: This comprehensive dataset comprises annotated images of a diverse range of precious gemstones meticulously curated for gemstone identification tasks. With 87 classes of gemstones for classification unique varieties including Chalcedony Blue, Amber, Aventurine Yellow, Dumortierite, Pearl, Aventurine Green, and many others, this dataset serves as a valuable resource for training and evaluating machine learning models in gemstone recognition.
Gemstone Variety: The dataset encompasses a wide spectrum of precious gemstones, ranging from well-known varieties like Emerald, Ruby, Sapphire, and Diamond to lesser-known gems such as Benitoite, Larimar, and Sphene.
Dataset Split: Train Set: 92% (46404 images) Validation Set: 4% (1932 images) Test Set: 4% (1932 images)
Preprocessing: Images in the dataset have been preprocessed to ensure consistency and quality:
Augmentations: To enhance model robustness and generalization, each training example has been augmented with various transformations:
File Formats Available:
Disclaimer:
The images included in this dataset were sourced from various online platforms, primarily from minerals.net and www.rasavgems.com websites, as well as other online datasets. We have curated and annotated these datasets for the purpose of gemstone identification and made them available in different formats. We do not claim ownership of the original images, and we do not claim to own these images. Any trademarks, logos, or copyrighted materials belong to their respective owners.
Researchers, enthusiasts and developers interested in gemstone identification, machine learning, and computer vision applications will find this dataset invaluable for training and benchmarking gemstone recognition algorithms.
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Possible configuration schemes for multiple PConv and PWConv.
Facebook
TwitterMIT Licensehttps://opensource.org/licenses/MIT
License information was derived automatically
This dataset is designed for the detection and classification of space debris, aiming to enhance space situational awareness and contribute to the mitigation of space debris hazards. It provides annotated images suitable for training machine learning models in object detection tasks.
Preprocessing:
Augmentations:
This dataset is suitable for developing and evaluating object detection models focused on identifying space debris and satellites. Potential applications include:
If you utilize this dataset in your research or projects, please cite it as follows:
@misc{space-debris-and-satellite-dataset,
title = {Space Debris and Satellite Dataset},
type = {Open Source Dataset},
author = {Mahmoud},
howpublished = {\url{https://universe.roboflow.com/mahmoud-xm4kv/space-debris-and-satilite}},
url = {https://universe.roboflow.com/mahmoud-xm4kv/space-debris-and-satilite},
journal = {Roboflow Universe},
publisher = {Roboflow},
year = {2024},
month = {sep},
note = {visited on 2024-10-04}
}
Facebook
Twitterhttps://creativecommons.org/publicdomain/zero/1.0/https://creativecommons.org/publicdomain/zero/1.0/
This dataset is a balanced and augmented version of the original Scoliosis Detection Dataset designed for deep learning and computer vision tasks, particularly spinal curvature classification using YOLOv5.
It contains dermatoscopic-style spine X-ray images categorized into four classes based on the severity of scoliosis:
1-derece → Mild scoliosis
2-derece → Moderate scoliosis
3-derece → Severe scoliosis
saglikli → Healthy (no scoliosis)
⚙️ Data Details
Train set: ../train/images
Validation set: ../valid/images
Test set: ../test/images
Total Classes: 4
Balanced Samples: Each class contains approximately 1259 images and labels
Augmentations Applied:
Rotation
Brightness and contrast adjustment
Horizontal flip
Random zoom and cropping
Gaussian noise
These augmentations were used to improve model robustness and reduce class imbalance.
🎯 Use Cases
This dataset is ideal for:
Scoliosis detection and classification research
Object detection experiments (YOLOv5, YOLOv8, EfficientDet)
Transfer learning on medical image datasets
Model comparison and explainability studies
📊 Source
Originally sourced and preprocessed using Roboflow, then restructured and balanced manually for research and experimentation.
Roboflow Project Link: 🔗 View on Roboflow
🧠 License
CC BY 4.0 — Free to use and share with attribution.
Facebook
TwitterMIT Licensehttps://opensource.org/licenses/MIT
License information was derived automatically
This dataset is a curated and preprocessed collection of spinal X-ray images for deep learning–based scoliosis and vertebra detection using YOLOv5, YOLOv8, or other object detection frameworks.
It contains high-quality annotated X-rays featuring multiple bounding boxes per image — each representing different spinal regions and conditions.
train: scoliosis yolov5/train/images
val: scoliosis yolov5/valid/images
test: scoliosis yolov5/test/images
nc: 3
names: ['Vertebra', 'scoliosis spine', 'normal spine']
/train/images/valid/images/test/images.txt with class, x_center, y_center, width, height).jpg / .pngClasses Description:
To improve model generalization and balance the dataset, the following augmentations were used:
This dataset is ideal for:
The dataset was processed and annotated using Roboflow, then refined and organized into YOLOv5 format for seamless training. Each image includes verified bounding boxes for vertebral and scoliosis regions.
Roboflow Project Link: 🔗 View on Roboflow (add your Roboflow link here)
CC BY 4.0 — Free to use, modify, and redistribute with proper attribution.
Facebook
TwitterUltralytics YOLOv8 is a cutting-edge, state-of-the-art (SOTA) model that builds upon the success of previous YOLO versions and introduces new features and improvements to further boost performance and flexibility. YOLOv8 is designed to be fast, accurate, and easy to use, making it an excellent choice for a wide range of object detection and tracking, instance segmentation, image classification and pose estimation tasks.
We hope that the resources here will help you get the most out of YOLOv8. Please browse the YOLOv8 Docs for details, raise an issue on GitHub for support, and join our Discord community for questions and discussions!
Not seeing a result you expected?
Learn how you can add new datasets to our index.
Facebook
TwitterFENCING SCOREBOARD DATASET (YOLOv8 FORMAT)
Project: CMU Fencing Classification Project Author: Michael Stefanov (Carnegie Mellon University) License: MIT Date: 2025
Description:
Labeled images of fencing scoreboards in lit and unlit states, used to train the YOLOv8 detection model. Includes augmented samples and negatives for robust learning.
Dataset Summary:
Total Images: ~2000 Splits: train (1600), valid (400) Classes: 1 ("scoreboard") Format: YOLOv8… See the full description on the dataset page: https://huggingface.co/datasets/mastefan/fencing-scoreboard-yolov8.