https://creativecommons.org/publicdomain/zero/1.0/https://creativecommons.org/publicdomain/zero/1.0/
This dataset was created by Alin Cijov
Released under CC0: Public Domain
Attribution-NoDerivs 4.0 (CC BY-ND 4.0)https://creativecommons.org/licenses/by-nd/4.0/
License information was derived automatically
Give Machines the Power to See People.
This isnât just a dataset â itâs a foundation for building the future of human-aware technology. Carefully crafted and annotated with precision, the People Detection dataset enables AI systems to recognize and understand human presence in dynamic, real-world environments.
Whether youâre building smart surveillance, autonomous vehicles, crowd analytics, or next-gen robotics, this dataset gives your model the eyes it needs.
Created using Roboflow. Optimized for clarity, performance, and scale. Source Dataset on Roboflow â
This is more than a dataset. Itâs a step toward a smarter world â One where machines can understand people.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Description
This synthetic dataset has been generated to facilitate object detection (in YOLO format) for research on dyslexia-related handwriting patterns. It builds upon an original corpus of uppercase and lowercase letters obtained from multiple sources: the NIST Special Database 19 111, the Kaggle dataset âA-Z Handwritten Alphabets in .csv formatâ 222, as well as handwriting samples from dyslexic primary school children of Seberang Jaya, Penang (Malaysia).
In the original dataset, uppercase letters originated from NIST Special Database 19, while lowercase letters came from the Kaggle dataset curated by S. Patel. Additional images (categorized as Normal, Reversal, and Corrected) were collected and labeled based on handwriting samples of dyslexic and non-dyslexic students, resulting in:
Building upon this foundation, the Synthetic Dyslexia Handwriting Dataset presented here was programmatically generated to produce labeled examples suitable for training and validating object detection models. Each synthetic image arranges multiple letters of various classes (Normal, Reversal, Corrected) in a âtext lineâ style on a black background, providing YOLO-compatible .txt
annotations that specify bounding boxes for each letter.
(x, y, width, height)
in YOLO format.0 = Normal
, 1 = Reversal
, and 2 = Corrected
.If you are using this synthetic dataset or the original Dyslexia Handwriting Dataset, please cite the following papers:
111 P. J. Grother, âNIST Special Database 19,â NIST, 2016. [Online]. Available:
https://www.nist.gov/srd/nist-special-database-19
222 S. Patel, âA-Z Handwritten Alphabets in .csv format,â Kaggle, 2017. [Online]. Available:
https://www.kaggle.com/sachinpatel21/az-handwritten-alphabets-in-csv-format
Researchers and practitioners are encouraged to integrate this synthetic dataset into their computer vision pipelines for tasks such as dyslexia pattern analysis, character recognition, and educational technology development. Please cite the original authors and publications if you utilize this synthetic dataset in your work.
The original RAR file was password-protected with the password: WanAsy321. This synthetic dataset, however, is provided openly for streamlined usage.
MIT Licensehttps://opensource.org/licenses/MIT
License information was derived automatically
This dataset consists of microscopic images of blood cells specifically designed for the detection of White Blood Cells (WBC). It is intended for object detection tasks where the goal is to accurately locate and identify WBCs within blood smear images. Researchers and developers can utilize this data to train machine learning models for medical applications such as automated blood cell analysis.
Images: The dataset contains high-resolution microscopic images of blood smears, where WBCs are scattered among Red Blood Cells (RBCs) and platelets. Each image is annotated with bounding boxes around the WBCs.
Annotations: The annotations are provided in YOLO format, where each bounding box is associated with a label for WBC.
images/: Contains the blood cell images in .jpg or .png format. labels/: Contains the annotation files in .txt format (YOLO format), with each file corresponding to an image. Image Size: Varies, but all images are in high resolution suitable for detection tasks.
Medical Image Analysis: This dataset can be used to build models for the automated detection of WBCs, which is a crucial step in diagnosing various blood-related disorders. Object Detection: Ideal for testing object detection algorithms like YOLO, Faster R-CNN, or SSD. Acknowledgments This dataset is created using publicly available microscopic blood cell images, annotated for educational and research purposes. It can be used for developing machine learning models for academic research, prototyping medical applications, or object detection benchmarking.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Most street object datasets require your model to do object detection and extract multiple objects from a single image. Which is fine if you're working with complex models like YOLO or R-CNN. This dataset however is for image classification training that could be done with any simple CNN model or even traditional ML models with enough processing and feature extraction.
A convenient csv file with image paths and encoded labels is provided for use in image data generators.
Dataset contains 9879 images with varying sizes categorized into 7 classes.
This dataset was obtained by performing some processing on the following dataset:
https://www.kaggle.com/datasets/ahmedyoussefff/street-objects-dataset/
https://universe.roboflow.com/project-mzmwg/street-objects-ag7dt
The preprocessing consisted of cropping each object specified by the YOLO format into its own separate image. Preprocessing code is available here: https://www.kaggle.com/code/owm4096/street-objects-classification-dataset-extraction
https://creativecommons.org/publicdomain/zero/1.0/https://creativecommons.org/publicdomain/zero/1.0/
Mendeley Link : https://data.mendeley.com/datasets/pwyyg8zmk5/2
Poribohon-BD is a vehicle dataset of 15 native vehicles of Bangladesh. The vehicles are: i) Bicycle, ii) Boat, iii) Bus, iv) Car, v) CNG, vi) Easy-bike, vii) Horse-cart, viii) Launch, ix) Leguna, x) Motorbike, xi) Rickshaw, xii) Tractor, xiii) Truck, xiv) Van, xv) Wheelbarrow. The dataset contains a total of 9058 images with a high diversity of poses, angles, lighting conditions, weather conditions, backgrounds. All of the images are in JPG format. The dataset also contains 9058 image annotation files. These files state the exact positions of the objects with labels in the corresponding image. The annotation has been performed manually and the annotated values are stored in XML files. LabelImg tool by Tzuta Lin has been used to label the images. Moreover, data augmentation techniques have been applied to keep the number of images comparable to each type of vehicle. Human faces have also been blurred to maintain privacy and confidentiality. The data files are divided into 15 individual folders. Each folder contains images and annotation files of one vehicle type. The 16th folder titled âMulti-class Vehiclesâ contains images and annotation files of different types of vehicles. Poribohon-BD is compatible with various CNN architectures such as YOLO, VGG-16, R-CNN, DPM.
Sichkar V. N. Real time detection and classification of traffic signs based on YOLO version 3 algorithm. Scientific and Technical Journal of Information Technologies, Mechanics and Optics, 2020, vol. 20, no. 3, pp. 418â424. DOI: 10.17586/2226-1494-2020-20-3-418-424 (Full-text available on ResearchGate here: Real time detection and classification of traffic signs based on YOLO version 3 algorithm
Sichkar V. N. Effect of various dimension convolutional layer filters on traffic sign classification accuracy. Scientific and Technical Journal of Information Technologies, Mechanics and Optics, 2019, vol. 19, no. 3, pp. 546â552. DOI: 10.17586/2226-1494-2019-19-3-546-552 (Full-text available on ResearchGate here: Effect of various dimension convolutional layer filters on traffic sign classification accuracy
Design, Train & Test deep CNN for Image Classification. Join the course & enjoy new opportunities to get deep learning skills: https://www.udemy.com/course/convolutional-neural-networks-for-image-classification/
https://github.com/sichkar-valentyn/1-million-images-for-Traffic-Signs-Classification-tasks/blob/main/images/slideshow_classification.gif?raw=true%20=470x516" alt="CNN Course" title="CNN Course">
https://github.com/sichkar-valentyn/1-million-images-for-Traffic-Signs-Classification-tasks/blob/main/images/concept_map.png?raw=true%20=570x410" alt="Concept map" title="Concept map">
https://www.udemy.com/course/convolutional-neural-networks-for-image-classification/
Explore one more dataset used for detection tasks here: https://www.kaggle.com/valentynsichkar/traffic-signs-dataset-in-yolo-format
This is ready to use preprocessed data for Traffic Signs saved into the nine pickle
files.
Original datasets are in the following files:
- train.pickle
- valid.pickle
- test.pickle
Code with detailed description on how datasets were preprocessed is in datasets_preparing.py
Before preprocessing training dataset was equalized making examples in the classes equal as it is shown on the figure below. Histogram of 43 classes for training dataset with their number of examples for Traffic Signs Classification before and after equalization by adding transformated images (brightness and rotation) from original dataset. After equalization, training dataset has increased up to 86989 examples.
https://www.googleapis.com/download/storage/v1/b/kaggle-user-content/o/inbox%2F3400968%2Fb5d9f0189353832e769c2bdd8e25243d%2Fhistogram.png?generation=1567275066871451&alt=media" alt="">
Resulted preprocessed nine files are as follows:
- data0.pickle - Shuffling
- data1.pickle - Shuffling, /255.0 Normalization
- data2.pickle - Shuffling, /255.0 + Mean Normalization
- data3.pickle - Shuffling, /255.0 + Mean + STD Normalization
- data4.pickle - Grayscale, Shuffling
- data5.pickle - Grayscale, Shuffling, Local Histogram Equalization
- data6.pickle - Grayscale, Shuffling, Local Histogram Equalization, /255.0 Normalization
- data7.pickle - Grayscale, Shuffling, Local Histogram Equalization, /255.0 + Mean Normalization
- data8.pickle - Grayscale, Shuffling, Local Histogram Equalization, /255.0 + Mean + STD Normalization
Datasets data0 - data3 have RGB images and datasets data4 - data8 have Gray images.
Shapes of data0 - data3 are as following (RGB):
- x_train: (86989, 3, 32, 32)
- y_train: (86989,)
- x_validation: (4410, 3, 32, 32)
- y_validation: (4410,)
- x_test: (12630, 3, 32, 32)
- y_test: (12630,)
Shapes of data4 - data8 are as following (Gray):
- x_train: (86989, 1, 32, 32)
- y_train: (86989,)
- x_validation: (4410, 1, 32, 32)
- y_validation: (4410,)
- x_test: (12630, 1, 32, 32)
- y_test: (12630,)
mean image
and standard deviation
were calculated from train dataset
and applied to validation and testing datasets for appropriate datasets. When using user's image for classification, it has to be preprocessed fi...
https://cdla.io/sharing-1-0/https://cdla.io/sharing-1-0/
https://i.imgur.com/eVRmfw9.gif" alt="Rock Paper Scissors">
About the Dataset
Dive into the Rock-Paper-Scissors SXSW collectionâa vibrant, crowd-sourced set of 7,521 hand-pose images, fine-tuned and battle-ready for your next computer-vision masterpiece! Exported from Roboflow on March 15, 2024, this dataset is packaged in TensorFlow Object Detection format so you can hit the ground running with training and experimentation.
~ Why Youâll Love It
Massive scale: 7,521 uniquely captured poses, each stretched to 640 Ă 640 px for consistency.
Built-in robustness: Every image is tripled with on-the-fly augmentationsâflips, crops, brightness tweaks, and exposure shiftsâso your model sees every angle, every time.
Three clear classes: rock
, paper
, and scissors
âperfect for straightforward multiclass detection.
~ How It Was Made Born from the Worldâs Largest Game of Rock, Paper, Scissors at SXSW 2023, this collection blends:
~ Plug-and-Play Uses
Rapid prototyping: Train YOLO, SSD, Faster R-CNN (or your favorite detector) in minutes. Edge deployment: Build real-time hand-gesture games for mobile or IoT devices. Research & transfer learning: Use as a springboard for sign-language or fine-grained gesture tasks. Active learning loops: Easily append new samples or tweak augmentations to push performance further.
~ License & Citation
License: CC0 1.0 Public Domainâuse, remix, and share without barriers.
Get ready to rock your next computer-vision projectâno paper cuts included!
Not seeing a result you expected?
Learn how you can add new datasets to our index.
https://creativecommons.org/publicdomain/zero/1.0/https://creativecommons.org/publicdomain/zero/1.0/
This dataset was created by Alin Cijov
Released under CC0: Public Domain