Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
The goal of this task is to train a model that can localize and classify each instance of Person and Car as accurately as possible.
from IPython.display import Markdown, display
display(Markdown("../input/Car-Person-v2-Roboflow/README.roboflow.txt"))
In this Notebook, I have processed the images with RoboFlow because in COCO formatted dataset was having different dimensions of image and Also data set was not splitted into different Format. To train a custom YOLOv7 model we need to recognize the objects in the dataset. To do so I have taken the following steps:
Image Credit - jinfagang
!git clone https://github.com/WongKinYiu/yolov7 # Downloading YOLOv7 repository and installing requirements
%cd yolov7
!pip install -qr requirements.txt
!pip install -q roboflow
!wget "https://github.com/WongKinYiu/yolov7/releases/download/v0.1/yolov7.pt"
import os
import glob
import wandb
import torch
from roboflow import Roboflow
from kaggle_secrets import UserSecretsClient
from IPython.display import Image, clear_output, display # to display images
print(f"Setup complete. Using torch {torch._version_} ({torch.cuda.get_device_properties(0).name if torch.cuda.is_available() else 'CPU'})")
https://camo.githubusercontent.com/dd842f7b0be57140e68b2ab9cb007992acd131c48284eaf6b1aca758bfea358b/68747470733a2f2f692e696d6775722e636f6d2f52557469567a482e706e67">
I will be integrating W&B for visualizations and logging artifacts and comparisons of different models!
try:
user_secrets = UserSecretsClient()
wandb_api_key = user_secrets.get_secret("wandb_api")
wandb.login(key=wandb_api_key)
anonymous = None
except:
wandb.login(anonymous='must')
print('To use your W&B account,
Go to Add-ons -> Secrets and provide your W&B access token. Use the Label name as WANDB.
Get your W&B access token from here: https://wandb.ai/authorize')
wandb.init(project="YOLOvR",name=f"7. YOLOv7-Car-Person-Custom-Run-7")
https://uploads-ssl.webflow.com/5f6bc60e665f54545a1e52a5/615627e5824c9c6195abfda9_computer-vision-cycle.png" alt="">
In order to train our custom model, we need to assemble a dataset of representative images with bounding box annotations around the objects that we want to detect. And we need our dataset to be in YOLOv7 format.
In Roboflow, We can choose between two paths:
https://raw.githubusercontent.com/Owaiskhan9654/Yolo-V7-Custom-Dataset-Train-on-Kaggle/main/Roboflow.PNG" alt="">
user_secrets = UserSecretsClient()
roboflow_api_key = user_secrets.get_secret("roboflow_api")
rf = Roboflow(api_key=roboflow_api_key)
project = rf.workspace("owais-ahmad").project("custom-yolov7-on-kaggle-on-custom-dataset-rakiq")
dataset = project.version(2).download("yolov7")
Here, I am able to pass a number of arguments: - img: define input image size - batch: determine
https://creativecommons.org/publicdomain/zero/1.0/https://creativecommons.org/publicdomain/zero/1.0/
Sharpen your Cricket AI: Unleash the power of YOLOv8 for precise cricket ball detection in images and videos with this comprehensive dataset.
Fuel Your Custom Training: Build a robust cricket ball detection model tailored to your specific needs. This dataset, featuring 1778 meticulously annotated images in YOLOv8 format, serves as the perfect launchpad.
In-Action Balls: Train your model to identify cricket balls in motion, capturing deliveries, fielding plays, and various gameplay scenarios.
Lighting Variations: Adapt to diverse lighting conditions (day, night, indoor) with a range of images showcasing balls under different illumination.
Background Complexity: Prepare your model for real-world environments. The dataset includes images featuring stadiums, practice nets, and various background clutter.
Ball States: Train effectively with images of new and used cricket balls, encompassing varying degrees of wear and tear.
Real-time Cricket Analysis: Power applications for in-depth player analysis, ball trajectory tracking, and automated umpiring systems.
Enhanced Broadcasting Experiences: Integrate seamless ball tracking, on-screen overlays, and real-time highlights into cricket broadcasts.
Automated Summarization: Streamline cricket video processing for automated highlight reels, focusing on key ball-related moments.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Accident Detection Model is made using YOLOv8, Google Collab, Python, Roboflow, Deep Learning, OpenCV, Machine Learning, Artificial Intelligence. It can detect an accident on any accident by live camera, image or video provided. This model is trained on a dataset of 3200+ images, These images were annotated on roboflow.
https://user-images.githubusercontent.com/78155393/233774342-287492bb-26c1-4acf-bc2c-9462e97a03ca.png" alt="Survey">
MIT Licensehttps://opensource.org/licenses/MIT
License information was derived automatically
This dataset contains 8,992 images of Uno cards and 26,976 labeled examples on various textured backgrounds.
This dataset was collected, processed, and released by Roboflow user Adam Crawshaw, released with a modified MIT license: https://firstdonoharm.dev/
https://i.imgur.com/P8jIKjb.jpg" alt="Image example">
Adam used this dataset to create an auto-scoring Uno application:
Fork or download this dataset and follow our How to train state of the art object detector YOLOv4 for more.
See here for how to use the CVAT annotation tool.
Roboflow makes managing, preprocessing, augmenting, and versioning datasets for computer vision seamless. :fa-spacer: Developers reduce 50% of their boilerplate code when using Roboflow's workflow, save training time, and increase model reproducibility. :fa-spacer:
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
## Overview
WD is a dataset for object detection tasks - it contains Weapons annotations for 2,451 images.
## Getting Started
You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
## License
This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).
Training YOLO v3 for Objects Detection with Custom Data. Build your own detector by labelling, training and testing on image, video and in real time with camera. Join here: https://www.udemy.com/course/training-yolo-v3-for-objects-detection-with-custom-data/
Detections on video and image are shown below. Trained weights can be found in the course mentioned above.
https://www.googleapis.com/download/storage/v1/b/kaggle-user-content/o/inbox%2F3400968%2Fbcdae0b57021d6ac3e86a9aa2e8c4b08%2Fts_detections.gif?generation=1581700736851192&alt=media" alt="Video" title="Detections of Traffic Signs on Video">
https://www.googleapis.com/download/storage/v1/b/kaggle-user-content/o/inbox%2F3400968%2F20ca377934c5ed5f8c1e4272c838b01a%2Fts_detections.jpg?generation=1581701085754638&alt=media" alt="Image" title="Detections of Traffic Signs on Image">
https://www.googleapis.com/download/storage/v1/b/kaggle-user-content/o/inbox%2F3400968%2Fc51fc6aba2c0cd6d22512f486880868a%2FConcept_map_YOLO_3.png?generation=1584694252456677&alt=media" alt="Concept Map of the Course" title="Concept Map of the Course YOLO v3">
https://www.udemy.com/course/training-yolo-v3-for-objects-detection-with-custom-data/
Explore one more dataset used for classification tasks here: https://www.kaggle.com/valentynsichkar/traffic-signs-preprocessed
This is ready to use Traffic Signs Dataset in YOLO format for Detection tasks. It can be used for training as well as for testing. Dataset consists of images in *.jpg
format and *.txt
files next to every image that have the same names as images files have. These *.txt
files include annotations of bounding boxes of Traffic Sings in the YOLO format:
[Class Number] [center in x] [center in y] [Width] [Height]
For example, file 00001.txt
includes three bounding boxes (each in a new line) that describe three Traffic Signs in 00001.jpg
image:
2 0.7378676470588236 0.5125 0.030147058823529412 0.055
2 0.3044117647058823 0.65375 0.041176470588235294 0.0725
3 0.736764705882353 0.453125 0.04264705882352941 0.06875
Traffic Sins in this Dataset are grouped into four categories:
prohibitory
danger
mandatory
other
Prohibitory category consists of following Traffic Signs: speed limit
, no overtaking
, no traffic both ways
, no trucks
.
Danger category consists of following Traffic Sings: priority at next intersection
, danger
, bend left
, bend right
, bend
, uneven road
, slippery road
, road narrows
, construction
, traffic signal
, pedestrian crossing
, school crossing
, cycles crossing
, snow
, animals
.
Mandatory category consists of following Traffic Sings: go right
, go left
, go straight
, go right or straight
, go left or straight
, keep right
, keep left
, roundabout
.
Other category consists of following Traffic Sings: restriction ends
, priority road
, give way
, stop
, no entry
.
Dataset itself is in zip
archive and organization is as following:
+--ts/
| 00000.txt
| 00000.jpg
| 00001.txt
| 00001.jpg
| ...
To train in Darknet framework, dataset is accompanied with following files:
ts_data.data
classes.names
train.txt
test.txt
yolov3_ts_test.cfg
yolov3_ts_train.cfg
Pay attention! You need to specify yours full paths before use these additional files. Find more details in the course mentioned above. Also, you can use especially designed Python file to get full path getting-full-path.py
.
Here is testing results after training in Darknet framework on this dataset. Total number of iterations are 8000
. For training used 631 images
and for validation during training used 112 images
.
mAP (mean average precision), calculated on test images, is as following:
total number detections
= 270
prohibitory
, ap (average precision) = 97.22% (TP = 78, FP = 2)
danger
, ap (average precision) = 100.00% (TP = 24, FP = 0)
mandatory
, ap (average precision) = 94.50% (TP = 23, FP = 2)
other
, ap (average precision) = 97.10% (TP = 39, FP = 1)
mAP = 97.21 %
Initial data is German Traffic Sign Detection Benchmark (GTDRB).
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Augmented custom dataset with images sourced from online sources and camera captures. The dataset was used to train YOLO models for road obstacle detection on African roads specificallly.
Attribution-NonCommercial-ShareAlike 4.0 (CC BY-NC-SA 4.0)https://creativecommons.org/licenses/by-nc-sa/4.0/
License information was derived automatically
This repository contains the supplementary materials presented in the publication “Learning Manufacturing Computer Vision Systems Using Tiny YOLO v4” by Medina, A., Bradley, R., Xu, W., Ponce, P., Anthony, B., and Molina, A. that can be found with the following DOI 10.3389/frobt.2024.1331249
There are three files in this repository:
This Dataset is for an example used for education purposes. It is a small dataset that is adapted from the following Kaggle repository, authored by Ruthger Righart https://www.kaggle.com/datasets/rrighart/jarlids/data. One of the activities proposed is to teach students how to find, download and review a free dataset, so this is the example given.
Another activity is to teach how to label images to create a custom dataset. The images (with extension .JPG) from the original repository are used. The labels (with extension .txt) were created by the authors of the Learning Manufacturing Computer Vision Systems Using Tiny YOLOv4 paper. The authors used the free tool labelImg, from GitHub repository (https://github.com/HumanSignal/labelImg), to label the images with object bounding boxes and corresponding labels in the YOLO format.
The dataset contains 238 images and corresponding labels, with files named “p
This notebook was created to give the user a step-by-step tutorial on how to train a YOLOv4 algorithm with a custom dataset using a free GPU on Google Collab, the prerequisite to use it are:
There are other requirements like cloning a GitHub repository and altering certain files on that repository; however, those steps are discussed within the notebook.
At the end of the notebook an example on how to test the trained model with images and/or videos is shown, however since Google Collab doesn’t have access to the physical computer of the user live stream video is not part of the example.
Disclaimer: This code is not optimized, and its intended purpose is to teach students how to run YOLO on a raspberry pi using the OpenCV library.
To use this code with different files or datasets, be sure to change the two parameters inside the net3 variable which are the cfg file used while training the algorithm and the weights file. You should also change the class list to include your classes, keeping in mind that the classes order must correspond to the order of the labeling process and class 0 is the first one on the list.
Also to change the Title of the created image prompt you shout go to the line calling the imshow method and change the ‘Tiny YOLOv4’ string.
This algorithm uses the first camera it finds and opens up a display image with the detected objects surrounded by a bounding box, on top of that box the top predicted class is going to show, to change color of bounding boxes or text change the rectangle method where it says GREEN as well as in the next code line ant change the number to change the thickness of the line.
This code has a hardcoded confidence threshold for both the YOLO objectevness score and the class score, this can be found in the NMSBoxes method and the if confidence line accordingly. The main value to change first is the if confidence value.
To close the image, you need to press the key ‘q’ as closing the display window is not going to work as it will reopen again.
Note: This code allows the pop-up window, which displays the detections, to be closed only when the "q" key is pressed. Simply closing the window will not work.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Custom Multi-Altitude Aerial Vehicles Dataset:
Created for publishing results for ICUAS 2023 paper "How High can you Detect? Improved accuracy and efficiency at varying altitudes for Aerial Vehicle Detection", following the abstract of the paper.
Abstract—Object detection in aerial images is a challenging task mainly because of two factors, the objects of interest being really small, e.g. people or vehicles, making them indistinguishable from the background; and the features of objects being quite different at various altitudes. Especially, when utilizing Unmanned Aerial Vehicles (UAVs) to capture footage, the need for increased altitude to capture a larger field of view is quite high. In this paper, we investigate how to find the best solution for detecting vehicles in various altitudes, while utilizing a single CNN model. The conditions for choosing the best solution are the following; higher accuracy for most of the altitudes and real-time processing ( > 20 Frames per second (FPS) ) on an Nvidia Jetson Xavier NX embedded device. We collected footage of moving vehicles from altitudes of 50-500 meters with a 50-meter interval, including a roundabout and rooftop objects as noise for high altitude challenges. Then, a YoloV7 model was trained on each dataset of each altitude along with a dataset including all the images from all the altitudes. Finally, by conducting several training and evaluation experiments and image resizes we have chosen the best method of training objects on multiple altitudes to be the mixup dataset with all the altitudes, trained on a higher image size resolution, and then performing the detection using a smaller image resize to reduce the inference performance. The main results
The creation of a custom dataset was necessary for altitude evaluation as no other datasets were available. To fulfill the requirements, the footage was captured using a small UAV hovering above a roundabout near the University of Cyprus campus, where several structures and buildings with solar panels and water tanks were visible at varying altitudes. The data were captured during a sunny day, ensuring bright and shadowless images. Images were extracted from the footage, and all data were annotated with a single class labeled as 'Car'. The dataset covered altitudes ranging from 50 to 500 meters with a 50-meter step, and all images were kept at their original high resolution of 3840x2160, presenting challenges for object detection. The data were split into 3 sets for training, validation, and testing, with the number of vehicles increasing as altitude increased, which was expected due to the larger field of view of the camera. Each folder consists of an aerial vehicle dataset captured at the corresponding altitude. For each altitude, the dataset annotations are generated in YOLO, COCO, and VOC formats. The dataset consists of the following images and detection objects:
Data
Subset
Images
Cars
50m
Train
130
269
50m
Test
32
66
50m
Valid
33
73
100m
Train
246
937
100m
Test
61
226
100m
Valid
62
250
150m
Train
244
1691
150m
Test
61
453
150m
Valid
61
426
200m
Train
246
1753
200m
Test
61
445
200m
Valid
62
424
250m
Train
245
3326
250m
Test
61
821
250m
Valid
61
823
300m
Train
246
6250
300m
Test
61
1553
300m
Valid
62
1585
350m
Train
246
10741
350m
Test
61
2591
350m
Valid
62
2687
400m
Train
245
20072
400m
Test
61
4974
400m
Valid
61
4924
450m
Train
246
31794
450m
Test
61
7887
450m
Valid
61
7880
500m
Train
270
49782
500m
Test
67
12426
500m
Valid
68
12541
mix_alt
Train
2364
126615
mix_alt
Test
587
31442
mix_alt
Valid
593
31613
It is advised to further enhance the dataset so that random augmentations are probabilistically applied to each image prior to adding it to the batch for training. Specifically, there are a number of possible transformations such as geometric (rotations, translations, horizontal axis mirroring, cropping, and zooming), as well as image manipulations (illumination changes, color shifting, blurring, sharpening, and shadowing).
MIT Licensehttps://opensource.org/licenses/MIT
License information was derived automatically
This dataset contains 74 images of aerial maritime photographs taken with via a Mavic Air 2 drone and 1,151 bounding boxes, consisting of docks, boats, lifts, jetskis, and cars. This is a multi class problem. This is an aerial object detection dataset. This is a maritime object detection dataset.
The drone was flown at 400 ft. No drones were harmed in the making of this dataset.
This dataset was collected and annotated by the Roboflow team, released with MIT license.
https://i.imgur.com/9ZYLQSO.jpg" alt="Image example">
This dataset is a great starter dataset for building an aerial object detection model with your drone.
Fork or download this dataset and follow our How to train state of the art object detector YOLOv4 for more. Stay tuned for particular tutorials on how to teach your UAV drone how to see and comprable airplane imagery and airplane footage.
See here for how to use the CVAT annotation tool that was used to create this dataset.
Roboflow makes managing, preprocessing, augmenting, and versioning datasets for computer vision seamless. :fa-spacer: Developers reduce 50% of their boilerplate code when using Roboflow's workflow, save training time, and increase model reproducibility. :fa-spacer:
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
This database contains 4976 planetary images of boulder fields located on Earth, Mars and Moon. The data was collected during the BOULDERING Marie Skłodowska-Curie Global fellowship between October 2021 and 2024. The data was already splitted into train, validation and test datasets, but feel free to re-organize the labels at your convenience.
For each image, all of the boulder outlines within the image were carefully mapped in QGIS. More information about the labelling procedure can be found in the following manuscript (https://agupubs.onlinelibrary.wiley.com/doi/full/10.1029/2023JE008013). This dataset differs from the previous dataset included along with the manuscript https://zenodo.org/records/8171052, as it contains more mapped images, especially of boulder populations around young impact structures on the Moon (cold spots). In addition, the boulder outlines were also pre-processed so that it can be ingested directly in YOLOv8.
A description of what is what is given in the README.txt file (in addition in how to load the custom datasets in Detectron2 and YOLO). Most of the other files are mostly self-explanatory. Please see previous dataset or manuscript for more information. If you want to have more information about specific lunar and martian planetary images, the IDs of the images are still available in the name of the file. Use this ID to find more information (e.g., M121118602_00875_image.png, ID M121118602 ca be used on https://pilot.wr.usgs.gov/). I will also upload the raw data from which this pre-processed dataset was generated (see https://zenodo.org/records/14250970).
Thanks to this database, you can easily train a Detectron2 Mask R-CNN or YOLO instance segmentation models to automatically detect boulders.
How to cite:
Please refer to the "how to cite" section of the readme file of https://github.com/astroNils/YOLOv8-BeyondEarth.
Structure:
. └── boulder2024/ ├── jupyter-notebooks/ │ └── REGISTERING_BOULDER_DATASET_IN_DETECTRON2.ipynb ├── test/ │ └── images/ │ ├── _image.png │ ├── ... │ └── labels/ │ ├── _image.txt │ ├── ... ├── train/ │ └── images/ │ ├── _image.png │ ├── ... │ └── labels/ │ ├── _image.txt │ ├── ... ├── validation/ │ └── images/ │ ├── _image.png │ ├── ... │ └── labels/ │ ├── _image.txt │ ├── ... ├── detectron2_inst_seg_boulder_dataset.json ├── README.txt ├── yolo_inst_seg_boulder_dataset.yaml
detectron2_inst_seg_boulder_dataset.json
is a json file containing the masks as expected by Detectron2 (see https://detectron2.readthedocs.io/en/latest/tutorials/datasets.html for more information on the format). In order to use this custom dataset, you need to register the dataset before using it in the training. There is an example how to do that in the jupyter-notebooks folder. You need to have detectron2, and all of its depedencies installed.
yolo_inst_seg_boulder_dataset.yaml
can be used as it is, however you need to update the paths in the .yaml file, to the test, train and validation folders. More information about the YOLO format can be found here (https://docs.ultralytics.com/datasets/segment/).
Attribution-NonCommercial 4.0 (CC BY-NC 4.0)https://creativecommons.org/licenses/by-nc/4.0/
License information was derived automatically
Dataset is available under the Attribution-NonCommercial 4.0 International (CC BY-NC 4.0) license.
=======================
Summary
=======================
This dataset contains data showing a group of one or more drones (drone swarm) roaming around closed 3D space.
The purpose of this dataset is to train and validate the position tracking of drones on images. This can be done using 3D coordinates of each drone or using segmentation masks for machine learning related approach.
The first part of the dataset is split across ten sequences (AirSim 1-10) which were generated using Unreal Engine with Microsoft AirSim plugin. Each sequence is slightly modified to introduce more diversity in the dataset.
The second part of the dataset (HML 1-3) is based on recordings performed in the Human Motion Laboratory located inside the Research and Development Center of the Polish-Japanese Academy of Information Technology.
The simulation sequences used a drone model DJI Mavic 2 Pro by MattMaksymowicz, CC Attribution, https://sketchfab.com/3d-models/dji-mavic-2-pro-3e5b8566dbe24f4ba65473179650abd1.
All documents and papers that use the dataset must acknowledge the use of the dataset by including a citation of the following paper:
Lindenheim-Locher W, Świtoński A, Krzeszowski T, Paleta G, Hasiec P, Josiński H, Paszkuta M, Wojciechowski K, Rosner J. YOLOv5 Drone Detection Using Multimodal Data Registered by the Vicon System. Sensors. 2023; 23(14):6396. https://doi.org/10.3390/s23146396
=======================
Dataset sequences
=======================
AirSim 1:
4 drones based on DJI Mavic 2 Pro.
Length: 20 seconds
Framerate: 25 FPS
Number of cameras: 8
AirSim 2:
10 drones based on custom model. All drones move the same path however, each one has some fixed offset applied in all axes.
Length: 10 seconds
Framerate: 25 FPS
Number of cameras: 8
AirSim 3:
10 drones based on custom model. All drones move the same path however, each one has some fixed offset applied in all axes.
Length: 10 seconds
Framerate: 25 FPS
Number of cameras: 8
AirSim 4:
8 drones based on custom model. All drones move the same path however, each one has some fixed offset applied in all axes.
Length: 20 seconds
Framerate: 25 FPS
Number of cameras: 8
AirSim 5:
6 drones based on custom model. All drones move the same path however, each one has some fixed offset applied in all axes.
Length: 20 seconds
Framerate: 25 FPS
Number of cameras: 8
AirSim 6:
8 drones based on custom model. All drones move the same path however, each one has some fixed offset applied in all axes.
Length: 20 seconds
Framerate: 25 FPS
Number of cameras: 8
AirSim 7:
8 drones based on three types of custom model.
Length: 20 seconds
Framerate: 25 FPS
Number of cameras: 8
AirSim 8:
8 drones based on three types of custom model.
Length: 20 seconds
Framerate: 25 FPS
Number of cameras: 8
AirSim 9:
8 drones based on three types of custom model.
Length: 20 seconds
Framerate: 25 FPS
Number of cameras: 8
AirSim 10:
2 drones based on a custom model.
Length: 20 seconds
Framerate: 25 FPS
Number of cameras: 8
HML 1:
1 real drone (DJI Mavic 2). It is controlled in real-time.
Length: 78 seconds
Framerate: 25 FPS
Number of cameras: 4
HML 2:
1 real drone. It is controlled in real-time.
Length: 60 seconds
Framerate: 25 FPS
Number of cameras: 4
HML 3:
1 real drone. It is controlled in real-time.
Length: 90 seconds
Framerate: 25 FPS
Number of cameras: 4
=======================
Additional information
=======================
AirSim 1-10 sets consist of two types of image data: RGB images and masks
RGB images are compressed to avi videos to save space. The name of the AVI filename contains the camera name.
The name of the mask directory is based on the drone whose masks it contains
Naming pattern of mask files:
- a camera that was used to capture
- frame ID
for example, cam_1_230.jpeg was taken by camera 1 and is 230 frames of the sequence.
HML 1-3 sets consist of two additional files:
- c3d motion data file
- xcp calibration file
=======================
Further information
=======================
For any questions, comments or other issues please contact Tomasz Krzeszowski
Attribution-NonCommercial-ShareAlike 4.0 (CC BY-NC-SA 4.0)https://creativecommons.org/licenses/by-nc-sa/4.0/
License information was derived automatically
https://www.googleapis.com/download/storage/v1/b/kaggle-user-content/o/inbox%2F23345571%2F4471e4ade50676d782d4787f77aa08ad%2F1000_F_256252609_6WIHRGbpzSaVQwioubxwgXdSJTNONNcK.jpg?generation=1739209341333909&alt=media" alt="">
This dataset contains 2,700 images focused on detecting potholes, cracks, and open manholes on roads. It has been augmented to enhance the variety and robustness of the data. The images are organized into training and validation sets, with three distinct categories:
Included in the Dataset: - Bounding Box Annotations in YOLO Format (.txt files) - Format: YOLOv8 & YOLO11 compatible - Purpose: Ready for training YOLO-based object detection models
Folder Structure Organized into:
Dual Format Support
Use Cases Targeted
Here's a clear breakdown of the folder structure:
https://www.googleapis.com/download/storage/v1/b/kaggle-user-content/o/inbox%2F23345571%2F023b40c98bf858c58394d6ed2393bfc3%2FScreenshot%202025-05-01%20202438.png?generation=1746109541780835&alt=media" alt="">
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
test dataset for train model on custom dataset, test
CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
License information was derived automatically
## Overview
Custom_Object_Train is a dataset for object detection tasks - it contains Ball annotations for 1,475 images.
## Getting Started
You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
## License
This dataset is available under the [Public Domain license](https://creativecommons.org/licenses/Public Domain).
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
The dataset and code used in this study are crucial for advancing the accurate detection of positive microchambers in large-field ddPCR imaging. The provided dataset includes annotated ddPCR images in YOLO format, stored in the ddpcr320/
folder. The codebase features the improved YOLOv5 model, integrating BiFPN, GhostConv, C3Ghost modules, SimAM attention mechanism, and network pruning, among other custom modifications. The train.py
and detect.py
scripts handle training and detection tasks, while dataset.ipynb
demonstrates the dataset creation and splitting processes, as well as dataset processing and augmentation. The graphical user interface, developed using PyQt5 and implemented in main_win.py
, facilitates image processing and result analysis for users. The project structure, ddpcr_yolov5
, is systematically organized, with detailed instructions provided in the README.md file.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Outdoor Obstacle Detection (OOD): is a custom dataset created to train models to detect 22 specific types of obstacles that can obstruct blind people in the way when walking in outdoor spaces. The dataset contains 10.000 images and 29.779 annotated instances and 22 classes: person, car, tree, spherical_roadblock, warning_column, waste_container, street_light, fire_hydrant, traffic_light, stop_sign, pole, bench, curb, stairs, bicycle, motorcycle, dog, bus, truck, train, bus_stop, crutch.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Goals 1- Create dataset and apply augmentations. 2- Train custom YOLO model. 3- Evaluate using still frames. 4- Evaluate usinf video.
MIT Licensehttps://opensource.org/licenses/MIT
License information was derived automatically
## Overview
DevNet Custom MV PPE Detection is a dataset for object detection tasks - it contains PPE annotations for 5,239 images.
## Getting Started
You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
## License
This dataset is available under the [MIT license](https://creativecommons.org/licenses/MIT).
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Football (soccer) player and football (soccer) ball detection dataset from Augmented Startups. * Project Type: Object Detection * Labeled/Annotated with: Bounding boxes
football
, player
This is a great starter-dataset for those wanting to test player and/or ball-tracking for football (soccer) games with the Deploy Tab, or the Deployment device and method of their choice.
Images can also be Cloned to another project to continue iterating on the project and model. World Cup, Premier League, La Liga, Major League Soccer (MLS) and/or Champions League computer vision projects, anyone?
Roboflow offers AutoML model training - Roboflow Train, and the ability to import and export up to 30 different annotation formats. Leaving you flexibility to deploy directly with a Roboflow Train model, or use Roboflow to prepare and manage datasets, and train and deploy with the custom model architecture of your choice + https://github.com/roboflow-ai/notebooks.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
The goal of this task is to train a model that can localize and classify each instance of Person and Car as accurately as possible.
from IPython.display import Markdown, display
display(Markdown("../input/Car-Person-v2-Roboflow/README.roboflow.txt"))
In this Notebook, I have processed the images with RoboFlow because in COCO formatted dataset was having different dimensions of image and Also data set was not splitted into different Format. To train a custom YOLOv7 model we need to recognize the objects in the dataset. To do so I have taken the following steps:
Image Credit - jinfagang
!git clone https://github.com/WongKinYiu/yolov7 # Downloading YOLOv7 repository and installing requirements
%cd yolov7
!pip install -qr requirements.txt
!pip install -q roboflow
!wget "https://github.com/WongKinYiu/yolov7/releases/download/v0.1/yolov7.pt"
import os
import glob
import wandb
import torch
from roboflow import Roboflow
from kaggle_secrets import UserSecretsClient
from IPython.display import Image, clear_output, display # to display images
print(f"Setup complete. Using torch {torch._version_} ({torch.cuda.get_device_properties(0).name if torch.cuda.is_available() else 'CPU'})")
https://camo.githubusercontent.com/dd842f7b0be57140e68b2ab9cb007992acd131c48284eaf6b1aca758bfea358b/68747470733a2f2f692e696d6775722e636f6d2f52557469567a482e706e67">
I will be integrating W&B for visualizations and logging artifacts and comparisons of different models!
try:
user_secrets = UserSecretsClient()
wandb_api_key = user_secrets.get_secret("wandb_api")
wandb.login(key=wandb_api_key)
anonymous = None
except:
wandb.login(anonymous='must')
print('To use your W&B account,
Go to Add-ons -> Secrets and provide your W&B access token. Use the Label name as WANDB.
Get your W&B access token from here: https://wandb.ai/authorize')
wandb.init(project="YOLOvR",name=f"7. YOLOv7-Car-Person-Custom-Run-7")
https://uploads-ssl.webflow.com/5f6bc60e665f54545a1e52a5/615627e5824c9c6195abfda9_computer-vision-cycle.png" alt="">
In order to train our custom model, we need to assemble a dataset of representative images with bounding box annotations around the objects that we want to detect. And we need our dataset to be in YOLOv7 format.
In Roboflow, We can choose between two paths:
https://raw.githubusercontent.com/Owaiskhan9654/Yolo-V7-Custom-Dataset-Train-on-Kaggle/main/Roboflow.PNG" alt="">
user_secrets = UserSecretsClient()
roboflow_api_key = user_secrets.get_secret("roboflow_api")
rf = Roboflow(api_key=roboflow_api_key)
project = rf.workspace("owais-ahmad").project("custom-yolov7-on-kaggle-on-custom-dataset-rakiq")
dataset = project.version(2).download("yolov7")
Here, I am able to pass a number of arguments: - img: define input image size - batch: determine