Usage:
python
!pip install /kaggle/input/ultralytics-whl/ultralytics-8.0.139-py3-none-any.whl
GitHub: https://github.com/ultralytics/ultralytics
PyPi: https://pypi.org/project/ultralytics/
http://www.gnu.org/licenses/agpl-3.0.htmlhttp://www.gnu.org/licenses/agpl-3.0.html
Licencia: GNU Affero General Public License v3 or later (AGPLv3+) (AGPL-3.0) Responsable: Ultralytics Etiquetasmachine-learning, deep-learning, computer-vision, ML, DL, AI, YOLO, YOLOv3, YOLOv5, YOLOv8, YOLOv9, YOLOv10, YOLO11, HUB, Ultralytics Requiere: Python >=3.8
Apache License, v2.0https://www.apache.org/licenses/LICENSE-2.0
License information was derived automatically
This dataset was created by Mahmoud Abdelshafy
Released under Apache 2.0
This is the yoloV5 model cloned from ultralytics/yolov5 for offline use.
Add this dataset to the notebook and run the following commands.
Add model weights. You can use your own or take it from here.
!mkdir /root/.config/Ultralytics/
!cp ../input/yolo-arial/Arial.ttf /root/.config/Ultralytics/Arial.ttf
You can now call torch.hub.load()
to load the yolov5 model offline. Make sure to set the parameters as
repo or dir = './path/to/local/yolov5'
model = 'custom'
source = 'local'
force_reload = True
path = './path/to/best.pt'
Using yolov5x6 weights
import torch
yolov5x6_model = torch.hub.load('../input/yolov5', 'custom', source='local', force_reload=True, path='../input/ultralyticsyolov5aweights/yolov5x6.pt')
From https://github.com/ultralytics/yolov5 2021/7/8
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Started out as a pumpkin detector to test training YOLOv5. Now suffering from extensive feature creep and probably ending up as a cat/dog/spider/pumpkin/randomobjects-detector. Or as a desaster.
The dataset does not fit https://docs.ultralytics.com/tutorials/training-tips-best-results/ well. There are no background images and the labeling is often only partial. Especially in the humans and pumpkin category where there are often lots of objects in one photo people apparently (and understandably) got bored and did not labe everything. And of course the images from the cat-category don't have the humans in it labeled since they come from a cat-identification model which ignored humans. It will need a lot of time to fixt that.
Dataset used: - Cat and Dog Data: Cat / Dog Tutorial NVIDIA Jetson https://github.com/dusty-nv/jetson-inference/blob/master/docs/pytorch-cat-dog.md © 2016-2019 NVIDIA according to bottom of linked page - Spider Data: Kaggle Animal 10 image set https://www.kaggle.com/datasets/alessiocorrado99/animals10 Animal pictures of 10 different categories taken from google images Kaggle project licensed GPL 2 - Pumpkin Data: Kaggle "Vegetable Images" https://www.researchgate.net/publication/352846889_DCNN-Based_Vegetable_Image_Classification_Using_Transfer_Learning_A_Comparative_Study https://www.kaggle.com/datasets/misrakahmed/vegetable-image-dataset Kaggle project licensed CC BY-SA 4.0 - Some pumpkin images manually copied from google image search - https://universe.roboflow.com/chess-project/chess-sample-rzbmc Provided by a Roboflow user License: CC BY 4.0 - https://universe.roboflow.com/steve-pamer-cvmbg/pumpkins-gfjw5 Provided by a Roboflow user License: CC BY 4.0 - https://universe.roboflow.com/nbduy/pumpkin-ryavl Provided by a Roboflow user License: CC BY 4.0 - https://universe.roboflow.com/homeworktest-wbx8v/cat_test-1x0bl/dataset/2 - https://universe.roboflow.com/220616nishikura/catdetector - https://universe.roboflow.com/atoany/cats-s4d4i/dataset/2 - https://universe.roboflow.com/personal-vruc2/agricultured-ioth22 - https://universe.roboflow.com/sreyoshiworkspace-radu9/pet_detection - https://universe.roboflow.com/artyom-hystt/my-dogs-lcpqe - license: Public Domain url: https://universe.roboflow.com/dolazy7-gmail-com-3vj05/sweetpumpkin/dataset/2 - https://universe.roboflow.com/tristram-dacayan/social-distancing-g4pbu - https://universe.roboflow.com/fyp-3edkl/social-distancing-2ygx5 License MIT - Spiders: https://universe.roboflow.com/lucas-lins-souza/animals-train-yruka
Currently I can't guarantee it's all correctly licenced. Checks are in progress. Inform me if you see one of your pictures and want it to be removed!
This dataset is cloned from https://github.com/ultralytics/yolov5
As of the dataset creation date, the original Yolo V5 repo on Github is updated frequently, and this version can be dated very soon
Thanks Ultralytics team, especially the main author Glenn Jocher for creating this powerful and easy-to-use library.
GPL-3.0
How to download a Yolo model. This downloads the latest Yolo version: https://www.kaggle.com/code/vbookshelf/exp-90-covid-downloading-yolo-v5/notebook?scriptVersionId=66895104
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Accident Detection Model is made using YOLOv8, Google Collab, Python, Roboflow, Deep Learning, OpenCV, Machine Learning, Artificial Intelligence. It can detect an accident on any accident by live camera, image or video provided. This model is trained on a dataset of 3200+ images, These images were annotated on roboflow.
https://user-images.githubusercontent.com/78155393/233774342-287492bb-26c1-4acf-bc2c-9462e97a03ca.png" alt="Survey">
Apache License, v2.0https://www.apache.org/licenses/LICENSE-2.0
License information was derived automatically
This dataset was created by Fathy sahlool
Released under Apache 2.0
For usage of Yolov5 without connecting to Internet.
Yolov5-master and four model weight files "yolov5s.pt", "yolov5m.pt", "yolov5l.pt", "yolov5x.pt".
Downloaded from https://github.com/ultralytics/yolov5
These are all YOLO weights pretrained on the MS COCO dataset. The yolov3-spp-ultralytics.pt
and yolov3-tiny.pt
are taken from Ultralytics YOLOv3. The yolov4.pt
is converted from yolov4.weights
by AlexeyAB/darknet into PyTorch .pt
format.
MIT Licensehttps://opensource.org/licenses/MIT
License information was derived automatically
FSOCO dataset split into train (80%), validation (10%), and test (10%) set. Ready for Ultralytics YOLO training.
MIT Licensehttps://opensource.org/licenses/MIT
License information was derived automatically
This dataset is a repackaged version of the original MP-IDB (The Malaria Parasite Image Database for Image Processing and Analysis), formatted for Ultralytics YOLO (You Only Look Once) instance segmentation annotation. The goal of this release is to make it easier for researchers and practitioners to apply state-of-the-art instance segmentation or object detection techniques to malaria cell detection and classification tasks.
⚠️ This dataset is a derivative work. All original images and annotations belong to the original MP-IDB authors. This version only converts them into Ultralytics YOLO-compatible format.
The original MP-IDB dataset was created and released by Andrea Loddo, Cecilia Di Ruberto, Michel Kocher, and Guy Prod’Hom, and is described in the following publication:
MP-IDB: The Malaria Parasite Image Database for Image Processing and Analysis
In Processing and Analysis of Biomedical Information, Springer, 2019.
DOI: 10.1007/978-3-030-13835-6_7
The dataset includes annotated microscopic blood smear images of four malaria species:
Each image contains cells in one or more of the following parasite life stages, indicated in filenames:
Expert pathologists provided the ground truth for each image.
This version of the dataset includes:
.txt
files)This reformatting is designed to save time for those building instance segmentation or object detection models for medical imaging and accelerate prototyping using YOLO and the Ultralytics Package.
The original MP-IDB dataset is released under the MIT License by Andrea Loddo and contributors. Please make sure to cite the original work if you use this dataset in your own research or application:
The dataset contains sample images from the Open Images Dataset v7. This dataset only contains images for the category 'ducks' and is arranged to fine-tune the YOLOv8 image segmentation models.
The dataset contains two main directors, i.e., images and labels. These directories further contain 'train' and 'val' directories. As the names suggest, these directories contain images and labels for the training and validation of image segmentation models.
Training Images: 400 Validation Images: 50
Class/es: Duck
The dataset also contains a config.yaml
file. This file contains paths for relevant directories that YOLOv8 needs to load datasets
https://creativecommons.org/publicdomain/zero/1.0/https://creativecommons.org/publicdomain/zero/1.0/
Almost 12000 images used in Google Recaptcha V2 collected by category more than 500 of which with manual markup for training object detection model such as YOLO.
If you find this dataset useful, please leave an upvote, that motivates me to collect such datasets✋
Feel free to using this data for your commercial or educational goals. P.S. https://github.com/Artistrazh/recaptcha_v2_solver is my project for solving Google Recaptcha V2 using yolov3, BLIP and this dataset.
https://github.com/ultralytics/yolov5
Only clone, not modify because the license.
Configure for one class as Wheat New configure with yolov5x.yaml
Using:
!cp -r ../input/all-in-one-yolov5-wheat/yolov5/* .
!pip install -r requirements.txt
if issue with torch version and torchvission, check version cuda and comeback the old version torch, example cuda 10.1 :
!pip install torch==1.5.1+cu101 torchvision==0.6.1+cu101 -f https://download.pytorch.org/whl/torch_stable.html
!git clone https://github.com/NVIDIA/apex && cd apex && pip install -v --no-cache-dir --global-option="--cpp_ext" --global-option="--cuda_ext" . --user && cd .. && rm -rf apex
!python train.py --img 1024 --batch 2 --epochs 12 --data ../input/all-in-one-yolov5-wheat/wheat0.yaml --cfg ../input/all-in-one-yolov5-wheat/yolov5x.yaml --name yolov5_fold0 --weights "yolov5x.pt" --adam
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
One of my passions is playing board games with my friends. However one of them lives abroad and so we like to stream the game when playing with him. However instead of just having a normal stream, I wanted to show some additional information about the monsters that are on the game board. This originated in a fun project to train CNNs in order to detect these monsters.
To have enough training data, I made a little project in UE4 to generate these training images. For each image there is a mask for every monster that appears in it. The dataset also includes annotations for the train images in the COCO format (annotations.json
) and labes for the bounding box in Darknet format in the folder labels
.
There is a training and validation subset for the images
, labels
and masks
folders. The structure is as follows: for the first training image containing an earth_demon
and harrower_infester
:
images/train/image_1.png
labels/train/label_1.png
. This file contains two lines. One line for each monster. A line is constructed as follows: class_id center_x center_y width height
. Note that the position and dimensions are relative to the image width and height.masks/train
. One is named image_1_mask_0_harrower_infester.png
and the other image_1_mask_1_earth_demon.png
.The code for generating this dataset and training a MaskRCNN and YoloV5 model can be found at https://github.com/ericdepotter/Gloomhaven-Monster-Recognizer.
I took pictures for the images of the monsters myself. The images of the game tiles I obtained from this collection of Gloomhaven assets.
This is a classic object detection or object segmentation problem.
Not seeing a result you expected?
Learn how you can add new datasets to our index.
Usage:
python
!pip install /kaggle/input/ultralytics-whl/ultralytics-8.0.139-py3-none-any.whl
GitHub: https://github.com/ultralytics/ultralytics
PyPi: https://pypi.org/project/ultralytics/