27 datasets found
  1. T

    imagenet2012

    • tensorflow.org
    Updated Jun 1, 2024
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    (2024). imagenet2012 [Dataset]. https://www.tensorflow.org/datasets/catalog/imagenet2012
    Explore at:
    Dataset updated
    Jun 1, 2024
    Description

    ILSVRC 2012, commonly known as 'ImageNet' is an image dataset organized according to the WordNet hierarchy. Each meaningful concept in WordNet, possibly described by multiple words or word phrases, is called a "synonym set" or "synset". There are more than 100,000 synsets in WordNet, majority of them are nouns (80,000+). In ImageNet, we aim to provide on average 1000 images to illustrate each synset. Images of each concept are quality-controlled and human-annotated. In its completion, we hope ImageNet will offer tens of millions of cleanly sorted images for most of the concepts in the WordNet hierarchy.

    The test split contains 100K images but no labels because no labels have been publicly released. We provide support for the test split from 2012 with the minor patch released on October 10, 2019. In order to manually download this data, a user must perform the following operations:

    1. Download the 2012 test split available here.
    2. Download the October 10, 2019 patch. There is a Google Drive link to the patch provided on the same page.
    3. Combine the two tar-balls, manually overwriting any images in the original archive with images from the patch. According to the instructions on image-net.org, this procedure overwrites just a few images.

    The resulting tar-ball may then be processed by TFDS.

    To assess the accuracy of a model on the ImageNet test split, one must run inference on all images in the split, export those results to a text file that must be uploaded to the ImageNet evaluation server. The maintainers of the ImageNet evaluation server permits a single user to submit up to 2 submissions per week in order to prevent overfitting.

    To evaluate the accuracy on the test split, one must first create an account at image-net.org. This account must be approved by the site administrator. After the account is created, one can submit the results to the test server at https://image-net.org/challenges/LSVRC/eval_server.php The submission consists of several ASCII text files corresponding to multiple tasks. The task of interest is "Classification submission (top-5 cls error)". A sample of an exported text file looks like the following:

    771 778 794 387 650
    363 691 764 923 427
    737 369 430 531 124
    755 930 755 59 168
    

    The export format is described in full in "readme.txt" within the 2013 development kit available here: https://image-net.org/data/ILSVRC/2013/ILSVRC2013_devkit.tgz Please see the section entitled "3.3 CLS-LOC submission format". Briefly, the format of the text file is 100,000 lines corresponding to each image in the test split. Each line of integers correspond to the rank-ordered, top 5 predictions for each test image. The integers are 1-indexed corresponding to the line number in the corresponding labels file. See labels.txt.

    To use this dataset:

    import tensorflow_datasets as tfds
    
    ds = tfds.load('imagenet2012', split='train')
    for ex in ds.take(4):
     print(ex)
    

    See the guide for more informations on tensorflow_datasets.

    https://storage.googleapis.com/tfds-data/visualization/fig/imagenet2012-5.1.0.png" alt="Visualization" width="500px">

  2. T

    imagenet2012_real

    • tensorflow.org
    Updated Jun 1, 2024
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    (2024). imagenet2012_real [Dataset]. https://www.tensorflow.org/datasets/catalog/imagenet2012_real
    Explore at:
    Dataset updated
    Jun 1, 2024
    Description

    This dataset contains ILSVRC-2012 (ImageNet) validation images augmented with a new set of "Re-Assessed" (ReaL) labels from the "Are we done with ImageNet" paper, see https://arxiv.org/abs/2006.07159. These labels are collected using the enhanced protocol, resulting in multi-label and more accurate annotations.

    Important note: about 3500 examples contain no label, these should be excluded from the averaging when computing the accuracy. One possible way of doing this is with the following NumPy code:

    is_correct = [pred in real_labels[i] for i, pred in enumerate(predictions) if real_labels[i]]
    real_accuracy = np.mean(is_correct)
    

    To use this dataset:

    import tensorflow_datasets as tfds
    
    ds = tfds.load('imagenet2012_real', split='train')
    for ex in ds.take(4):
     print(ex)
    

    See the guide for more informations on tensorflow_datasets.

    https://storage.googleapis.com/tfds-data/visualization/fig/imagenet2012_real-1.0.0.png" alt="Visualization" width="500px">

  3. T

    imagenet_a

    • tensorflow.org
    Updated Jun 1, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    (2024). imagenet_a [Dataset]. https://www.tensorflow.org/datasets/catalog/imagenet_a
    Explore at:
    Dataset updated
    Jun 1, 2024
    Description

    ImageNet-A is a set of images labelled with ImageNet labels that were obtained by collecting new data and keeping only those images that ResNet-50 models fail to correctly classify. For more details please refer to the paper.

    The label space is the same as that of ImageNet2012. Each example is represented as a dictionary with the following keys:

    • 'image': The image, a (H, W, 3)-tensor.
    • 'label': An integer in the range [0, 1000).
    • 'file_name': A unique sting identifying the example within the dataset.

    To use this dataset:

    import tensorflow_datasets as tfds
    
    ds = tfds.load('imagenet_a', split='train')
    for ex in ds.take(4):
     print(ex)
    

    See the guide for more informations on tensorflow_datasets.

    https://storage.googleapis.com/tfds-data/visualization/fig/imagenet_a-0.1.0.png" alt="Visualization" width="500px">

  4. T

    imagenette

    • tensorflow.org
    • opendatalab.com
    • +1more
    Updated Jun 1, 2024
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    (2024). imagenette [Dataset]. https://www.tensorflow.org/datasets/catalog/imagenette
    Explore at:
    Dataset updated
    Jun 1, 2024
    Description

    Imagenette is a subset of 10 easily classified classes from the Imagenet dataset. It was originally prepared by Jeremy Howard of FastAI. The objective behind putting together a small version of the Imagenet dataset was mainly because running new ideas/algorithms/experiments on the whole Imagenet take a lot of time.

    This version of the dataset allows researchers/practitioners to quickly try out ideas and share with others. The dataset comes in three variants:

    • Full size
    • 320 px
    • 160 px

    Note: The v2 config correspond to the new 70/30 train/valid split (released in Dec 6 2019).

    To use this dataset:

    import tensorflow_datasets as tfds
    
    ds = tfds.load('imagenette', split='train')
    for ex in ds.take(4):
     print(ex)
    

    See the guide for more informations on tensorflow_datasets.

    https://storage.googleapis.com/tfds-data/visualization/fig/imagenette-full-size-v2-1.0.0.png" alt="Visualization" width="500px">

  5. T

    imagenet_resized

    • tensorflow.org
    Updated Jun 1, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    (2024). imagenet_resized [Dataset]. https://www.tensorflow.org/datasets/catalog/imagenet_resized
    Explore at:
    Dataset updated
    Jun 1, 2024
    Description

    This dataset consists of the ImageNet dataset resized to fixed size. The images here are the ones provided by Chrabaszcz et. al. using the box resize method.

    For downsampled ImageNet for unsupervised learning see downsampled_imagenet.

    WARNING: The integer labels used are defined by the authors and do not match those from the other ImageNet datasets provided by Tensorflow datasets. See the original label list, and the labels used by this dataset. Additionally, the original authors 1 index there labels which we convert to 0 indexed by subtracting one.

    To use this dataset:

    import tensorflow_datasets as tfds
    
    ds = tfds.load('imagenet_resized', split='train')
    for ex in ds.take(4):
     print(ex)
    

    See the guide for more informations on tensorflow_datasets.

    https://storage.googleapis.com/tfds-data/visualization/fig/imagenet_resized-8x8-0.1.0.png" alt="Visualization" width="500px">

  6. T

    imagenet_r

    • tensorflow.org
    Updated Jun 1, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    (2024). imagenet_r [Dataset]. https://www.tensorflow.org/datasets/catalog/imagenet_r
    Explore at:
    Dataset updated
    Jun 1, 2024
    Description

    ImageNet-R is a set of images labelled with ImageNet labels that were obtained by collecting art, cartoons, deviantart, graffiti, embroidery, graphics, origami, paintings, patterns, plastic objects, plush objects, sculptures, sketches, tattoos, toys, and video game renditions of ImageNet classes. ImageNet-R has renditions of 200 ImageNet classes resulting in 30,000 images. by collecting new data and keeping only those images that ResNet-50 models fail to correctly classify. For more details please refer to the paper.

    The label space is the same as that of ImageNet2012. Each example is represented as a dictionary with the following keys:

    • 'image': The image, a (H, W, 3)-tensor.
    • 'label': An integer in the range [0, 1000).
    • 'file_name': A unique sting identifying the example within the dataset.

    To use this dataset:

    import tensorflow_datasets as tfds
    
    ds = tfds.load('imagenet_r', split='train')
    for ex in ds.take(4):
     print(ex)
    

    See the guide for more informations on tensorflow_datasets.

    https://storage.googleapis.com/tfds-data/visualization/fig/imagenet_r-0.2.0.png" alt="Visualization" width="500px">

  7. Hymenoptera dataset

    • kaggle.com
    Updated Jul 11, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Tensorflow Notebooks (2022). Hymenoptera dataset [Dataset]. https://www.kaggle.com/datasets/tensorflownotebooks/hymenoptera-dataset
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Jul 11, 2022
    Dataset provided by
    Kagglehttp://kaggle.com/
    Authors
    Tensorflow Notebooks
    Description

    This dataset is used in the Pytorch example Transfer Learning for Computer Vision Tutorial

  8. h

    imagenet_sketch

    • huggingface.co
    • opendatalab.com
    • +1more
    Updated May 25, 2024
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Songwei Ge (2024). imagenet_sketch [Dataset]. https://huggingface.co/datasets/songweig/imagenet_sketch
    Explore at:
    Dataset updated
    May 25, 2024
    Authors
    Songwei Ge
    License

    https://choosealicense.com/licenses/unknown/https://choosealicense.com/licenses/unknown/

    Description

    ImageNet-Sketch data set consists of 50000 images, 50 images for each of the 1000 ImageNet classes. We construct the data set with Google Image queries "sketch of _", where _ is the standard class name. We only search within the "black and white" color scheme. We initially query 100 images for every class, and then manually clean the pulled images by deleting the irrelevant images and images that are for similar but different classes. For some classes, there are less than 50 images after manually cleaning, and then we augment the data set by flipping and rotating the images.

  9. T

    imagenet_lt

    • tensorflow.org
    Updated Dec 10, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    (2022). imagenet_lt [Dataset]. https://www.tensorflow.org/datasets/catalog/imagenet_lt
    Explore at:
    Dataset updated
    Dec 10, 2022
    Description

    ImageNet-LT is a subset of original ImageNet ILSVRC 2012 dataset. The training set is subsampled such that the number of images per class follows a long-tailed distribution. The class with the maximum number of images contains 1,280 examples, whereas the class with the minumum number of images contains only 5 examples. The dataset also has a balanced validation set, which is also a subset of the ImageNet ILSVRC 2012 training set and contains 20 images per class. The test set of this dataset is the same as the validation set of the original ImageNet ILSVRC 2012 dataset.

    The original ImageNet ILSVRC 2012 dataset must be downloaded manually, and its path should be set with --manual_dir in order to generate this dataset.

    To use this dataset:

    import tensorflow_datasets as tfds
    
    ds = tfds.load('imagenet_lt', split='train')
    for ex in ds.take(4):
     print(ex)
    

    See the guide for more informations on tensorflow_datasets.

    https://storage.googleapis.com/tfds-data/visualization/fig/imagenet_lt-1.0.0.png" alt="Visualization" width="500px">

  10. h

    tiny_imagenet_keras

    • huggingface.co
    Updated Dec 17, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    AbstractPhila (2024). tiny_imagenet_keras [Dataset]. https://huggingface.co/datasets/AbstractPhil/tiny_imagenet_keras
    Explore at:
    Dataset updated
    Dec 17, 2024
    Authors
    AbstractPhila
    License

    MIT Licensehttps://opensource.org/licenses/MIT
    License information was derived automatically

    Description

    They are tensorflow shapes and labels. (None, 32, 32, 3), (None, 1) Training is 100k images Verification is 10k images. https://github.com/tensorflow/datasets Use this instead. I think the labels are int64. You can check and verify the integrity if you want, and I'd advise it. They are a direct copy of tiny imagenet 200 classes, the training and validation images prepared and separated already. I've found better sources of them, so I wouldn't advise using this. It is available to all though.… See the full description on the dataset page: https://huggingface.co/datasets/AbstractPhil/tiny_imagenet_keras.

  11. T

    imagenet_pi

    • tensorflow.org
    Updated Apr 6, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    (2023). imagenet_pi [Dataset]. https://www.tensorflow.org/datasets/catalog/imagenet_pi
    Explore at:
    Dataset updated
    Apr 6, 2023
    Description

    ImageNet-PI is a relabelled version of the standard ILSVRC2012 ImageNet dataset in which the labels are provided by a collection of 16 deep neural networks with different architectures pre-trained on the standard ILSVRC2012. Specifically, the pre-trained models are downloaded from tf.keras.applications.

    In addition to the new labels, ImageNet-PI also provides meta-data about the annotation process in the form of confidences of the models on their labels and additional information about each model.

    For more information see: ImageNet-PI

    To use this dataset:

    import tensorflow_datasets as tfds
    
    ds = tfds.load('imagenet_pi', split='train')
    for ex in ds.take(4):
     print(ex)
    

    See the guide for more informations on tensorflow_datasets.

  12. T

    imagenet2012_corrupted

    • tensorflow.org
    Updated Jun 1, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    (2024). imagenet2012_corrupted [Dataset]. https://www.tensorflow.org/datasets/catalog/imagenet2012_corrupted
    Explore at:
    Dataset updated
    Jun 1, 2024
    Description

    Imagenet2012Corrupted is a dataset generated by adding common corruptions to the images in the ImageNet dataset. In the original paper, there are 15 + 4 different corruptions, and each has 5 levels of severity. We also implement the 4 extra corruptions gaussian blur, saturate, spatter, and speckle noise. The randomness is fixed so that regeneration is deterministic.

    To use this dataset:

    import tensorflow_datasets as tfds
    
    ds = tfds.load('imagenet2012_corrupted', split='train')
    for ex in ds.take(4):
     print(ex)
    

    See the guide for more informations on tensorflow_datasets.

    https://storage.googleapis.com/tfds-data/visualization/fig/imagenet2012_corrupted-gaussian_noise_1-3.1.0.png" alt="Visualization" width="500px">

  13. Z

    DeepWeeds: A Multiclass Weed Species Image Dataset for Deep Learning

    • data.niaid.nih.gov
    Updated May 16, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Philippa, Bronson (2023). DeepWeeds: A Multiclass Weed Species Image Dataset for Deep Learning [Dataset]. https://data.niaid.nih.gov/resources?id=zenodo_7939059
    Explore at:
    Dataset updated
    May 16, 2023
    Dataset provided by
    Calvert, Brendan
    Rahimi Azghadi, Mostafa
    Kenny, Owen
    Philippa, Bronson
    Konovalov, Dimitriv A.
    Olsen, Alex
    Johns, Jamie
    Wood, Jake C.
    Banks, Wesley
    Ridd, Peter
    Girgenti, Benjamin
    White, Ronald D.
    Whinney, James
    License

    http://www.apache.org/licenses/LICENSE-2.0http://www.apache.org/licenses/LICENSE-2.0

    Description

    DeepWeeds: A Multiclass Weed Species Image Dataset for Deep Learning

    This repository makes available the source code and public dataset for the work, "DeepWeeds: A Multiclass Weed Species Image Dataset for Deep Learning", published with open access by Scientific Reports: https://www.nature.com/articles/s41598-018-38343-3. The DeepWeeds dataset consists of 17,509 images capturing eight different weed species native to Australia in situ with neighbouring flora. In our work, the dataset was classified to an average accuracy of 95.7% with the ResNet50 deep convolutional neural network.

    The source code, images and annotations are licensed under CC BY 4.0 license. The contents of this repository are released under an Apache 2 license.

    Download the dataset images and our trained models

    images.zip (468 MB)

    models.zip (477 MB)

    Due to the size of the images and models they are hosted outside of the Github repository. The images and models must be downloaded into directories named "images" and "models", respectively, at the root of the repository. If you execute the python script (deepweeds.py), as instructed below, this step will be performed for you automatically.

    TensorFlow Datasets

    Alternatively, you can access the DeepWeeds dataset with TensorFlow Datasets, TensorFlow's official collection of ready-to-use datasets. DeepWeeds was officially added to the TensorFlow Datasets catalog in August 2019.

    Weeds and locations

    The selected weed species are local to pastoral grasslands across the state of Queensland. They include: "Chinee apple", "Snake weed", "Lantana", "Prickly acacia", "Siam weed", "Parthenium", "Rubber vine" and "Parkinsonia". The images were collected from weed infestations at the following sites across Queensland: "Black River", "Charters Towers", "Cluden", "Douglas", "Hervey Range", "Kelso", "McKinlay" and "Paluma". The table and figure below break down the dataset by weed, location and geographical distribution.

    Data organization

    Images are assigned unique filenames that include the date/time the image was photographed and an ID number for the instrument which produced the image. The format is like so: YYYYMMDD-HHMMSS-ID, where the ID is simply an integer from 0 to 3. The unique filenames are strings of 17 characters, such as 20170320-093423-1.

    labels

    The labels.csv file assigns species labels to each image. It is a comma separated text file in the format:

    Filename,Label,Species ... 20170207-154924-0,jpg,7,Snake weed 20170610-123859-1.jpg,1,Lantana 20180119-105722-1.jpg,8,Negative ...

    Note: The specific label subsets of training (60%), validation (20%) and testing (20%) for the five-fold cross validation used in the paper are also provided here as CSV files in the same format as "labels.csv".

    models

    We provide the most successful ResNet50 and InceptionV3 models saved in Keras' hdf5 model format. The ResNet50 model, which provided the best results, has also been converted to UFF format in order to construct a TensorRT inference engine.

    resnet.hdf5 inception.hdf5 resnet.uff

    deepweeds.py

    This python script trains and evaluates Keras' base implementation of ResNet50 and InceptionV3 on the DeepWeeds dataset, pre-trained with ImageNet weights. The performance of the networks are cross validated for 5 folds. The final classification accuracy is taken to be the average across the five folds. Similarly, the final confusion matrix from the associated paper aggregates across the five independent folds. The script also provides the ability to measure the inference speeds within the TensorFlow environment.

    The script can be executed to carry out these computations using the following commands.

    To train and evaluate the ResNet50 model with five-fold cross validation, use python3 deepweeds.py cross_validate --model resnet.

    To train and evaluate the InceptionV3 model with five-fold cross validation, use python3 deepweeds.py cross_validate --model inception.

    To measure inference times for the ResNet50 model, use python3 deepweeds.py inference --model models/resnet.hdf5.

    To measure inference times for the InceptionV3 model, use python3 deepweeds.py inference --model models/inception.hdf5.

    Dependencies

    The required Python packages to execute deepweeds.py are listed in requirements.txt.

    tensorrt

    This folder includes C++ source code for creating and executing a ResNet50 TensorRT inference engine on an NVIDIA Jetson TX2 platform. To build and run on your Jetson TX2, execute the following commands:

    cd tensorrt/src make -j4 cd ../bin ./resnet_inference

    Citations

    If you use the DeepWeeds dataset in your work, please cite it as:

    IEEE style citation: “A. Olsen, D. A. Konovalov, B. Philippa, P. Ridd, J. C. Wood, J. Johns, W. Banks, B. Girgenti, O. Kenny, J. Whinney, B. Calvert, M. Rahimi Azghadi, and R. D. White, “DeepWeeds: A Multiclass Weed Species Image Dataset for Deep Learning,” Scientific Reports, vol. 9, no. 2058, 2 2019. [Online]. Available: https://doi.org/10.1038/s41598-018-38343-3

    BibTeX

    @article{DeepWeeds2019, author = {Alex Olsen and Dmitry A. Konovalov and Bronson Philippa and Peter Ridd and Jake C. Wood and Jamie Johns and Wesley Banks and Benjamin Girgenti and Owen Kenny and James Whinney and Brendan Calvert and Mostafa {Rahimi Azghadi} and Ronald D. White}, title = {{DeepWeeds: A Multiclass Weed Species Image Dataset for Deep Learning}}, journal = {Scientific Reports}, year = 2019, number = 2058, month = 2, volume = 9, issue = 1, day = 14, url = "https://doi.org/10.1038/s41598-018-38343-3", doi = "10.1038/s41598-018-38343-3" }

  14. R

    Cifar 100 Dataset

    • universe.roboflow.com
    • opendatalab.com
    • +3more
    zip
    Updated Aug 11, 2022
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Popular Benchmarks (2022). Cifar 100 Dataset [Dataset]. https://universe.roboflow.com/popular-benchmarks/cifar100
    Explore at:
    zipAvailable download formats
    Dataset updated
    Aug 11, 2022
    Dataset authored and provided by
    Popular Benchmarks
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Variables measured
    Animals People CommonObjects
    Description

    CIFAR-100

    The CIFAR-10 and CIFAR-100 dataset contains labeled subsets of the 80 million tiny images dataset. They were collected by Alex Krizhevsky, Vinod Nair, and Geoffrey Hinton. * More info on CIFAR-100: https://www.cs.toronto.edu/~kriz/cifar.html * TensorFlow listing of the dataset: https://www.tensorflow.org/datasets/catalog/cifar100 * GitHub repo for converting CIFAR-100 tarball files to png format: https://github.com/knjcode/cifar2png

    All images were sized 32x32 in the original dataset

    The CIFAR-10 dataset consists of 60,000 32x32 colour images in 10 classes, with 6,000 images per class. There are 50,000 training images and 10,000 test images [in the original dataset].

    This dataset is just like the CIFAR-10, except it has 100 classes containing 600 images each. There are 500 training images and 100 testing images per class. The 100 classes in the CIFAR-100 are grouped into 20 superclasses. Each image comes with a "fine" label (the class to which it belongs) and a "coarse" label (the superclass to which it belongs). However, this project does not contain the superclasses. * Superclasses version: https://universe.roboflow.com/popular-benchmarks/cifar100-with-superclasses/

    More background on the dataset: https://i.imgur.com/5w8A0Vm.png" alt="CIFAR-100 Dataset Classes and Superclassees">

    Version 1 (original-images_Original-CIFAR100-Splits):

    • Original images, with the original splits for CIFAR-100: train (83.33% of images - 50,000 images) set and test (16.67% of images - 10,000 images) set only.
    • This version was not trained

    Version 2 (original-images_trainSetSplitBy80_20):

    • Original, raw images, with the train set split to provide 80% of its images to the training set (approximately 40,000 images) and 20% of its images to the validation set (approximately 10,000 images)
    • Trained from Roboflow Classification Model's ImageNet training checkpoint
    • https://blog.roboflow.com/train-test-split/ https://i.imgur.com/kSPeKGn.png" alt="Train/Valid/Test Split Rebalancing">

    Citation:

    @TECHREPORT{Krizhevsky09learningmultiple,
      author = {Alex Krizhevsky},
      title = {Learning multiple layers of features from tiny images},
      institution = {},
      year = {2009}
    }
    
  15. o

    mobilenet-v1.onnx model

    • explore.openaire.eu
    Updated Apr 10, 2019
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Https Github Com Tensorflow Models Blob Master Research Slim Nets Mobilenet V. https://github.com/tensorflow/models/blob/master/research/slim/nets/mobilenet_v1.md (2019). mobilenet-v1.onnx model [Dataset]. http://doi.org/10.5281/zenodo.3157894
    Explore at:
    Dataset updated
    Apr 10, 2019
    Authors
    Https Github Com Tensorflow Models Blob Master Research Slim Nets Mobilenet V. https://github.com/tensorflow/models/blob/master/research/slim/nets/mobilenet_v1.md
    Description

    Application: Image Classification ML Task: mobilenetv1 Framework: ONNX (via tensorflow) Training Information: Quality: 70.9% Precision: fp32 Is Quantized: no Is ONNX: yes Dataset: http://www.image-net.org/challenges/LSVRC/2012/ script to convert from tensorflow: https://gist.github.com/guschmue/788ae7f602c1f15ce3998b8d5f56ed2e

  16. R

    Mnist Dataset

    • universe.roboflow.com
    • tensorflow.org
    • +3more
    zip
    Updated Aug 8, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Popular Benchmarks (2022). Mnist Dataset [Dataset]. https://universe.roboflow.com/popular-benchmarks/mnist-cjkff/model/2
    Explore at:
    zipAvailable download formats
    Dataset updated
    Aug 8, 2022
    Dataset authored and provided by
    Popular Benchmarks
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Variables measured
    Digits
    Description

    THE MNIST DATABASE of handwritten digits

    Authors:

    • Yann LeCun, Courant Institute, NYU
    • Corinna Cortes, Google Labs, New York
    • Christopher J.C. Burges, Microsoft Research, Redmond

    Dataset Obtained From: http://yann.lecun.com/exdb/mnist/

    All images were sized 28x28 in the original dataset

    The MNIST database of handwritten digits, available from this page, has a training set of 60,000 examples, and a test set of 10,000 examples. It is a subset of a larger set available from NIST. The digits have been size-normalized and centered in a fixed-size image.

    It is a good database for people who want to try learning techniques and pattern recognition methods on real-world data while spending minimal efforts on preprocessing and formatting.

    Version 1 (original-images_trainSetSplitBy80_20):

    • Original, raw images, with the train set split to provide 80% of its images to the training set and 20% of its images to the validation set
    • Trained from Roboflow Classification Model's ImageNet training checkpoint

    Version 2 (original-images_ModifiedClasses_trainSetSplitBy80_20):

    • Original, raw images, with the train set split to provide 80% of its images to the training set and 20% of its images to the validation set
    • Modify Classes, a Roboflow preprocessing feature, was employed to change class names from 0, 1, 2, 3, 4, 5, 6, 7, 8, 9 to one, two, three, four, five, six, seven, eight, nine
    • Trained from the Roboflow Classification Model's ImageNet training checkpoint

    Version 3 (original-images_Original-MNIST-Splits):

    • Original images, with the original splits for MNIST: train (86% of images - 60,000 images) set and test (14% of images - 10,000 images) set only.
    • This version was not trained

    Citation:

    @article{lecun2010mnist,
     title={MNIST handwritten digit database},
     author={LeCun, Yann and Cortes, Corinna and Burges, CJ},
     journal={ATT Labs [Online]. Available: http://yann.lecun.com/exdb/mnist},
     volume={2},
     year={2010}
    }
    
  17. a

    palm zip

    • data-ocpw.opendata.arcgis.com
    Updated Sep 13, 2021
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    OC Public Works (2021). palm zip [Dataset]. https://data-ocpw.opendata.arcgis.com/content/f5130f997eb44a29a2ad474c9db62a4c
    Explore at:
    Dataset updated
    Sep 13, 2021
    Dataset authored and provided by
    OC Public Works
    Description

    The model file “palm_zip.dlpk” was initially a zip file, which has been modified as .dlpk in order to meet the requirement of the ArcGIS deep learning package format. Please change “dlpk” back to “zip” and extract all. You might want to refer the model outside ArcGIS platform. The model has been trained by transfer learning with samples clipped from the ECW format 2011 Eagle Aerial image of Orange County, on the basis of Tensorflow.Keras ResNet50 model with dataset of ImageNet. There is two classes in this model, "'others" and "palm", corresponding to 0 and 1 respectively in the model. All Palm species including King Palm, Queen Palm, Mexican Palm, etc. belong to “palm”, while all other tree species and land covers belong to “others”. The training samples have been resized to 224 by 224 pixels, so please be aware of your image size and content. It is suggested that only those objects with maximal prediction score class as ‘1’ (referring to Palm) and score greater than 0.9 be selected as Palm.

  18. h

    stanford-dogs

    • huggingface.co
    • tensorflow.org
    Updated Dec 4, 2023
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Alan Blanchet (2023). stanford-dogs [Dataset]. https://huggingface.co/datasets/Alanox/stanford-dogs
    Explore at:
    Dataset updated
    Dec 4, 2023
    Authors
    Alan Blanchet
    License

    MIT Licensehttps://opensource.org/licenses/MIT
    License information was derived automatically

    Description

    The Stanford Dogs dataset contains images of 120 breeds of dogs from around the world. This dataset has been built using images and annotation from ImageNet for the task of fine-grained image categorization.

  19. T

    imagenet2012_multilabel

    • tensorflow.org
    Updated Dec 10, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    (2022). imagenet2012_multilabel [Dataset]. https://www.tensorflow.org/datasets/catalog/imagenet2012_multilabel
    Explore at:
    Dataset updated
    Dec 10, 2022
    Description

    This dataset contains ILSVRC-2012 (ImageNet) validation images annotated with multi-class labels from "Evaluating Machine Accuracy on ImageNet", ICML, 2020. The multi-class labels were reviewed by a panel of experts extensively trained in the intricacies of fine-grained class distinctions in the ImageNet class hierarchy (see paper for more details). Compared to the original labels, these expert-reviewed multi-class labels enable a more semantically coherent evaluation of accuracy.

    Version 3.0.0 of this dataset contains more corrected labels from "When does dough become a bagel? Analyzing the remaining mistakes on ImageNet as well as the ImageNet-Major (ImageNet-M) 68-example split under 'imagenet-m'.

    Only 20,000 of the 50,000 ImageNet validation images have multi-label annotations. The set of multi-labels was first generated by a testbed of 67 trained ImageNet models, and then each individual model prediction was manually annotated by the experts as either correct (the label is correct for the image),wrong (the label is incorrect for the image), or unclear (no consensus was reached among the experts).

    Additionally, during annotation, the expert panel identified a set of problematic images. An image was problematic if it met any of the below criteria:

    • The original ImageNet label (top-1 label) was incorrect or unclear
    • Image was a drawing, painting, sketch, cartoon, or computer-rendered
    • Image was excessively edited
    • Image had inappropriate content

    The problematic images are included in this dataset but should be ignored when computing multi-label accuracy. Additionally, since the initial set of 20,000 annotations is class-balanced, but the set of problematic images is not, we recommend computing the per-class accuracies and then averaging them. We also recommend counting a prediction as correct if it is marked as correct or unclear (i.e., being lenient with the unclear labels).

    One possible way of doing this is with the following NumPy code:

    import tensorflow_datasets as tfds
    
    ds = tfds.load('imagenet2012_multilabel', split='validation')
    
    # We assume that predictions is a dictionary from file_name to a class index between 0 and 999
    
    num_correct_per_class = {}
    num_images_per_class = {}
    
    for example in ds:
      # We ignore all problematic images
      if example[‘is_problematic’].numpy():
        continue
    
      # The label of the image in ImageNet
      cur_class = example['original_label'].numpy()
    
      # If we haven't processed this class yet, set the counters to 0
      if cur_class not in num_correct_per_class:
        num_correct_per_class[cur_class] = 0
        assert cur_class not in num_images_per_class
        num_images_per_class[cur_class] = 0
    
      num_images_per_class[cur_class] += 1
    
      # Get the predictions for this image
      cur_pred = predictions[example['file_name'].numpy()]
    
      # We count a prediction as correct if it is marked as correct or unclear
      # (i.e., we are lenient with the unclear labels)
      if cur_pred is in example['correct_multi_labels'].numpy() or cur_pred is in example['unclear_multi_labels'].numpy():
        num_correct_per_class[cur_class] += 1
    
    # Check that we have collected accuracy data for each of the 1,000 classes
    num_classes = 1000
    assert len(num_correct_per_class) == num_classes
    assert len(num_images_per_class) == num_classes
    
    # Compute the per-class accuracies and then average them
    final_avg = 0
    for cid in range(num_classes):
     assert cid in num_correct_per_class
     assert cid in num_images_per_class
     final_avg += num_correct_per_class[cid] / num_images_per_class[cid]
    final_avg /= num_classes
    
    

    To use this dataset:

    import tensorflow_datasets as tfds
    
    ds = tfds.load('imagenet2012_multilabel', split='train')
    for ex in ds.take(4):
     print(ex)
    

    See the guide for more informations on tensorflow_datasets.

    https://storage.googleapis.com/tfds-data/visualization/fig/imagenet2012_multilabel-3.0.0.png" alt="Visualization" width="500px">

  20. o

    deepNIR: Dataset for generating synthetic NIR images and improved fruit...

    • explore.openaire.eu
    Updated Mar 15, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Inkyu Sa; Jong Yoon Lim; Ho Seok Ahn (2022). deepNIR: Dataset for generating synthetic NIR images and improved fruit detection system using deep learning techniques [Dataset]. http://doi.org/10.5281/zenodo.6324489
    Explore at:
    Dataset updated
    Mar 15, 2022
    Authors
    Inkyu Sa; Jong Yoon Lim; Ho Seok Ahn
    Description

    In this paper, we present datasets that can be utilised for synthetic near infrared (NIR) image and bounding box level fruit detection system. It is undeniable fact that high-caliber machine learning software frameworks such as Tensorflow or Pytorch and large scale dataset such as ImageNet and COCO, and accelerated GPU hardware support have pushed the limit of machine learning for more than decades. Among these breakthroughs quality dataset is one of important key building blocks that can lead to success in model generalisation and deployment for data-driven deep neural networks. Particularly, synthetic data generation such as generative adversarial networks often requires relatively larger scale data than other supervised approaches. In addition, posing constrains such as geometrical facial constrains in fake face generation or consistent and radiometrically calibrated reflectances from satellite imagery commonly yield better results. We share NIR+RGB dataset that are re-processed from other two public datasets (nirscene and SEN12MS) and our own novel sweetpepper dataset to be able to timely adopt to other following studies. We oversampled from original nirscene dataset at 10, 100, 200, and 400 ratios and total of 127k pair of images. For SEN12MS satellite multispectral dataset, we selected one largest subset; Summer (45k) and All seasons (180k). Our sweetpeppr dataset consists of 1,615 pairs of NIR+RGB images. We demonstrate these NIR+RGB datasets are sufficient to be used for synthetic NIR generation quantitatively and qualitatively. We achieved Frechet Inception Distance (FID) of 11.36, 26.53, and 40.15 for nirscene1, SEN12MS, and sweetpepper dataset respectively. We also release 11 fruits' bounding box annotations that can be exported as various formats using cloud service. 4 newly added fruits [blueberry, cherry, kiwi, and wheat] compounds 11 novel bounding box dastaset together with our previous work in deepFruits project [apple, avocado, capsicum, mango, orange, rockmelon, strawberry]. The total number of bounding box instances is 162k and all bounding box dataset is ready for use from cloud service. For evaluation of these dataset, Yolov5 single stage detector is exploited and reported impressive mean-average-precision, mAP[0.5:0.95] results of [min:0.49, max:0.812]. We hope these dataset is useful and serves as one of baseline for the following up studies.

Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
(2024). imagenet2012 [Dataset]. https://www.tensorflow.org/datasets/catalog/imagenet2012

imagenet2012

Related Article
Explore at:
443 scholarly articles cite this dataset (View in Google Scholar)
Dataset updated
Jun 1, 2024
Description

ILSVRC 2012, commonly known as 'ImageNet' is an image dataset organized according to the WordNet hierarchy. Each meaningful concept in WordNet, possibly described by multiple words or word phrases, is called a "synonym set" or "synset". There are more than 100,000 synsets in WordNet, majority of them are nouns (80,000+). In ImageNet, we aim to provide on average 1000 images to illustrate each synset. Images of each concept are quality-controlled and human-annotated. In its completion, we hope ImageNet will offer tens of millions of cleanly sorted images for most of the concepts in the WordNet hierarchy.

The test split contains 100K images but no labels because no labels have been publicly released. We provide support for the test split from 2012 with the minor patch released on October 10, 2019. In order to manually download this data, a user must perform the following operations:

  1. Download the 2012 test split available here.
  2. Download the October 10, 2019 patch. There is a Google Drive link to the patch provided on the same page.
  3. Combine the two tar-balls, manually overwriting any images in the original archive with images from the patch. According to the instructions on image-net.org, this procedure overwrites just a few images.

The resulting tar-ball may then be processed by TFDS.

To assess the accuracy of a model on the ImageNet test split, one must run inference on all images in the split, export those results to a text file that must be uploaded to the ImageNet evaluation server. The maintainers of the ImageNet evaluation server permits a single user to submit up to 2 submissions per week in order to prevent overfitting.

To evaluate the accuracy on the test split, one must first create an account at image-net.org. This account must be approved by the site administrator. After the account is created, one can submit the results to the test server at https://image-net.org/challenges/LSVRC/eval_server.php The submission consists of several ASCII text files corresponding to multiple tasks. The task of interest is "Classification submission (top-5 cls error)". A sample of an exported text file looks like the following:

771 778 794 387 650
363 691 764 923 427
737 369 430 531 124
755 930 755 59 168

The export format is described in full in "readme.txt" within the 2013 development kit available here: https://image-net.org/data/ILSVRC/2013/ILSVRC2013_devkit.tgz Please see the section entitled "3.3 CLS-LOC submission format". Briefly, the format of the text file is 100,000 lines corresponding to each image in the test split. Each line of integers correspond to the rank-ordered, top 5 predictions for each test image. The integers are 1-indexed corresponding to the line number in the corresponding labels file. See labels.txt.

To use this dataset:

import tensorflow_datasets as tfds

ds = tfds.load('imagenet2012', split='train')
for ex in ds.take(4):
 print(ex)

See the guide for more informations on tensorflow_datasets.

https://storage.googleapis.com/tfds-data/visualization/fig/imagenet2012-5.1.0.png" alt="Visualization" width="500px">

Search
Clear search
Close search
Google apps
Main menu