14 datasets found
  1. Imagenet extracted features with ResNet

    • kaggle.com
    zip
    Updated Jul 21, 2019
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    J.Guzman (2019). Imagenet extracted features with ResNet [Dataset]. https://www.kaggle.com/jgoodman8/imagenet-features-resnet
    Explore at:
    zip(0 bytes)Available download formats
    Dataset updated
    Jul 21, 2019
    Authors
    J.Guzman
    License

    Attribution-ShareAlike 4.0 (CC BY-SA 4.0)https://creativecommons.org/licenses/by-sa/4.0/
    License information was derived automatically

    Description

    Overview

    This dataset contains features extracted from the Imagenet dataset using a pre-trained ResNet neural network. The network was configured with an input layer of (200, 200, 3). Feature extraction was performed using the Python package Py Image Feature Extractor.

    Source

    Related datasets

  2. Z

    Model Zoo: A Dataset of Diverse Populations of Resnet-18 Models - Tiny...

    • data.niaid.nih.gov
    • zenodo.org
    Updated Aug 28, 2022
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Giró-i-Nieto, Xavier (2022). Model Zoo: A Dataset of Diverse Populations of Resnet-18 Models - Tiny ImageNet [Dataset]. https://data.niaid.nih.gov/resources?id=zenodo_7023277
    Explore at:
    Dataset updated
    Aug 28, 2022
    Dataset provided by
    Taskiran, Diyar
    Schürholt, Konstantin
    Giró-i-Nieto, Xavier
    Knyazev, Boris
    Borth, Damian
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Abstract

    In the last years, neural networks have evolved from laboratory environments to the state-of-the-art for many real-world problems. Our hypothesis is that neural network models (i.e., their weights and biases) evolve on unique, smooth trajectories in weight space during training. Following, a population of such neural network models (refereed to as “model zoo”) would form topological structures in weight space. We think that the geometry, curvature and smoothness of these structures contain information about the state of training and can be reveal latent properties of individual models. With such zoos, one could investigate novel approaches for (i) model analysis, (ii) discover unknown learning dynamics, (iii) learn rich representations of such populations, or (iv) exploit the model zoos for generative modelling of neural network weights and biases. Unfortunately, the lack of standardized model zoos and available benchmarks significantly increases the friction for further research about populations of neural networks. With this work, we publish a novel dataset of model zoos containing systematically generated and diverse populations of neural network models for further research. In total the proposed model zoo dataset is based on six image datasets, consist of 27 model zoos with varying hyperparameter combinations are generated and includes 50’360 unique neural network models resulting in over 2’585’360 collected model states. Additionally, to the model zoo data we provide an in-depth analysis of the zoos and provide benchmarks for multiple downstream tasks as mentioned before.

    Dataset

    This dataset is part of a larger collection of model zoos and contains the zoo of 1000 ResNet18 models trained on Tiny Imagenet. All zoos with extensive information and code can be found at www.modelzoos.cc.

    The complete zoo is 2.6TB large. Due to the size, this repository contains the checkpoints of the first 115 models at their last epoch 60. For a link to the full dataset as well as more information on the zoos and code to access and use the zoos, please see www.modelzoos.cc.

  3. LGV pretrained models

    • figshare.com
    bin
    Updated Jun 2, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Martin Gubri (2023). LGV pretrained models [Dataset]. http://doi.org/10.6084/m9.figshare.20497821.v1
    Explore at:
    binAvailable download formats
    Dataset updated
    Jun 2, 2023
    Dataset provided by
    Figsharehttp://figshare.com/
    Authors
    Martin Gubri
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    LGV models used as surrogate in the original paper.

    Those resnet50 models were collected along the SGD trajectory with a high learning rate. The zip file contains three random seeds in respective subfolders. Each one contains a subfolder with the original pretrained model from which the model collection started. These pretrained models were trained by Ashukha, A., et al. Pitfalls of In-Domain Uncertainty Estimation and Ensembling in Deep Learning (2020).

  4. f

    Performance on ImageNet dataset. We train the models following the standard...

    • figshare.com
    • plos.figshare.com
    xls
    Updated Jun 11, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Sheng Ran; Tao Huang; Wuyue Yang (2025). Performance on ImageNet dataset. We train the models following the standard training strategy with pre-trained teacher networks ResNet-34 and ResNet-50 provided by Torchvision [42]. [Dataset]. http://doi.org/10.1371/journal.pone.0325599.t001
    Explore at:
    xlsAvailable download formats
    Dataset updated
    Jun 11, 2025
    Dataset provided by
    PLOS ONE
    Authors
    Sheng Ran; Tao Huang; Wuyue Yang
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Performance on ImageNet dataset. We train the models following the standard training strategy with pre-trained teacher networks ResNet-34 and ResNet-50 provided by Torchvision [42].

  5. S

    Data from: Deep Learning, Feature Learning, and Clustering Analysis for SEM...

    • scidb.cn
    Updated Oct 17, 2020
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Rossella Aversa; Piero Coronica; Cristiano De Nobili; Stefano Cozzini (2020). Deep Learning, Feature Learning, and Clustering Analysis for SEM Image Classification [Dataset]. http://doi.org/10.11922/sciencedb.j00104.00062
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Oct 17, 2020
    Dataset provided by
    Science Data Bank
    Authors
    Rossella Aversa; Piero Coronica; Cristiano De Nobili; Stefano Cozzini
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    One table and six figures. Table 1 shows the number of images for each label in the 1μ–2μ data set, adopting the same labelling used in [11, 12, 13], reported here for completeness: 0 = Porous sponges, 1 = Patterned surfaces, 2 = Particles, 3 = Films and coated surfaces, 4 = Powders, 5 = Tips, 6 = Nanowires, 7 = Biological, 8 = MEMS devices and electrodes, 9 = Fibres.Figure 1 shows test accuracy as a function of the number of training epochs obtained by training from scratch Inception-v3 (magenta), Inception-v4 (orange), Inception-Resnet (green), and AlexNet (black) on SEM data set. All the models were trained with the best combination of hyperparameters, according to the memory capability of the available hardware. In Figure 2, Main: Test accuracy as a function of the number of training epochs obtained when fine tuning on the SEM data set Inception-v3 (magenta) and Inception-v4 (orange) starting from the ImageNet checkpoint, and Inception-v3 (blue) from the SEM checkpoint that, as expected, converges very rapidly. Inset: Test accuracy as a function of the number of training epochs obtained when performing feature extraction of Inception-v3 (magenta), Inception-v4 (orange), and Inception-Resnet (green) on the SEM data set starting from the ImageNet checkpoint. All the models were trained with the best combination of hyperparameters, according to the memory capability of the hardware available. Figure 3 shows intrinsic Dimension of the 1μ–2μ_1001 data set, varying the sample size, computed before autoencoding (green lines) and after autoencoding (red lines). The three brightness levels for each color correspond to the percentage of points used in the linear fi t: 90%, 70%, and 50%. Figure 4 shows ddisc heatmap for a manually labelled subset of images. Figure 5 presents heatmaps of the distances obtained via Inception-v3. The image captions specify the methods used and indicate the correlation index with ddisc. Figure 6 shows NMI scores of the clustering obtained by the five hierarchical algorithms (solid lines) considered as a function of k, the number of clusters. The scores of the artificial scenarios are reported as orange (good case) and green (uniform case) dashed lines.

  6. u

    Image-Based Honey Bee Larval Viral and Bacterial Diagnosis Using Machine...

    • agdatacommons.nal.usda.gov
    bin
    Updated Jul 23, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    United States Department of Agriculture (2025). Image-Based Honey Bee Larval Viral and Bacterial Diagnosis Using Machine Learning [Dataset]. https://agdatacommons.nal.usda.gov/articles/dataset/Image-Based_Honey_Bee_Larval_Viral_and_Bacterial_Diagnosis_Using_Machine_Learning/29622290
    Explore at:
    binAvailable download formats
    Dataset updated
    Jul 23, 2025
    Dataset provided by
    National Center for Biotechnology Information
    Authors
    United States Department of Agriculture
    License

    https://rightsstatements.org/vocab/UND/1.0/https://rightsstatements.org/vocab/UND/1.0/

    Description

    Honey bees are essential pollinators of ecosystems and agriculture worldwide. With an estimated 50-80% of crops pollinated by honey bees, they generate approximately $20 billion in market value in the U.S. alone. However, commercial beekeepers often face an uphill battle, losing anywhere from 40-90% of their hives annually, significantly impacted by brood diseases caused by bacterial, viral, and fungal pathogens. Accurate diagnosis of these brood diseases, especially distinguishing bacterial diseases like European Foulbrood (EFB) from viral infections with a superficial resemblance to EFB (EFB-like disease), remains challenging. Incorrect diagnoses often lead to prophylactic antibiotic treatment across entire apiaries, exacerbating antibiotic resistance, disrupting native gut microbiota, and increasing susceptibility to opportunistic pathogens. Correct field diagnosis of brood disease is challenging and requires years of experience to identify and differentiate various disease states according to subtle differences in larval symptomology. To explore the feasibility of an image-based AI diagnosis tool, we collaborated with apiary inspectors and researchers to generate a dataset of 2,759 honey bee larvae images from Michigan apiaries, molecularly verified through 16S rRNA microbiome sequencing and qPCR viral screening. Our dataset included EFB cases and viral infections (ABPV, DWVA, and DWVB), which were augmented to 8,430 and 8,124 images respectively. We leveraged transfer learning techniques, fine-tuning deep convolutional neural networks (ResNet-50v2, ResNet-101v2, InceptionResNet-v2) pre-trained on ImageNet to discriminate between EFB and viral infections. Our proof-of-concept models achieved 73-88% accuracy on the training/validation sets. When tested on an independent dataset from Illinois containing additional viral pathogens not present in training data, the models showed higher accuracy for EFB (72-88%) than viral infections (28-68%), highlighting both the promise and current limitations of this approach. Implementing AI-based diagnostic tools could reduce unnecessary antibiotic treatments and help maintain the microbiome integrity critical to colony health. However, expanding training datasets to include all major pathogens, healthy larvae, and diverse geographic regions will be essential for developing field-ready diagnostic tools.

  7. Z

    DeepWeeds: A Multiclass Weed Species Image Dataset for Deep Learning

    • data.niaid.nih.gov
    Updated May 16, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Rahimi Azghadi, Mostafa (2023). DeepWeeds: A Multiclass Weed Species Image Dataset for Deep Learning [Dataset]. https://data.niaid.nih.gov/resources?id=zenodo_7939059
    Explore at:
    Dataset updated
    May 16, 2023
    Dataset provided by
    Calvert, Brendan
    White, Ronald D.
    Konovalov, Dimitriv A.
    Whinney, James
    Rahimi Azghadi, Mostafa
    Kenny, Owen
    Olsen, Alex
    Johns, Jamie
    Wood, Jake C.
    Banks, Wesley
    Ridd, Peter
    Philippa, Bronson
    Girgenti, Benjamin
    License

    http://www.apache.org/licenses/LICENSE-2.0http://www.apache.org/licenses/LICENSE-2.0

    Description

    DeepWeeds: A Multiclass Weed Species Image Dataset for Deep Learning

    This repository makes available the source code and public dataset for the work, "DeepWeeds: A Multiclass Weed Species Image Dataset for Deep Learning", published with open access by Scientific Reports: https://www.nature.com/articles/s41598-018-38343-3. The DeepWeeds dataset consists of 17,509 images capturing eight different weed species native to Australia in situ with neighbouring flora. In our work, the dataset was classified to an average accuracy of 95.7% with the ResNet50 deep convolutional neural network.

    The source code, images and annotations are licensed under CC BY 4.0 license. The contents of this repository are released under an Apache 2 license.

    Download the dataset images and our trained models

    images.zip (468 MB)

    models.zip (477 MB)

    Due to the size of the images and models they are hosted outside of the Github repository. The images and models must be downloaded into directories named "images" and "models", respectively, at the root of the repository. If you execute the python script (deepweeds.py), as instructed below, this step will be performed for you automatically.

    TensorFlow Datasets

    Alternatively, you can access the DeepWeeds dataset with TensorFlow Datasets, TensorFlow's official collection of ready-to-use datasets. DeepWeeds was officially added to the TensorFlow Datasets catalog in August 2019.

    Weeds and locations

    The selected weed species are local to pastoral grasslands across the state of Queensland. They include: "Chinee apple", "Snake weed", "Lantana", "Prickly acacia", "Siam weed", "Parthenium", "Rubber vine" and "Parkinsonia". The images were collected from weed infestations at the following sites across Queensland: "Black River", "Charters Towers", "Cluden", "Douglas", "Hervey Range", "Kelso", "McKinlay" and "Paluma". The table and figure below break down the dataset by weed, location and geographical distribution.

    Data organization

    Images are assigned unique filenames that include the date/time the image was photographed and an ID number for the instrument which produced the image. The format is like so: YYYYMMDD-HHMMSS-ID, where the ID is simply an integer from 0 to 3. The unique filenames are strings of 17 characters, such as 20170320-093423-1.

    labels

    The labels.csv file assigns species labels to each image. It is a comma separated text file in the format:

    Filename,Label,Species ... 20170207-154924-0,jpg,7,Snake weed 20170610-123859-1.jpg,1,Lantana 20180119-105722-1.jpg,8,Negative ...

    Note: The specific label subsets of training (60%), validation (20%) and testing (20%) for the five-fold cross validation used in the paper are also provided here as CSV files in the same format as "labels.csv".

    models

    We provide the most successful ResNet50 and InceptionV3 models saved in Keras' hdf5 model format. The ResNet50 model, which provided the best results, has also been converted to UFF format in order to construct a TensorRT inference engine.

    resnet.hdf5 inception.hdf5 resnet.uff

    deepweeds.py

    This python script trains and evaluates Keras' base implementation of ResNet50 and InceptionV3 on the DeepWeeds dataset, pre-trained with ImageNet weights. The performance of the networks are cross validated for 5 folds. The final classification accuracy is taken to be the average across the five folds. Similarly, the final confusion matrix from the associated paper aggregates across the five independent folds. The script also provides the ability to measure the inference speeds within the TensorFlow environment.

    The script can be executed to carry out these computations using the following commands.

    To train and evaluate the ResNet50 model with five-fold cross validation, use python3 deepweeds.py cross_validate --model resnet.

    To train and evaluate the InceptionV3 model with five-fold cross validation, use python3 deepweeds.py cross_validate --model inception.

    To measure inference times for the ResNet50 model, use python3 deepweeds.py inference --model models/resnet.hdf5.

    To measure inference times for the InceptionV3 model, use python3 deepweeds.py inference --model models/inception.hdf5.

    Dependencies

    The required Python packages to execute deepweeds.py are listed in requirements.txt.

    tensorrt

    This folder includes C++ source code for creating and executing a ResNet50 TensorRT inference engine on an NVIDIA Jetson TX2 platform. To build and run on your Jetson TX2, execute the following commands:

    cd tensorrt/src make -j4 cd ../bin ./resnet_inference

    Citations

    If you use the DeepWeeds dataset in your work, please cite it as:

    IEEE style citation: “A. Olsen, D. A. Konovalov, B. Philippa, P. Ridd, J. C. Wood, J. Johns, W. Banks, B. Girgenti, O. Kenny, J. Whinney, B. Calvert, M. Rahimi Azghadi, and R. D. White, “DeepWeeds: A Multiclass Weed Species Image Dataset for Deep Learning,” Scientific Reports, vol. 9, no. 2058, 2 2019. [Online]. Available: https://doi.org/10.1038/s41598-018-38343-3

    BibTeX

    @article{DeepWeeds2019, author = {Alex Olsen and Dmitry A. Konovalov and Bronson Philippa and Peter Ridd and Jake C. Wood and Jamie Johns and Wesley Banks and Benjamin Girgenti and Owen Kenny and James Whinney and Brendan Calvert and Mostafa {Rahimi Azghadi} and Ronald D. White}, title = {{DeepWeeds: A Multiclass Weed Species Image Dataset for Deep Learning}}, journal = {Scientific Reports}, year = 2019, number = 2058, month = 2, volume = 9, issue = 1, day = 14, url = "https://doi.org/10.1038/s41598-018-38343-3", doi = "10.1038/s41598-018-38343-3" }

  8. Classification report of EfficientNet.

    • plos.figshare.com
    xls
    Updated Sep 6, 2024
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Najam Aziz; Nasru Minallah; Jaroslav Frnda; Madiha Sher; Muhammad Zeeshan; Amara Haroon Durrani (2024). Classification report of EfficientNet. [Dataset]. http://doi.org/10.1371/journal.pone.0307825.t002
    Explore at:
    xlsAvailable download formats
    Dataset updated
    Sep 6, 2024
    Dataset provided by
    PLOShttp://plos.org/
    Authors
    Najam Aziz; Nasru Minallah; Jaroslav Frnda; Madiha Sher; Muhammad Zeeshan; Amara Haroon Durrani
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Brain tumors pose significant global health concerns due to their high mortality rates and limited treatment options. These tumors, arising from abnormal cell growth within the brain, exhibits various sizes and shapes, making their manual detection from magnetic resonance imaging (MRI) scans a subjective and challenging task for healthcare professionals, hence necessitating automated solutions. This study investigates the potential of deep learning, specifically the DenseNet architecture, to automate brain tumor classification, aiming to enhance accuracy and generalizability for clinical applications. We utilized the Figshare brain tumor dataset, comprising 3,064 T1-weighted contrast-enhanced MRI images from 233 patients with three prevalent tumor types: meningioma, glioma, and pituitary tumor. Four pre-trained deep learning models—ResNet, EfficientNet, MobileNet, and DenseNet—were evaluated using transfer learning from ImageNet. DenseNet achieved the highest test set accuracy of 96%, outperforming ResNet (91%), EfficientNet (91%), and MobileNet (93%). Therefore, we focused on improving the performance of the DenseNet, while considering it as base model. To enhance the generalizability of the base DenseNet model, we implemented a fine-tuning approach with regularization techniques, including data augmentation, dropout, batch normalization, and global average pooling, coupled with hyperparameter optimization. This enhanced DenseNet model achieved an accuracy of 97.1%. Our findings demonstrate the effectiveness of DenseNet with transfer learning and fine-tuning for brain tumor classification, highlighting its potential to improve diagnostic accuracy and reliability in clinical settings.

  9. f

    Comparison of identification accuracy and training time between different...

    • plos.figshare.com
    xls
    Updated Jun 21, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Lin Lin; Jie Zhang; Xu Gao; Jiancheng Shi; Cheng Chen; Nantian Huang (2023). Comparison of identification accuracy and training time between different training models on the PLAID dataset. [Dataset]. http://doi.org/10.1371/journal.pone.0281482.t002
    Explore at:
    xlsAvailable download formats
    Dataset updated
    Jun 21, 2023
    Dataset provided by
    PLOS ONE
    Authors
    Lin Lin; Jie Zhang; Xu Gao; Jiancheng Shi; Cheng Chen; Nantian Huang
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Comparison of identification accuracy and training time between different training models on the PLAID dataset.

  10. f

    The F1 for different data balance algorithms using transferred...

    • plos.figshare.com
    xls
    Updated Jun 21, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Lin Lin; Jie Zhang; Xu Gao; Jiancheng Shi; Cheng Chen; Nantian Huang (2023). The F1 for different data balance algorithms using transferred CBAM-ResNet34. [Dataset]. http://doi.org/10.1371/journal.pone.0281482.t003
    Explore at:
    xlsAvailable download formats
    Dataset updated
    Jun 21, 2023
    Dataset provided by
    PLOS ONE
    Authors
    Lin Lin; Jie Zhang; Xu Gao; Jiancheng Shi; Cheng Chen; Nantian Huang
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    The F1 for different data balance algorithms using transferred CBAM-ResNet34.

  11. f

    The PLIAD dataset of appliance types and instance statistics.

    • plos.figshare.com
    xls
    Updated Jun 21, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Lin Lin; Jie Zhang; Xu Gao; Jiancheng Shi; Cheng Chen; Nantian Huang (2023). The PLIAD dataset of appliance types and instance statistics. [Dataset]. http://doi.org/10.1371/journal.pone.0281482.t001
    Explore at:
    xlsAvailable download formats
    Dataset updated
    Jun 21, 2023
    Dataset provided by
    PLOS ONE
    Authors
    Lin Lin; Jie Zhang; Xu Gao; Jiancheng Shi; Cheng Chen; Nantian Huang
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    The PLIAD dataset of appliance types and instance statistics.

  12. Tencent ML-Images

    • opendatalab.com
    zip
    Updated May 3, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Tencent AI Lab (2023). Tencent ML-Images [Dataset]. https://opendatalab.com/OpenDataLab/Tencent_ML-Images
    Explore at:
    zipAvailable download formats
    Dataset updated
    May 3, 2023
    Dataset provided by
    Tencent AI Lab
    腾讯https://tencent.com/
    University of Hong Kong
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    This repository introduces the open-source project dubbed Tencent ML-Images, which publishes ML-Images: the largest open-source multi-label image database, including 17,609,752 training and 88,739 validation image URLs, which are annotated with up to 11,166 categories Resnet-101 model: it is pre-trained on ML-Images, and achieves the top-1 accuracy 80.73% on ImageNet via transfer learning

  13. f

    The comparison of the proposed method and other power fingerprint...

    • plos.figshare.com
    xls
    Updated Jun 21, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Lin Lin; Jie Zhang; Xu Gao; Jiancheng Shi; Cheng Chen; Nantian Huang (2023). The comparison of the proposed method and other power fingerprint identification methods. [Dataset]. http://doi.org/10.1371/journal.pone.0281482.t004
    Explore at:
    xlsAvailable download formats
    Dataset updated
    Jun 21, 2023
    Dataset provided by
    PLOS ONE
    Authors
    Lin Lin; Jie Zhang; Xu Gao; Jiancheng Shi; Cheng Chen; Nantian Huang
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    The comparison of the proposed method and other power fingerprint identification methods.

  14. f

    Data_Sheet_1_Rethinking Breiman's Dilemma in Neural Networks: Phase...

    • frontiersin.figshare.com
    pdf
    Updated Jun 1, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Weizhi Zhu; Yifei Huang; Yuan Yao (2023). Data_Sheet_1_Rethinking Breiman's Dilemma in Neural Networks: Phase Transitions of Margin Dynamics.pdf [Dataset]. http://doi.org/10.3389/fams.2020.575073.s001
    Explore at:
    pdfAvailable download formats
    Dataset updated
    Jun 1, 2023
    Dataset provided by
    Frontiers
    Authors
    Weizhi Zhu; Yifei Huang; Yuan Yao
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Margin enlargement of training data has been an important strategy for perceptrons in machine learning for the purpose of boosting the confidence of training toward a good generalization ability. Yet Breiman (1999) shows a dilemma: a uniform improvement on margin distribution does not necessarily reduce generalization errors. In this paper, we revisit Breiman's dilemma in deep neural networks with recently proposed spectrally normalized margins from a novel perspective based on phase transitions of normalized margin distributions in training dynamics. Normalized margin distribution of a classifier of the data can be divided into two parts: low/small margins such as some negative margins for misclassified samples vs. high/large margins for high confident correctly classified samples, which often behave differently during the training process. Low margins for training and test datasets are often effectively reduced in training, along with reductions of training and test errors, whereas high margins may exhibit different dynamics, reflecting the trade-off between the expressive power of models and the complexity of data. When data complexity is comparable to the model expressiveness, high margin distributions for both training and test data undergo similar decrease-increase phase transitions during training. In such cases, one can predict the trend of generalization or test error through margin-based generalization bounds with restricted Rademacher complexities, shown in two ways in this paper with early stopping time exploiting such phase transitions. On the other hand, over-expressive models may have both low and high training margins undergoing uniform improvements with a distinct phase transition in test margin dynamics. This reconfirms the Breiman's dilemma associated with over-parameterized neural networks where margins fail to predict overfitting. Experiments are conducted with some basic convolutional networks, AlexNet, VGG-16, and ResNet-18, on several datasets, including Cifar10/100 and mini-ImageNet.

  15. Not seeing a result you expected?
    Learn how you can add new datasets to our index.

Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
J.Guzman (2019). Imagenet extracted features with ResNet [Dataset]. https://www.kaggle.com/jgoodman8/imagenet-features-resnet
Organization logo

Imagenet extracted features with ResNet

Features extracted from the Imagenet dataset using ResNet

Explore at:
zip(0 bytes)Available download formats
Dataset updated
Jul 21, 2019
Authors
J.Guzman
License

Attribution-ShareAlike 4.0 (CC BY-SA 4.0)https://creativecommons.org/licenses/by-sa/4.0/
License information was derived automatically

Description

Overview

This dataset contains features extracted from the Imagenet dataset using a pre-trained ResNet neural network. The network was configured with an input layer of (200, 200, 3). Feature extraction was performed using the Python package Py Image Feature Extractor.

Source

Related datasets

Search
Clear search
Close search
Google apps
Main menu