19 datasets found
  1. cnn_c1

    • kaggle.com
    Updated Mar 9, 2021
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    satya (2021). cnn_c1 [Dataset]. https://www.kaggle.com/satyapr/cnn-c1/code
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Mar 9, 2021
    Dataset provided by
    Kagglehttp://kaggle.com/
    Authors
    satya
    Description

    Dataset

    This dataset was created by satya

    Contents

  2. MNIST-224by224-train-test-dataset

    • kaggle.com
    Updated Nov 26, 2021
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    DHRUV Desh (2021). MNIST-224by224-train-test-dataset [Dataset]. https://www.kaggle.com/dhruvdesh/mnist224by224testdataset/code
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Nov 26, 2021
    Dataset provided by
    Kagglehttp://kaggle.com/
    Authors
    DHRUV Desh
    License

    https://creativecommons.org/publicdomain/zero/1.0/https://creativecommons.org/publicdomain/zero/1.0/

    Description

    Context

    I needed 224 by 224 version of MNIST dataset for one of my projects so I made this.

    Content

    The dataset has 35 files in idx3-ubyte format with 2000 images each and dimension 224x224. 5 of these are test data files and 30 are train data files.

    Acknowledgements

    The actual MNIST dataset creators

  3. f

    Model comparison results using MNIST-C and MNIST-C-shape datasets.

    • plos.figshare.com
    xls
    Updated Jun 13, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Seoyoung Ahn; Hossein Adeli; Gregory J. Zelinsky (2024). Model comparison results using MNIST-C and MNIST-C-shape datasets. [Dataset]. http://doi.org/10.1371/journal.pcbi.1012159.t001
    Explore at:
    xlsAvailable download formats
    Dataset updated
    Jun 13, 2024
    Dataset provided by
    PLOS Computational Biology
    Authors
    Seoyoung Ahn; Hossein Adeli; Gregory J. Zelinsky
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Recognition accuracy (means and standard deviations from 5 trained models, hereafter referred to as model “runs”) from ORA and two CNN baselines, both of which were trained using identical CNN encoders (one a 2-layer CNN and the other a Resnet-18), and a CapsNet model following the implementation in [51].

  4. P

    N-MNIST Dataset

    • paperswithcode.com
    Updated Mar 31, 2023
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    (2023). N-MNIST Dataset [Dataset]. https://paperswithcode.com/dataset/n-mnist
    Explore at:
    Dataset updated
    Mar 31, 2023
    Description

    Brief Description The Neuromorphic-MNIST (N-MNIST) dataset is a spiking version of the original frame-based MNIST dataset. It consists of the same 60 000 training and 10 000 testing samples as the original MNIST dataset, and is captured at the same visual scale as the original MNIST dataset (28x28 pixels). The N-MNIST dataset was captured by mounting the ATIS sensor on a motorized pan-tilt unit and having the sensor move while it views MNIST examples on an LCD monitor as shown in this video. A full description of the dataset and how it was created can be found in the paper below. Please cite this paper if you make use of the dataset.

    Orchard, G.; Cohen, G.; Jayawant, A.; and Thakor, N. “Converting Static Image Datasets to Spiking Neuromorphic Datasets Using Saccades", Frontiers in Neuroscience, vol.9, no.437, Oct. 2015

  5. Robustness assessment of a C++ implementation of a quantized (int8) version...

    • zenodo.org
    • data.niaid.nih.gov
    zip
    Updated Nov 22, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    David de Andrés; David de Andrés; Juan Carlos Ruiz; Juan Carlos Ruiz (2023). Robustness assessment of a C++ implementation of a quantized (int8) version of the LeNet-5 convolutional neural network [Dataset]. http://doi.org/10.5281/zenodo.10196616
    Explore at:
    zipAvailable download formats
    Dataset updated
    Nov 22, 2023
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    David de Andrés; David de Andrés; Juan Carlos Ruiz; Juan Carlos Ruiz
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Time period covered
    Jun 24, 2023 - Jun 26, 2023
    Description

    The architecture of the LeNet-5 convolutional neural network (CNN) was defined by LeCun in its paper "Gradient-based learning applied to document recognition" (https://ieeexplore.ieee.org/document/726791) to classify images of hand written digits (MNIST dataset).

    This architecture has been customized to use Rectified Linear Unit (ReLU) as activation functions instead of Sigmoid, and 8-bit integers for weights and activations instead of floating-point.

    It consists of the following layers:

    • conv1: Convolution 2D, 1 input channel (28x28), 3 output channels (28x28), kernel size 5, stride 1, padding 2.
    • relu1: Rectified Linear Unit (3@28x28).
    • max1: Subsampling buy max pooling (3@14x14).
    • conv2: Convolution 2D, 3 input channels (14x14), 6 output channels (14x14), kernel size 5, stride 1, padding 2.
    • relu2: Rectified Linear Unit (6@14x14).
    • max2: Subsampling buy max pooling (6@7x7).
    • fc1: Fully connected (294, 147)
    • fc2: Fully connected (147, 10)

    The fault hypotheses for this work include the occurrence of:

    • BF: single, double-adjacent and triple-adjacent bit-flip faults
    • S0: single, double-adjacent and triple-adjacent stuck-at-0 faults
    • S1: single, double-adjacent and triple-adjacent stuck-at-1 faults

    In the memory cells containing all the parameters of the CNN:

    • w: weights (int8)
    • zw: zero point of the weights (int8)
    • b: biases (int32)
    • z: zero point (int8)
    • m: m (int32)

    Images 200 to 249 from the MNIST dataset have been used as workload.

    This dataset contains the raw data obtained from running exhaustive fault injection campaigns for all considered fault models, targeting all considered locations and for all the images in the workload.

    In addition, the raw data have been lightly processed to obtain global data related to the particular bits and parameters affected by the faults, and the obtained failure modes.

    Files information

    • golden_run.csv: Prediction obtained for all the images considered in the workload in the absence of faults (Golden Run). This is intended to act as oracle to determine the impact of injected faults.
    • single_faults/bit_flip folder: Prediction obtained for all the images considered in the workload in presence of single bit-flip faults. There is one file for each parameter of each layer.
    • single_faults/stuck_at_0 folder: Prediction obtained for all the images considered in the workload in presence of single stuck-at-0 faults. There is one file for each parameter of each layer.
    • single_faults/stuck_at_1 folder: Prediction obtained for all the images considered in the workload in presence of single stuck-at-1 faults. There is one file for each parameter of each layer.
    • double_adjacent_faults/bit_flip folder: Prediction obtained for all the images considered in the workload in presence of double adjacent bit-flip faults. There is one file for each parameter of each layer.
    • double_adjacent_faults/stuck_at_0 folder: Prediction obtained for all the images considered in the workload in presence of double adjacent stuck-at-0 faults. There is one file for each parameter of each layer.
    • double_adjacent_faults/stuck_at_1 folder: Prediction obtained for all the images considered in the workload in presence of double adjacent stuck-at-1 faults. There is one file for each parameter of each layer.
    • triple_adjacent_faults/bit_flip folder: Prediction obtained for all the images considered in the workload in presence of triple adjacent bit-flip faults. There is one file for each parameter of each layer.
    • triple_adjacent_faults/stuck_at_0 folder: Prediction obtained for all the images considered in the workload in presence of triple adjacent stuck-at-0 faults. There is one file for each parameter of each layer.
    • triple_adjacent_faults/stuck_at_1 folder: Prediction obtained for all the images considered in the workload in presence of triple adjacent stuck-at-1 faults. There is one file for each parameter of each layer.

    Methodology information

    First, the CNN was used to classify all the images of the workload in the absence of faults to get a reference to determine the impact of faults. This is golden_run.csv file.

    After that, one fault injection experiment was executed for each bit of each element of each parameter of the CNN.

    Each experiment consisted in:

    • Affecting the bits (inverting it in case of bit-flip faults, setting it to 0 or 1 in case of stuck-at-0 or atuck-at-1 faults) identified by the mask.
    • Classifying all the images of the workload in the presence of this fault. The obtained output was stored in a given .csv file.
    • Removing the fault from the CNN by restoring the affected bits to its previous value.

    List of variables (Name : Description (Possible values))

    • IMGID: Integer number identifying the considered image (200-249).
    • TENSORID: Integer number identiying the parameter affected by the fault (0 - No fault, 1 - conv1.w, 2 - conv1.zw, 3 - conv1.m, 4 - conv1.b, 5 - conv1.z, 6 - conv2.w, 7 - conv2.zw, 8 - conv2.m, 9 - conv2.b, 10 - conv2.z, 11 - fc1.w, 12 - fc1.zw, 13 - fc1.m, 14 - fc.b, 15 - fc1.z, 16 - fc2.w, 17 - fc2.zw, 18 - fc2.m, 19 - fc2.b, 20 - fc2.z)
    • ELEMID: Integer number identiying the element of the parameter affected by the fault (-1 - No fault, [0-2] - {conv1.b, conv1.m, conv1.zw}, [0-74] - conv1.w, 0 - conv1.z, [0-5] - {conv2.b, conv2.m, conv2.zw}, [0-149] - conv2.w, 0 - {conv1.z, conv2.z, fc1.z, fc2.z}, [0-146] - {fc1.b, fc1.m, fc1.zw}, [0-43217] - fc1.w, [0-9] - {fc2.b, fc2.m, fc2.zw}, [0-1469] - fc2.w)
    • MASK: 8-digit hexadecimal number identifying those bits affected by the fault ([00000000 - No fault, FFFFFFFF - all 32 bits faulty])
    • FAULT: String identiying the type of fault (NF - No fault, BF - bit-flip, S0 - Stuck-at-0, S1 - Stuck-at-1)
    • OUTPUT: 10 integer numbers provided by the CNN as output after processing the image. The highest value identifies the selected category for classification.
    • SOFTMAX: 10 decimal numbers obtained after applying the softmax function to the provided output. They represent the probability of the image of belonging to the corresponding category for classification.
    • PRED: Integer number representing the category predicted for the processed image.
    • LABEL: integer number representing the actual category for the processed image.

  6. o

    [ATHENA] CNN WDs on MNIST

    • explore.openaire.eu
    Updated Oct 23, 2020
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Ying Meng; Pooyan Jamshidi (2020). [ATHENA] CNN WDs on MNIST [Dataset]. http://doi.org/10.5281/zenodo.4121900
    Explore at:
    Dataset updated
    Oct 23, 2020
    Authors
    Ying Meng; Pooyan Jamshidi
    Description

    CNN weak defenses and the undefended model (model-mnist-cnn-clean.h5) trained on MNIST.

  7. Baseline models and optimized CNN models for 8 datasets

    • zenodo.org
    • scidb.cn
    • +1more
    application/gzip
    Updated Jan 24, 2020
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Martin Thoma; Martin Thoma (2020). Baseline models and optimized CNN models for 8 datasets [Dataset]. http://doi.org/10.5281/zenodo.582892
    Explore at:
    application/gzipAvailable download formats
    Dataset updated
    Jan 24, 2020
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Martin Thoma; Martin Thoma
    License

    Attribution-ShareAlike 4.0 (CC BY-SA 4.0)https://creativecommons.org/licenses/by-sa/4.0/
    License information was derived automatically

    Description

    Datasets:

    • Asirra
    • CIFAR-10
    • CIFAR-100
    • GTSRB
    • HASYv2
    • MNIST
    • STL-10
    • SVHN
  8. file-mnist-cnn

    • kaggle.com
    Updated Mar 16, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    NgoNguyen1512dino (2024). file-mnist-cnn [Dataset]. https://www.kaggle.com/datasets/ngonguyen1512dino/file-mnist-cnn/data
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Mar 16, 2024
    Dataset provided by
    Kagglehttp://kaggle.com/
    Authors
    NgoNguyen1512dino
    Description

    Dataset

    This dataset was created by NgoNguyen1512dino

    Contents

  9. f

    DataSheet1_A resource-efficient quantum convolutional neural network.PDF

    • frontiersin.figshare.com
    pdf
    Updated Apr 5, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Yanqi Song; Jing Li; Yusen Wu; Sujuan Qin; Qiaoyan Wen; Fei Gao (2024). DataSheet1_A resource-efficient quantum convolutional neural network.PDF [Dataset]. http://doi.org/10.3389/fphy.2024.1362690.s001
    Explore at:
    pdfAvailable download formats
    Dataset updated
    Apr 5, 2024
    Dataset provided by
    Frontiers
    Authors
    Yanqi Song; Jing Li; Yusen Wu; Sujuan Qin; Qiaoyan Wen; Fei Gao
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Quantum Convolutional Neural Network (QCNN) has achieved significant success in solving various complex problems, such as quantum many-body physics and image recognition. In comparison to the classical Convolutional Neural Network (CNN) model, the QCNN model requires excellent numerical performance or efficient computational resources to showcase its potential quantum advantages, particularly in classical data processing tasks. In this paper, we propose a computationally resource-efficient QCNN model referred to as RE-QCNN. Specifically, through a comprehensive analysis of the complexity associated with the forward and backward propagation processes in the quantum convolutional layer, our results demonstrate a significant reduction in computational resources required for this layer compared to the classical CNN model. Furthermore, our model is numerically benchmarked on recognizing images from the MNIST and Fashion-MNIST datasets, achieving high accuracy in these multi-class classification tasks.

  10. Mnist_play

    • kaggle.com
    zip
    Updated Nov 6, 2021
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Tianyi Peng (2021). Mnist_play [Dataset]. https://www.kaggle.com/tianyipeng/mnist-play
    Explore at:
    zip(6767 bytes)Available download formats
    Dataset updated
    Nov 6, 2021
    Authors
    Tianyi Peng
    Description

    Dataset

    This dataset was created by Tianyi Peng

    Contents

    It contains the following files:

  11. MNIST_CNN_9959

    • kaggle.com
    Updated May 15, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Phan Văn Hoàng (2024). MNIST_CNN_9959 [Dataset]. https://www.kaggle.com/datasets/phanvnhong/mnist-cnn-9959
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    May 15, 2024
    Dataset provided by
    Kagglehttp://kaggle.com/
    Authors
    Phan Văn Hoàng
    License

    MIT Licensehttps://opensource.org/licenses/MIT
    License information was derived automatically

    Description

    Dataset

    This dataset was created by Phan Văn Hoàng

    Released under MIT

    Contents

  12. f

    Parameter settings of experiments.

    • plos.figshare.com
    xls
    Updated Aug 16, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Shuyang Ren; Eunsam Kim; Choonhwa Lee (2024). Parameter settings of experiments. [Dataset]. http://doi.org/10.1371/journal.pone.0308991.t002
    Explore at:
    xlsAvailable download formats
    Dataset updated
    Aug 16, 2024
    Dataset provided by
    PLOS ONE
    Authors
    Shuyang Ren; Eunsam Kim; Choonhwa Lee
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Various deep learning techniques, including blockchain-based approaches, have been explored to unlock the potential of edge data processing and resultant intelligence. However, existing studies often overlook the resource requirements of blockchain consensus processing in typical Internet of Things (IoT) edge network settings. This paper presents our FLCoin approach. Specifically, we propose a novel committee-based method for consensus processing in which committee members are elected via the FL process. Additionally, we employed a two-layer blockchain architecture for federated learning (FL) processing to facilitate the seamless integration of blockchain and FL techniques. Our analysis reveals that the communication overhead remains stable as the network size increases, ensuring the scalability of our blockchain-based FL system. To assess the performance of the proposed method, experiments were conducted using the MNIST dataset to train a standard five-layer CNN model. Our evaluation demonstrated the efficiency of FLCoin. With an increasing number of nodes participating in the model training, the consensus latency remained below 3 s, resulting in a low total training time. Notably, compared with a blockchain-based FL system utilizing PBFT as the consensus protocol, our approach achieved a 90% improvement in communication overhead and a 35% reduction in training time cost. Our approach ensures an efficient and scalable solution, enabling the integration of blockchain and FL into IoT edge networks. The proposed architecture provides a solid foundation for building intelligent IoT services.

  13. f

    Main notations used in the paper.

    • plos.figshare.com
    xls
    Updated Aug 16, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Shuyang Ren; Eunsam Kim; Choonhwa Lee (2024). Main notations used in the paper. [Dataset]. http://doi.org/10.1371/journal.pone.0308991.t001
    Explore at:
    xlsAvailable download formats
    Dataset updated
    Aug 16, 2024
    Dataset provided by
    PLOS ONE
    Authors
    Shuyang Ren; Eunsam Kim; Choonhwa Lee
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Various deep learning techniques, including blockchain-based approaches, have been explored to unlock the potential of edge data processing and resultant intelligence. However, existing studies often overlook the resource requirements of blockchain consensus processing in typical Internet of Things (IoT) edge network settings. This paper presents our FLCoin approach. Specifically, we propose a novel committee-based method for consensus processing in which committee members are elected via the FL process. Additionally, we employed a two-layer blockchain architecture for federated learning (FL) processing to facilitate the seamless integration of blockchain and FL techniques. Our analysis reveals that the communication overhead remains stable as the network size increases, ensuring the scalability of our blockchain-based FL system. To assess the performance of the proposed method, experiments were conducted using the MNIST dataset to train a standard five-layer CNN model. Our evaluation demonstrated the efficiency of FLCoin. With an increasing number of nodes participating in the model training, the consensus latency remained below 3 s, resulting in a low total training time. Notably, compared with a blockchain-based FL system utilizing PBFT as the consensus protocol, our approach achieved a 90% improvement in communication overhead and a 35% reduction in training time cost. Our approach ensures an efficient and scalable solution, enabling the integration of blockchain and FL into IoT edge networks. The proposed architecture provides a solid foundation for building intelligent IoT services.

  14. MNIST Preprocessed

    • kaggle.com
    Updated Jul 24, 2019
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Valentyn Sichkar (2019). MNIST Preprocessed [Dataset]. https://www.kaggle.com/valentynsichkar/mnist-preprocessed/kernels
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Jul 24, 2019
    Dataset provided by
    Kagglehttp://kaggle.com/
    Authors
    Valentyn Sichkar
    Description

    📰 Related Paper

    Sichkar V. N. Effect of various dimension convolutional layer filters on traffic sign classification accuracy. Scientific and Technical Journal of Information Technologies, Mechanics and Optics, 2019, vol. 19, no. 3, pp. DOI: 10.17586/2226-1494-2019-19-3-546-552 (Full-text available here ResearchGate.net/profile/Valentyn_Sichkar)

    Test online with custom Traffic Sign here: https://valentynsichkar.name/mnist.html


    :mortar_board: Related course for classification tasks

    Design, Train & Test deep CNN for Image Classification. Join the course & enjoy new opportunities to get deep learning skills: https://www.udemy.com/course/convolutional-neural-networks-for-image-classification/

    https://github.com/sichkar-valentyn/1-million-images-for-Traffic-Signs-Classification-tasks/blob/main/images/slideshow_classification.gif?raw=true%20=470x516" alt="CNN Course" title="CNN Course">


    🗺️ Concept Map of the Course

    https://github.com/sichkar-valentyn/1-million-images-for-Traffic-Signs-Classification-tasks/blob/main/images/concept_map.png?raw=true%20=570x410" alt="Concept map" title="Concept map">


    👉 Join the Course

    https://www.udemy.com/course/convolutional-neural-networks-for-image-classification/


    Content

    This is ready to use preprocessed data saved into pickle file.
    Preprocessing stages are as follows:
    - Normalizing whole data by dividing / 255.0.
    - Dividing whole data into three datasets: train, validation and test.
    - Normalizing whole data by subtracting mean image and dividing by standard deviation.
    - Transposing every dataset to make channels come first.


    mean image and standard deviation were calculated from train dataset and applied to all datasets.
    When using user's image for classification, it has to be preprocessed firstly in the same way: normalized, subtracted with mean image and divided by standard deviation.


    Data written as dictionary with following keys:
    x_train: (59000, 1, 28, 28)
    y_train: (59000,)
    x_validation: (1000, 1, 28, 28)
    y_validation: (1000,)
    x_test: (1000, 1, 28, 28)
    y_test: (1000,)


    Contains pretrained weights model_params_ConvNet1.pickle for the model with following architecture:
    Input --> Conv --> ReLU --> Pool --> Affine --> ReLU --> Affine --> Softmax


    Parameters:

    • Input is 1-channeled GrayScale image.
    • 32 filters of Convolutional Layer.
    • Stride for Pool is 2 and height = width = 2.
    • Number of hidden neurons is 500.
    • Number of output neurons is 10.


    Architecture also can be understood as follows:
    https://www.googleapis.com/download/storage/v1/b/kaggle-user-content/o/inbox%2F3400968%2Fc23041248e82134b7d43ed94307b720e%2FModel_1_Architecture_MNIST.png?generation=1563654250901965&alt=media" alt="">

    Acknowledgements

    Initial data is MNIST that was collected by Yann LeCun, Corinna Cortes, Christopher J.C. Burges.

  15. f

    CNN2 architecture.

    • plos.figshare.com
    xls
    Updated Jul 22, 2024
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Bernardo Pulido-Gaytan; Andrei Tchernykh (2024). CNN2 architecture. [Dataset]. http://doi.org/10.1371/journal.pone.0306420.t004
    Explore at:
    xlsAvailable download formats
    Dataset updated
    Jul 22, 2024
    Dataset provided by
    PLOS ONE
    Authors
    Bernardo Pulido-Gaytan; Andrei Tchernykh
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    The widespread adoption of cloud computing necessitates privacy-preserving techniques that allow information to be processed without disclosure. This paper proposes a method to increase the accuracy and performance of privacy-preserving Convolutional Neural Networks with Homomorphic Encryption (CNN-HE) by Self-Learning Activation Functions (SLAF). SLAFs are polynomials with trainable coefficients updated during training, together with synaptic weights, for each polynomial independently to learn task-specific and CNN-specific features. We theoretically prove its feasibility to approximate any continuous activation function to the desired error as a function of the SLAF degree. Two CNN-HE models are proposed: CNN-HE-SLAF and CNN-HE-SLAF-R. In the first model, all activation functions are replaced by SLAFs, and CNN is trained to find weights and coefficients. In the second one, CNN is trained with the original activation, then weights are fixed, activation is substituted by SLAF, and CNN is shortly re-trained to adapt SLAF coefficients. We show that such self-learning can achieve the same accuracy 99.38% as a non-polynomial ReLU over non-homomorphic CNNs and lead to an increase in accuracy (99.21%) and higher performance (6.26 times faster) than the state-of-the-art CNN-HE CryptoNets on the MNIST optical character recognition benchmark dataset.

  16. CSGO Guns Dataset

    • kaggle.com
    Updated Oct 2, 2020
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Yash Khatri (2020). CSGO Guns Dataset [Dataset]. https://www.kaggle.com/khatriyash/csgo-guns-dataset/activity
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Oct 2, 2020
    Dataset provided by
    Kagglehttp://kaggle.com/
    Authors
    Yash Khatri
    Description

    Start here if...

    You have some experience with ANN, but you’re new to computer vision. This is the perfect introduction to working with bigger images than MNIST and also working with raw images.

    Description

    Your goal is to correctly identify the type of CSGO gun from a dataset of about 980 labeled images. We encourage you to experiment with different algorithms to learn first-hand what works well and how techniques compare.

    Acknowledgment

    This is a Dataset created by Yash khatri by extracting images from CSGO third party software.

    Inspiration

    My interest in Deep learning and CSGO.

  17. f

    Time cost comparison of training with PBFT and FLCoin.

    • plos.figshare.com
    xls
    Updated Aug 16, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Shuyang Ren; Eunsam Kim; Choonhwa Lee (2024). Time cost comparison of training with PBFT and FLCoin. [Dataset]. http://doi.org/10.1371/journal.pone.0308991.t003
    Explore at:
    xlsAvailable download formats
    Dataset updated
    Aug 16, 2024
    Dataset provided by
    PLOS ONE
    Authors
    Shuyang Ren; Eunsam Kim; Choonhwa Lee
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Time cost comparison of training with PBFT and FLCoin.

  18. LLM prompts in the context of machine learning

    • kaggle.com
    Updated Jul 1, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Jordan Nelson (2024). LLM prompts in the context of machine learning [Dataset]. https://www.kaggle.com/datasets/jordanln/llm-prompts-in-the-context-of-machine-learning
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Jul 1, 2024
    Dataset provided by
    Kaggle
    Authors
    Jordan Nelson
    License

    https://creativecommons.org/publicdomain/zero/1.0/https://creativecommons.org/publicdomain/zero/1.0/

    Description

    This dataset is an extension of my previous work on creating a dataset for natural language processing tasks. It leverages binary representation to characterise various machine learning models. The attributes in the dataset are derived from a dictionary, which was constructed from a corpus of prompts typically provided to a large language model (LLM). These prompts reference specific machine learning algorithms and their implementations. For instance, consider a user asking an LLM or a generative AI to create a Multi-Layer Perceptron (MLP) model for a particular application. By applying this concept to multiple machine learning models, we constructed our corpus. This corpus was then transformed into the current dataset using a bag-of-words approach. In this dataset, each attribute corresponds to a word from our dictionary, represented as a binary value: 1 indicates the presence of the word in a given prompt, and 0 indicates its absence. At the end of each entry, there is a label. Each entry in the dataset pertains to a single class, where each class represents a distinct machine learning model or algorithm. This dataset is intended for multi-class classification tasks, not multi-label classification, as each entry is associated with only one label and does not belong to multiple labels simultaneously. This dataset has been utilised with a Convolutional Neural Network (CNN) using the Keras Automodel API, achieving impressive training and testing accuracy rates exceeding 97%. Post-training, the model's predictive performance was rigorously evaluated in a production environment, where it continued to demonstrate exceptional accuracy. For this evaluation, we employed a series of questions, which are listed below. These questions were intentionally designed to be similar to ensure that the model can effectively distinguish between different machine learning models, even when the prompts are closely related.

    KNN How would you create a KNN model to classify emails as spam or not spam based on their content and metadata? How could you implement a KNN model to classify handwritten digits using the MNIST dataset? How would you use a KNN approach to build a recommendation system for suggesting movies to users based on their ratings and preferences? How could you employ a KNN algorithm to predict the price of a house based on features such as its location, size, and number of bedrooms etc? Can you create a KNN model for classifying different species of flowers based on their petal length, petal width, sepal length, and sepal width? How would you utilise a KNN model to predict the sentiment (positive, negative, or neutral) of text reviews or comments? Can you create a KNN model for me that could be used in malware classification? Can you make me a KNN model that can detect a network intrusion when looking at encrypted network traffic? Can you make a KNN model that would predict the stock price of a given stock for the next week? Can you create a KNN model that could be used to detect malware when using a dataset relating to certain permissions a piece of software may have access to?

    Decision Tree Can you describe the steps involved in building a decision tree model to classify medical images as malignant or benign for cancer diagnosis and return a model for me? How can you utilise a decision tree approach to develop a model for classifying news articles into different categories (e.g., politics, sports, entertainment) based on their textual content? What approach would you take to create a decision tree model for recommending personalised university courses to students based on their academic strengths and weaknesses? Can you describe how to create a decision tree model for identifying potential fraud in financial transactions based on transaction history, user behaviour, and other relevant data? In what ways might you apply a decision tree model to classify customer complaints into different categories determining the severity of language used? Can you create a decision tree classifier for me? Can you make me a decision tree model that will help me determine the best course of action across a given set of strategies? Can you create a decision tree model for me that can recommend certain cars to customers based on their preferences and budget? How can you make a decision tree model that will predict the movement of star constellations in the sky based on data provided by the NASA website? How do I create a decision tree for time-series forecasting?

    Random Forest Can you describe the steps involved in building a random forest model to classify different types of anomalies in network traffic data for cybersecurity purposes and return the code for me? In what ways could you implement a random forest model to predict the severity of traffic congestion in urban areas based on historical traffic patterns, weather...

  19. f

    Data_Sheet_1_Scaling Equilibrium Propagation to Deep ConvNets by Drastically...

    • frontiersin.figshare.com
    pdf
    Updated Jun 1, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Axel Laborieux; Maxence Ernoult; Benjamin Scellier; Yoshua Bengio; Julie Grollier; Damien Querlioz (2023). Data_Sheet_1_Scaling Equilibrium Propagation to Deep ConvNets by Drastically Reducing Its Gradient Estimator Bias.pdf [Dataset]. http://doi.org/10.3389/fnins.2021.633674.s001
    Explore at:
    pdfAvailable download formats
    Dataset updated
    Jun 1, 2023
    Dataset provided by
    Frontiers
    Authors
    Axel Laborieux; Maxence Ernoult; Benjamin Scellier; Yoshua Bengio; Julie Grollier; Damien Querlioz
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Equilibrium Propagation is a biologically-inspired algorithm that trains convergent recurrent neural networks with a local learning rule. This approach constitutes a major lead to allow learning-capable neuromophic systems and comes with strong theoretical guarantees. Equilibrium propagation operates in two phases, during which the network is let to evolve freely and then “nudged” toward a target; the weights of the network are then updated based solely on the states of the neurons that they connect. The weight updates of Equilibrium Propagation have been shown mathematically to approach those provided by Backpropagation Through Time (BPTT), the mainstream approach to train recurrent neural networks, when nudging is performed with infinitely small strength. In practice, however, the standard implementation of Equilibrium Propagation does not scale to visual tasks harder than MNIST. In this work, we show that a bias in the gradient estimate of equilibrium propagation, inherent in the use of finite nudging, is responsible for this phenomenon and that canceling it allows training deep convolutional neural networks. We show that this bias can be greatly reduced by using symmetric nudging (a positive nudging and a negative one). We also generalize Equilibrium Propagation to the case of cross-entropy loss (by opposition to squared error). As a result of these advances, we are able to achieve a test error of 11.7% on CIFAR-10, which approaches the one achieved by BPTT and provides a major improvement with respect to the standard Equilibrium Propagation that gives 86% test error. We also apply these techniques to train an architecture with unidirectional forward and backward connections, yielding a 13.2% test error. These results highlight equilibrium propagation as a compelling biologically-plausible approach to compute error gradients in deep neuromorphic systems.

  20. Not seeing a result you expected?
    Learn how you can add new datasets to our index.

Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
satya (2021). cnn_c1 [Dataset]. https://www.kaggle.com/satyapr/cnn-c1/code
Organization logo

cnn_c1

train cnn model accuracy 98.3 mnist handwritten

Explore at:
CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
Dataset updated
Mar 9, 2021
Dataset provided by
Kagglehttp://kaggle.com/
Authors
satya
Description

Dataset

This dataset was created by satya

Contents

Search
Clear search
Close search
Google apps
Main menu