3 datasets found
  1. h

    dlgenai-nppe-data

    • huggingface.co
    Updated Nov 16, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Rohan kumar (2025). dlgenai-nppe-data [Dataset]. https://huggingface.co/datasets/Rohan014/dlgenai-nppe-data
    Explore at:
    Dataset updated
    Nov 16, 2025
    Authors
    Rohan kumar
    License

    MIT Licensehttps://opensource.org/licenses/MIT
    License information was derived automatically

    Description

    Age & Gender Prediction - Project Data

    Training logs and metrics for the 25-t3-nppe1 project.

      Project Information
    

    Project ID: 25-t3-nppe1 Model: ResNet50 (Finetuned) Framework: PyTorch + PyTorch Lightning Space: https://huggingface.co/spaces/Rohan014/dlgenai-nppe

      Model Performance
    

    Validation RMSE: 8.5 Gender F1 Score: 0.62 Final Score: 0.65

      Training Configuration
    

    Epochs: 4 Batch Size: 64 Learning Rate: 0.0001 Optimizer: AdamW Scheduler:… See the full description on the dataset page: https://huggingface.co/datasets/Rohan014/dlgenai-nppe-data.

  2. Geospatial Deep Learning Seminar Online Course

    • ckan.americaview.org
    Updated Nov 2, 2021
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    ckan.americaview.org (2021). Geospatial Deep Learning Seminar Online Course [Dataset]. https://ckan.americaview.org/dataset/geospatial-deep-learning-seminar-online-course
    Explore at:
    Dataset updated
    Nov 2, 2021
    Dataset provided by
    CKANhttps://ckan.org/
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    This seminar is an applied study of deep learning methods for extracting information from geospatial data, such as aerial imagery, multispectral imagery, digital terrain data, and other digital cartographic representations. We first provide an introduction and conceptualization of artificial neural networks (ANNs). Next, we explore appropriate loss and assessment metrics for different use cases followed by the tensor data model, which is central to applying deep learning methods. Convolutional neural networks (CNNs) are then conceptualized with scene classification use cases. Lastly, we explore semantic segmentation, object detection, and instance segmentation. The primary focus of this course is semantic segmenation for pixel-level classification. The associated GitHub repo provides a series of applied examples. We hope to continue to add examples as methods and technologies further develop. These examples make use of a vareity of datasets (e.g., SAT-6, topoDL, Inria, LandCover.ai, vfillDL, and wvlcDL). Please see the repo for links to the data and associated papers. All examples have associated videos that walk through the process, which are also linked to the repo. A variety of deep learning architectures are explored including UNet, UNet++, DeepLabv3+, and Mask R-CNN. Currenlty, two examples use ArcGIS Pro and require no coding. The remaining five examples require coding and make use of PyTorch, Python, and R within the RStudio IDE. It is assumed that you have prior knowledge of coding in the Python and R enviroinments. If you do not have experience coding, please take a look at our Open-Source GIScience and Open-Source Spatial Analytics (R) courses, which explore coding in Python and R, respectively. After completing this seminar you will be able to: explain how ANNs work including weights, bias, activation, and optimization. describe and explain different loss and assessment metrics and determine appropriate use cases. use the tensor data model to represent data as input for deep learning. explain how CNNs work including convolutional operations/layers, kernel size, stride, padding, max pooling, activation, and batch normalization. use PyTorch, Python, and R to prepare data, produce and assess scene classification models, and infer to new data. explain common semantic segmentation architectures and how these methods allow for pixel-level classification and how they are different from traditional CNNs. use PyTorch, Python, and R (or ArcGIS Pro) to prepare data, produce and assess semantic segmentation models, and infer to new data.

  3. RepVGG

    • kaggle.com
    zip
    Updated Jan 22, 2021
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Mobassir (2021). RepVGG [Dataset]. https://www.kaggle.com/mobassir/repvgg
    Explore at:
    zip(2349894983 bytes)Available download formats
    Dataset updated
    Jan 22, 2021
    Authors
    Mobassir
    Description

    RepVGG: Making VGG-style ConvNets Great Again (PyTorch)

    taken from : https://github.com/DingXiaoH/RepVGG

    This is a super simple ConvNet architecture that achieves over 80% top-1 accuracy on ImageNet with a stack of 3x3 conv and ReLU! This repo contains the pretrained models, code for building the model, training, and the conversion from training-time model to inference-time.

    The MegEngine version: https://github.com/megvii-model/RepVGG.

    Update (Jan 13, 2021): you can get the equivalent kernel and bias in a differentiable way at any time (get_equivalent_kernel_bias in repvgg.py). This may help training-based pruning or quantization.

    Citation:

    @article{ding2101repvgg,
     title={RepVGG: Making VGG-style ConvNets Great Again},
     author={Ding, Xiaohan and Zhang, Xiangyu and Ma, Ningning and Han, Jungong and Ding, Guiguang and Sun, Jian},
     journal={arXiv preprint arXiv:2101.03697}
    }
    

    Abstract

    We present a simple but powerful architecture of convolutional neural network, which has a VGG-like inference-time body composed of nothing but a stack of 3x3 convolution and ReLU, while the training-time model has a multi-branch topology. Such decoupling of the training-time and inference-time architecture is realized by a structural re-parameterization technique so that the model is named RepVGG. On ImageNet, RepVGG reaches over 80\% top-1 accuracy, which is the first time for a plain model, to the best of our knowledge. On NVIDIA 1080Ti GPU, RepVGG models run 83% faster than ResNet-50 or 101% faster than ResNet-101 with higher accuracy and show favorable accuracy-speed trade-off compared to the state-of-the-art models like EfficientNet and RegNet.

    https://github.com/DingXiaoH/RepVGG/raw/main/arch.PNG" alt="">

    https://github.com/DingXiaoH/RepVGG/raw/main/speed_acc.PNG" alt="">

    https://github.com/DingXiaoH/RepVGG/raw/main/table.PNG" alt="">

    Use our pretrained models

    You may download all of the ImageNet-pretrained models reported in the paper from Google Drive (https://drive.google.com/drive/folders/1Avome4KvNp0Lqh2QwhXO6L5URQjzCjUq?usp=sharing) or Baidu Cloud (https://pan.baidu.com/s/1nCsZlMynnJwbUBKn0ch7dQ, the access code is "rvgg"). For the ease of transfer learning on other tasks, they are all training-time models (with identity and 1x1 branches). You may test the accuracy by running python test.py [imagenet-folder with train and val folders] train [path to weights file] -a [model name] Here "train" indicates the training-time architecture. For example, python test.py [imagenet-folder with train and val folders] train RepVGG-B2-train.pth -a RepVGG-B2

    Convert the training-time models into inference-time

    You may convert a trained model into the inference-time structure with python convert.py [weights file of the training-time model to load] [path to save] -a [model name] For example, python convert.py RepVGG-B2-train.pth RepVGG-B2-deploy.pth -a RepVGG-B2 Then you may test the inference-time model by python test.py [imagenet-folder with train and val folders] deploy RepVGG-B2-deploy.pth -a RepVGG-B2 Note that the argument "deploy" builds an inference-time model.

    ImageNet training settings

    We trained for 120 epochs with cosine learning rate decay from 0.1 to 0. We used 8 GPUs, global batch size of 256, weight decay of 1e-4 (no weight decay on fc.bias, bn.bias, rbr_dense.bn.weight and rbr_1x1.bn.weight) (weight decay on rbr_identity.weight makes little difference, and it is better to use it in most of the cases), and the same simple data preprocssing as the PyTorch official example: trans = transforms.Compose([ transforms.RandomResizedCrop(224), transforms.RandomHorizontalFlip(), transforms.ToTensor(), transforms.Normalize(mean=[0.485, 0.456, 0.406], std=[0.229, 0.224, 0.225]) ]) The multi-processing training script in this repo is based on the official PyTorch example for the better readability. The modifications include the model-building part, cosine learning rate scheduler, and the SGD optimizer that uses no weight decay on some parameters. You may find these code segments useful for your training code. This training script has not been tested because I don't have raw ImageNet training data. I would really appreciate it if you share with me your re-implementation results with this script. For example, python train.py -a RepVGG-A0 --dist-url 'tcp://127.0.0.1:23333' --dist-backend 'nccl' --multiprocessing-distributed --world-size 1 --rank 0 [imagenet-folder with train and val folders]

    Use like this in your own code

    from repvgg import repvgg_model_convert, create_RepVGG_A0
    train_model = create_RepVGG_A0(deploy=False)
    trai...
    
  4. Not seeing a result you expected?
    Learn how you can add new datasets to our index.

Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
Rohan kumar (2025). dlgenai-nppe-data [Dataset]. https://huggingface.co/datasets/Rohan014/dlgenai-nppe-data

dlgenai-nppe-data

Rohan014/dlgenai-nppe-data

Explore at:
Dataset updated
Nov 16, 2025
Authors
Rohan kumar
License

MIT Licensehttps://opensource.org/licenses/MIT
License information was derived automatically

Description

Age & Gender Prediction - Project Data

Training logs and metrics for the 25-t3-nppe1 project.

  Project Information

Project ID: 25-t3-nppe1 Model: ResNet50 (Finetuned) Framework: PyTorch + PyTorch Lightning Space: https://huggingface.co/spaces/Rohan014/dlgenai-nppe

  Model Performance

Validation RMSE: 8.5 Gender F1 Score: 0.62 Final Score: 0.65

  Training Configuration

Epochs: 4 Batch Size: 64 Learning Rate: 0.0001 Optimizer: AdamW Scheduler:… See the full description on the dataset page: https://huggingface.co/datasets/Rohan014/dlgenai-nppe-data.

Search
Clear search
Close search
Google apps
Main menu