5 datasets found
  1. FastQuantileLayerKeras

    • kaggle.com
    Updated Jan 14, 2021
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Erik (2021). FastQuantileLayerKeras [Dataset]. https://www.kaggle.com/snippsy/fastquantilelayerkeras/tasks
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Jan 14, 2021
    Dataset provided by
    Kagglehttp://kaggle.com/
    Authors
    Erik
    Description
  2. Spatial distribution of particulate matter, collected using low cost...

    • zenodo.org
    bin
    Updated Apr 24, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Janani Venkatraman Jagatha; Janani Venkatraman Jagatha; Christoph Schneider; Christoph Schneider; Sebastian Schubert; Luxi Jin; Sebastian Schubert; Luxi Jin (2025). Spatial distribution of particulate matter, collected using low cost sensors, in Downtown-Singapore [Dataset]. http://doi.org/10.5281/zenodo.14280847
    Explore at:
    binAvailable download formats
    Dataset updated
    Apr 24, 2025
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Janani Venkatraman Jagatha; Janani Venkatraman Jagatha; Christoph Schneider; Christoph Schneider; Sebastian Schubert; Luxi Jin; Sebastian Schubert; Luxi Jin
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Area covered
    Downtown Core, Singapore
    Description

    The dataset consists of particulate matter concentration and meteorology data, measured in Singapore, Chinatown, and Central business district from March 13, 2018, to March 16, 2018. The data collectors walked from the Outram district - Chinatown to the Central Business District in Singapore. The measurements were carried out using a hand-held air quality sensor ensemble (URBMOBI 3.0).

    The dataset contains information from two URBMOBI 3.0 devices and one reference-grade device (Grimm 1.109). The data from the sensors and Grimm are denoted by the subscript, 's1', 's2', and 'gr', respectively.

    singapore_all_pm_25.geojson : The observed PM concentration and meteorology, aggregated using a 25 m buffer around the measurement points.

    Information on working with geojson file can be found under GeoJSON .

    Units:
    PM : µg/m³
    Scaled_PM_MM : Dimensionless entity scaled using Min-Max-Scaler (https://scikit-learn.org/stable/modules/generated/sklearn.preprocessing.MinMaxScaler.html)
    Scaled_PM_SS : Dimensionless entity scaled using Standard-Scaler (https://scikit-learn.org/stable/modules/generated/sklearn.preprocessing.StandardScaler.html)
    Air temperature: °C
    Relative humidity: %

    The measurements are part of the "Effects of heavy precipitation events on near-surface climate and particulate matter concentrations in Singapore". It is funded by the support from Humboldt-Universität zu Berlin for seed funding for collaborative projects between National University of Singapore and Humboldt-Universität zu Berlin.

  3. Perovskite Solar Cells Ageing Dataset

    • zenodo.org
    • data.niaid.nih.gov
    bin
    Updated Jul 26, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Noor Titan Putri Hartono; Noor Titan Putri Hartono; Hans Köbler; Hans Köbler; Paolo Graniero; Paolo Graniero; Mark Khenkin; Mark Khenkin; Rutger Schlatmann; Rutger Schlatmann; Carolin Ulbrich; Carolin Ulbrich; Antonio Abate; Antonio Abate (2023). Perovskite Solar Cells Ageing Dataset [Dataset]. http://doi.org/10.5281/zenodo.8185883
    Explore at:
    binAvailable download formats
    Dataset updated
    Jul 26, 2023
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Noor Titan Putri Hartono; Noor Titan Putri Hartono; Hans Köbler; Hans Köbler; Paolo Graniero; Paolo Graniero; Mark Khenkin; Mark Khenkin; Rutger Schlatmann; Rutger Schlatmann; Carolin Ulbrich; Carolin Ulbrich; Antonio Abate; Antonio Abate
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    This dataset contains cleaned 2,245 ageing test traces (time vs. MPPT PCE/ maximum power point tracking power conversion efficiency) for perovskite solar cells with various device stacks and architectures in the pickle (.pkl) format.

    The dataset can be loaded with the following commands on Python.

    import pickle5 as pickle
    import pandas as pd 
    import numpy as np
    
    with open('20230303_mySeriesDrop.pkl', "rb") as fh:
      mySeriesDrop = pickle.load(fh)

    The following command can be used to call a specific row (row 0) within the dataset.

    mySeriesDrop[0]

    The next steps to use the dataset is using scaling/ normalisation (for instance using sklearn.preprocessing.MaxAbsScaler) and smoothing (for instance using Savitzky-Golay filter).

    The code to run the complete analysis, including self-organising map clustering, can be accessed here: https://doi.org/10.5281/zenodo.8181602.

  4. 1200 pixels spectral datasets

    • zenodo.org
    zip
    Updated May 21, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Hui Zhang; Hui Zhang (2024). 1200 pixels spectral datasets [Dataset]. http://doi.org/10.5281/zenodo.11082600
    Explore at:
    zipAvailable download formats
    Dataset updated
    May 21, 2024
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Hui Zhang; Hui Zhang
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    In the Zip, spectral. npy was the average spectral data of red ginseng, mycotoxins and interference impurities, and label. npy was the corresponding label. Spectral data format was [1200,510] and label data format was [1200,1]. The example of data usage (sklearn in Python database was used to establish the classification model) was as follows:

    import numpy as np
    from sklearn. model_selection import train_test_split
    from sklearn. preprocessing import StandardScaler
    from sklearn. neighbors import KNeighborsClassifier
    from sklearn. metrics import classification_report, accuracy_score

    # Load spectral data and labels
    x = np.load('.../spectral.npy')[:,1:-1]
    y = np.load('.../label.npy')

    x_train, x_test, y_train, y_test = train_test_split(x, y, test_size=0.2, random_state=42)

    # Data standardization
    scaler = StandardScaler()
    x_train = scaler.fit_transform(x_train)
    x_test = scaler.transform(x_test)

    # Train the KNN model
    knn_model = KNeighborsClassifier(n_neighbors=5)
    knn_model. fit(x_train, y_train)

    # Predict
    y_pred = knn_model.predict(x_test)

    # Print classification reports and accuracy rates
    print("Classification Report:")
    print(classification_report(y_test, y_pred))
    print("Accuracy Score:")
    print(accuracy_score(y_test, y_pred))

  5. h

    Dataset-EfficientDrivingTimeDeterminationSystem

    • huggingface.co
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    ACHMAD AKBAR, Dataset-EfficientDrivingTimeDeterminationSystem [Dataset]. https://huggingface.co/datasets/jellysquish/Dataset-EfficientDrivingTimeDeterminationSystem
    Explore at:
    Authors
    ACHMAD AKBAR
    Description

    import re import pandas as pd from sklearn.tree import DecisionTreeClassifier from sklearn.preprocessing import LabelEncoder from google.colab import drive from sklearn.tree import export_text from sklearn.metrics import accuracy_score

      1. Mount Google Drive
    

    drive.mount('/content/drive')

      2. Baca file Excel
    

    file_path = '/content/drive/MyDrive/Colab Notebooks/AI_GACOR_Cleaned.xlsx' data = pd.read_excel(file_path)

      3. Encode kolom 'Hari'
    

    label_encoder_hari =… See the full description on the dataset page: https://huggingface.co/datasets/jellysquish/Dataset-EfficientDrivingTimeDeterminationSystem.

  6. Not seeing a result you expected?
    Learn how you can add new datasets to our index.

Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
Erik (2021). FastQuantileLayerKeras [Dataset]. https://www.kaggle.com/snippsy/fastquantilelayerkeras/tasks
Organization logo

FastQuantileLayerKeras

Equivalent to sklearn.preprocessing.QuantileTransform as Keras layer.

Explore at:
CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
Dataset updated
Jan 14, 2021
Dataset provided by
Kagglehttp://kaggle.com/
Authors
Erik
Description
Search
Clear search
Close search
Google apps
Main menu