79 datasets found
  1. Unlabeled Sentinel 2 time series dataset (training, T30TUVU):...

    • zenodo.org
    • data.niaid.nih.gov
    zip
    Updated Oct 3, 2024
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Iris Dumeur; Iris Dumeur; Silvia Valero; Silvia Valero; Jordi Inglada; Jordi Inglada (2024). Unlabeled Sentinel 2 time series dataset (training, T30TUVU): Self-Supervised Spatio-Temporal Representation Learning of Satellite Image Time Series [Dataset]. http://doi.org/10.5281/zenodo.7892410
    Explore at:
    zipAvailable download formats
    Dataset updated
    Oct 3, 2024
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Iris Dumeur; Iris Dumeur; Silvia Valero; Silvia Valero; Jordi Inglada; Jordi Inglada
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    This is a part of the unlabeled Sentinel 2 (S2) L2A dataset composed of patch time series acquired over France used to pretrain U-BARN. For further details, see section IV.A of the pre-print article "Self-Supervised Spatio-Temporal Representation Learning Of Satellite Image Time Series" available here. Each patch is constituted of the 10 bands [B2,B3,B4,B5,B6,B7,B8,B8A,B11,B12] and the three masks ['CLM_R1', 'EDG_R1', 'SAT_R1']. The global dataset is composed of two disjoint datasets: training (9 tiles) and validation dataset (4 tiles).

    In this repo, only data from the S2 tile T30UVU are available. To download the full pretraining dataset, see: 10.5281/zenodo.7891924

    Dataset nameS2 tilesROI sizeTemporal extent
    Train

    T30TXT,T30TYQ,T30TYS,T30UVU,

    T31TDJ,T31TDL,T31TFN,T31TGJ,T31UEP

    1024*10242018-2020
    ValT30TYR,T30UWU,T31TEK,T31UER256*2562016-2019
  2. f

    Classification performance of considered classifiers on the original...

    • plos.figshare.com
    xls
    Updated Sep 29, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Karina Shyrokykh; Max Girnyk; Lisa Dellmuth (2023). Classification performance of considered classifiers on the original collected dataset. [Dataset]. http://doi.org/10.1371/journal.pone.0290762.t002
    Explore at:
    xlsAvailable download formats
    Dataset updated
    Sep 29, 2023
    Dataset provided by
    PLOS ONE
    Authors
    Karina Shyrokykh; Max Girnyk; Lisa Dellmuth
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Classification performance of considered classifiers on the original collected dataset.

  3. Text Classification labeled and unlabeled datasets

    • kaggle.com
    zip
    Updated Jan 7, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Anna Jazayeri (2024). Text Classification labeled and unlabeled datasets [Dataset]. https://www.kaggle.com/datasets/annajazayeri/text-classification-labeled-and-unlabeled-datasets/suggestions
    Explore at:
    zip(27499 bytes)Available download formats
    Dataset updated
    Jan 7, 2024
    Authors
    Anna Jazayeri
    License

    MIT Licensehttps://opensource.org/licenses/MIT
    License information was derived automatically

    Description

    Dataset

    This dataset was created by Anna Jazayeri

    Released under MIT

    Contents

  4. T

    Unlabeled Pupil Recordings

    • dataverse.tdl.org
    tsv, zip
    Updated May 8, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Therese Koch; Therese Koch (2024). Unlabeled Pupil Recordings [Dataset]. http://doi.org/10.18738/T8/AH9GO3
    Explore at:
    zip(102074210), zip(514405971), zip(507492643), tsv(374), zip(500749363), zip(200919264), zip(56407250), zip(170066336), zip(12056751), zip(122287530), zip(345308852), zip(61865067), zip(43095677), zip(372808832), zip(154944749), zip(264459124), zip(119740673), zip(254282710), zip(105878036), zip(80091893), zip(415086453)Available download formats
    Dataset updated
    May 8, 2024
    Dataset provided by
    Texas Data Repository
    Authors
    Therese Koch; Therese Koch
    License

    CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
    License information was derived automatically

    Description

    This folder contains song recordings from typical adult zebra finches used to validate the song similarity method proposed in AVN. There is a variable number of song files from each bird. Most contain recordings from a single day of song, but for some birds who sang less frequently, recordings from multiple days were pooled, as indicated in the READ_ME.txt file in each bird's folder. These birds were all 'pupils' for the purpose of AVN validation. Their tutors are listed in Bird_list.csv, and recordings of their tutors are provided in the "Unlabeled Tutor Recordings" dataset within the AVN dataverse.

  5. birdclef-unlabeled-stft-mingroup

    • kaggle.com
    zip
    Updated May 3, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Kirill Chemrov (2024). birdclef-unlabeled-stft-mingroup [Dataset]. https://www.kaggle.com/datasets/chemrovkirill/birdclef-unlabeled-stft-mingroup
    Explore at:
    zip(0 bytes)Available download formats
    Dataset updated
    May 3, 2024
    Authors
    Kirill Chemrov
    Description

    Dataset

    This dataset was created by Kirill Chemrov

    Contents

  6. f

    Comparative analysis over various datasets.

    • plos.figshare.com
    xls
    Updated Jan 10, 2025
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Tao Yu; Wei Huang; Xin Tang; Duosi Zheng (2025). Comparative analysis over various datasets. [Dataset]. http://doi.org/10.1371/journal.pone.0316557.t003
    Explore at:
    xlsAvailable download formats
    Dataset updated
    Jan 10, 2025
    Dataset provided by
    PLOS ONE
    Authors
    Tao Yu; Wei Huang; Xin Tang; Duosi Zheng
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    In credit risk assessment, unsupervised classification techniques can be introduced to reduce human resource expenses and expedite decision-making. Despite the efficacy of unsupervised learning methods in handling unlabeled datasets, their performance remains limited owing to challenges such as imbalanced data, local optima, and parameter adjustment complexities. Thus, this paper introduces a novel hybrid unsupervised classification method, named the two-stage hybrid system with spectral clustering and semi-supervised support vector machine (TSC-SVM), which effectively addresses the unsupervised imbalance problem in credit risk assessment by targeting global optimal solutions. Furthermore, a multi-view combined unsupervised method is designed to thoroughly mine data and enhance the robustness of label predictions. This method mitigates discrepancies in prediction outcomes from three distinct perspectives. The effectiveness, efficiency, and robustness of the proposed TSC-SVM model are demonstrated through various real-world applications. The proposed algorithm is anticipated to expand the customer base for financial institutions while reducing economic losses.

  7. f

    The primers and unlabeled probes sequences.

    • plos.figshare.com
    xls
    Updated May 31, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Ta-Hsien Lee; Tzong-Shoon Wu; Ching-Ping Tseng; Jiantai Timothy Qiu (2023). The primers and unlabeled probes sequences. [Dataset]. http://doi.org/10.1371/journal.pone.0042051.t001
    Explore at:
    xlsAvailable download formats
    Dataset updated
    May 31, 2023
    Dataset provided by
    PLOS ONE
    Authors
    Ta-Hsien Lee; Tzong-Shoon Wu; Ching-Ping Tseng; Jiantai Timothy Qiu
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    The primers and unlabeled probes sequences.

  8. Data from: unlabeled

    • kaggle.com
    zip
    Updated Aug 22, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Elbadry2025 (2024). unlabeled [Dataset]. https://www.kaggle.com/datasets/elbadry2025/unlabeled
    Explore at:
    zip(29504770454 bytes)Available download formats
    Dataset updated
    Aug 22, 2024
    Authors
    Elbadry2025
    License

    MIT Licensehttps://opensource.org/licenses/MIT
    License information was derived automatically

    Description

    Dataset

    This dataset was created by Elbadry2025

    Released under MIT

    Contents

  9. f

    Unlabeled Cells.

    • figshare.com
    • plos.figshare.com
    xls
    Updated Jun 3, 2023
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Guido A. Zampighi; Lorenzo Zampighi; Salvatore Lanzavecchia (2023). Unlabeled Cells. [Dataset]. http://doi.org/10.1371/journal.pone.0023753.t002
    Explore at:
    xlsAvailable download formats
    Dataset updated
    Jun 3, 2023
    Dataset provided by
    PLOS ONE
    Authors
    Guido A. Zampighi; Lorenzo Zampighi; Salvatore Lanzavecchia
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    *Mean ± SD.#Gaussian Center ± HWHM (Half Width at Half Maximum).

  10. 4

    Data and code underlying the publication: DCAST: Diverse Class-Aware...

    • data.4tu.nl
    zip
    Updated Nov 25, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Yasin Tepeli; Joana Gonçalves (2024). Data and code underlying the publication: DCAST: Diverse Class-Aware Self-Training Mitigates Selection Bias for Fairer Learning [Dataset]. http://doi.org/10.4121/8648064e-aa7b-4a09-a755-7eb2d90bef66.v1
    Explore at:
    zipAvailable download formats
    Dataset updated
    Nov 25, 2024
    Dataset provided by
    4TU.ResearchData
    Authors
    Yasin Tepeli; Joana Gonçalves
    License

    https://www.gnu.org/licenses/gpl-3.0.htmlhttps://www.gnu.org/licenses/gpl-3.0.html

    Description

    This repository consists of Data/Code to reproduce the results of the thesis chapter "DCAST: Diverse Class-Aware Self-Training Mitigates Selection Bias for Fairer Learning".

    The data is shared at: https://doi.org/10.6084/m9.figshare.27003601

    The code is shared at: https://github.com/joanagoncalveslab/DCAST

  11. h

    autoeval-eval-conceptual_captions-unlabeled-ccbde0-1800162251

    • huggingface.co
    Updated Jun 9, 2023
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Evaluation on the Hub (2023). autoeval-eval-conceptual_captions-unlabeled-ccbde0-1800162251 [Dataset]. https://huggingface.co/datasets/autoevaluate/autoeval-eval-conceptual_captions-unlabeled-ccbde0-1800162251
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Jun 9, 2023
    Dataset authored and provided by
    Evaluation on the Hub
    Description

    Dataset Card for AutoTrain Evaluator

    This repository contains model predictions generated by AutoTrain for the following task and dataset:

    Task: Summarization Model: 0ys/mt5-small-finetuned-amazon-en-es Dataset: conceptual_captions Config: unlabeled Split: train

    To run new evaluation jobs, visit Hugging Face's automatic model evaluator.

      Contributions
    

    Thanks to @DonaldDaz for evaluating this model.

  12. f

    Setting of parameters.

    • plos.figshare.com
    xls
    Updated Oct 27, 2023
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Liang Chen; Caiming Zhong; Zehua Zhang (2023). Setting of parameters. [Dataset]. http://doi.org/10.1371/journal.pone.0292960.t002
    Explore at:
    xlsAvailable download formats
    Dataset updated
    Oct 27, 2023
    Dataset provided by
    PLOS ONE
    Authors
    Liang Chen; Caiming Zhong; Zehua Zhang
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Clustering is an unsupervised machine learning technique whose goal is to cluster unlabeled data. But traditional clustering methods only output a set of results and do not provide any explanations of the results. Although in the literature a number of methods based on decision tree have been proposed to explain the clustering results, most of them have some disadvantages, such as too many branches and too deep leaves, which lead to complex explanations and make it difficult for users to understand. In this paper, a hypercube overlay model based on multi-objective optimization is proposed to achieve succinct explanations of clustering results. The model designs two objective functions based on the number of hypercubes and the compactness of instances and then uses multi-objective optimization to find a set of nondominated solutions. Finally, an Utopia point is defined to determine the most suitable solution, in which each cluster can be covered by as few hypercubes as possible. Based on these hypercubes, an explanations of each cluster is provided. Upon verification on synthetic and real datasets respectively, it shows that the model can provide a concise and understandable explanations to users.

  13. lstm-boxing-dataset(unlabeled)

    • kaggle.com
    zip
    Updated Dec 10, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    YasserOuzzine (2024). lstm-boxing-dataset(unlabeled) [Dataset]. https://www.kaggle.com/datasets/yasserouzzine2/lstm-boxing-datasetunlabeled/code
    Explore at:
    zip(231523106 bytes)Available download formats
    Dataset updated
    Dec 10, 2024
    Authors
    YasserOuzzine
    Description

    Dataset

    This dataset was created by YasserOuzzine

    Contents

  14. Z

    Unlabeled Sentinel 2 time series dataset (training, T30TYS): Self-Supervised...

    • data.niaid.nih.gov
    • explore.openaire.eu
    • +1more
    Updated May 11, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Iris Dumeur (2023). Unlabeled Sentinel 2 time series dataset (training, T30TYS): Self-Supervised Spatio-Temporal Representation Learning of Satellite Image Time Series [Dataset]. https://data.niaid.nih.gov/resources?id=ZENODO_7924192
    Explore at:
    Dataset updated
    May 11, 2023
    Dataset provided by
    Silvia Valero
    Jordi Inglada
    Iris Dumeur
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    This is a part of the unlabeled Sentinel 2 (S2) L2A dataset composed of patch time series acquired over France used to pretrain U-BARN. For further details, see section IV.A of the pre-print article "Self-Supervised Spatio-Temporal Representation Learning Of Satellite Image Time Series" available here. Each patch is constituted of the 10 bands [B2,B3,B4,B5,B6,B7,B8,B8A,B11,B12] and the three masks ['CLM_R1', 'EDG_R1', 'SAT_R1']. The global dataset is composed of two disjoint datasets: training (9 tiles) and validation dataset (4 tiles).

    In this repo, only data from the S2 tile T30TYS are available. To download the full pretraining dataset, see: 10.5281/zenodo.7891924

    Global unlabeled dataset description
    
    
        Dataset name
        S2 tiles
        ROI size
        Temporal extent
    
    
        Train
    

    T30TXT,T30TYQ,T30TYS,T30UVU,

    T31TDJ,T31TDL,T31TFN,T31TGJ,T31UEP

        1024*1024
        2018-2020
    
    
        Val
        T30TYR,T30UWU,T31TEK,T31UER
        256*256
        2016-2019
    
  15. h

    VietMed_unlabeled

    • huggingface.co
    Updated Oct 6, 2024
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Phan Tuấn Anh (2024). VietMed_unlabeled [Dataset]. https://huggingface.co/datasets/doof-ferb/VietMed_unlabeled
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Oct 6, 2024
    Authors
    Phan Tuấn Anh
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    unofficial mirror of VietMed (Vietnamese speech data in medical domain) unlabeled set

    official announcement: https://arxiv.org/abs/2404.05659 official download: https://huggingface.co/datasets/leduckhai/VietMed this repo contains the unlabeled set: 966h - 230k samples i also gather the metadata: see info.csv my extraction code: https://github.com/phineas-pta/fine-tune-whisper-vi/blob/main/misc/vietmed-unlabeled.py need to do: check misspelling, restore foreign words phonetised… See the full description on the dataset page: https://huggingface.co/datasets/doof-ferb/VietMed_unlabeled.

  16. e

    Unlabeled Point Lobos slides

    • knb.ecoinformatics.org
    • search.dataone.org
    • +1more
    Updated Dec 17, 2014
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Landels-Hill Big Creek Reserve; University of California Natural Reserve System; Chuck Bancroft (2014). Unlabeled Point Lobos slides [Dataset]. http://doi.org/10.5063/AA/nrs.596.1
    Explore at:
    Dataset updated
    Dec 17, 2014
    Dataset provided by
    Knowledge Network for Biocomplexity
    Authors
    Landels-Hill Big Creek Reserve; University of California Natural Reserve System; Chuck Bancroft
    Time period covered
    Dec 1, 1958 - Feb 1, 1995
    Area covered
    Description

    In first 2 binders: Slides of landscapes from Point Lobos State Reserve, Garrapata State Park, and unlabeled sites on the Central Coast

  17. Unlabeled Generative Dog Images

    • kaggle.com
    zip
    Updated Jun 4, 2021
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Ching-Yuan Bai (2021). Unlabeled Generative Dog Images [Dataset]. https://www.kaggle.com/andrewcybai/unlabeled-generative-dog-images
    Explore at:
    zip(0 bytes)Available download formats
    Dataset updated
    Jun 4, 2021
    Authors
    Ching-Yuan Bai
    License

    Attribution-ShareAlike 4.0 (CC BY-SA 4.0)https://creativecommons.org/licenses/by-sa/4.0/
    License information was derived automatically

    Description

    Dataset

    This dataset was created by Ching-Yuan Bai

    Released under CC BY-SA 4.0

    Contents

  18. Micro-ECoG, unlabeled 2s windows

    • figshare.com
    • search.datacite.org
    bin
    Updated Jan 25, 2019
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Nicholas Rogers (2019). Micro-ECoG, unlabeled 2s windows [Dataset]. http://doi.org/10.6084/m9.figshare.7633418.v4
    Explore at:
    binAvailable download formats
    Dataset updated
    Jan 25, 2019
    Dataset provided by
    figshare
    Figsharehttp://figshare.com/
    Authors
    Nicholas Rogers
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Submillimeter grid ECoG recorded in human and mouse. Data in 2.0 s windows, unlabeled. The electrodes were arranged in a square grid with spacing specified in each file. The arrangement of the channels on the grid is in the "grid" variable allowing each channel to be mapped relative to the others.Details regarding methods and use of the data available in the linked publication.

  19. Data from: SPATIALLY ADAPTIVE SEMI-SUPERVISED LEARNING WITH GAUSSIAN...

    • data.staging.idas-ds1.appdat.jsc.nasa.gov
    • data.nasa.gov
    • +1more
    Updated Feb 19, 2025
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    data.staging.idas-ds1.appdat.jsc.nasa.gov (2025). SPATIALLY ADAPTIVE SEMI-SUPERVISED LEARNING WITH GAUSSIAN PROCESSES FOR HYPERSPECTRAL DATA ANALYSIS [Dataset]. https://data.staging.idas-ds1.appdat.jsc.nasa.gov/dataset/spatially-adaptive-semi-supervised-learning-with-gaussian-processes-for-hyperspectral-data
    Explore at:
    Dataset updated
    Feb 19, 2025
    Dataset provided by
    NASAhttp://nasa.gov/
    Description

    SPATIALLY ADAPTIVE SEMI-SUPERVISED LEARNING WITH GAUSSIAN PROCESSES FOR HYPERSPECTRAL DATA ANALYSIS GOO JUN * AND JOYDEEP GHOSH* Abstract. A semi-supervised learning algorithm for the classification of hyperspectral data, Gaussian process expectation maximization (GP-EM), is proposed. Model parameters for each land cover class is first estimated by a supervised algorithm using Gaussian process regressions to find spatially adaptive parameters, and the estimated parameters are then used to initialize a spatially adaptive mixture-of-Gaussians model. The mixture model is updated by expectationmaximization iterations using the unlabeled data, and the spatially adaptive parameters for unlabeled instances are obtained by Gaussian process regressions with soft assignments. Two sets of hyperspectral data taken from the Botswana area by the NASA EO-1 satellite are used for experiments. Empirical evaluations show that the proposed framework performs significantly better than baseline algorithms that do not use spatial information, and the results are also better than any previously reported results by other algorithms on the same data.

  20. o

    Data from: Amos: A large-scale abdominal multi-organ benchmark for versatile...

    • explore.openaire.eu
    • zenodo.org
    Updated Nov 28, 2022
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    JI YUANFENG (2022). Amos: A large-scale abdominal multi-organ benchmark for versatile medical image segmentation [Dataset]. http://doi.org/10.5281/zenodo.7155724
    Explore at:
    Dataset updated
    Nov 28, 2022
    Authors
    JI YUANFENG
    Description

    Despite the considerable progress in automatic abdominal multi-organ segmentation from CT/MRI scans in recent years, a comprehensive evaluation of the models' capabilities is hampered by the lack of a large-scale benchmark from diverse clinical scenarios. Constraint by the high cost of collecting and labeling 3D medical data, most of the deep learning models to date are driven by datasets with a limited number of organs of interest or samples, which still limits the power of modern deep models and makes it difficult to provide a fully comprehensive and fair estimate of various methods. To mitigate the limitations, we present AMOS, a large-scale, diverse, clinical dataset for abdominal organ segmentation. AMOS provides 500 CT and 100 MRI scans collected from multi-center, multi-vendor, multi-modality, multi-phase, multi-disease patients, each with voxel-level annotations of 15 abdominal organs, providing challenging examples and test-bed for studying robust segmentation algorithms under diverse targets and scenarios. We further benchmark several state-of-the-art medical segmentation models to evaluate the status of the existing methods on this new challenging dataset. We have made our datasets, benchmark servers, and baselines publicly available, and hope to inspire future research. The paper can be found at https://arxiv.org/pdf/2206.08023.pdf In addition to providing the labeled 600 CT and MRI scans, we expect to provide 2000 CT and 1200 MRI scans without labels to support more learning tasks (semi-supervised, un-supervised, domain adaption, ...). The link can be found in: labeled data (500CT+100MRI) unlabeled data Part I (900CT) unlabeled data Part II (1100CT) (Now there are 1000CT, we will replenish to 1100CT) unlabeled data Part III (1200MRI) if you found this dataset useful for your research, please cite: @inproceedings{NEURIPS2022_ee604e1b, author = {Ji, Yuanfeng and Bai, Haotian and GE, Chongjian and Yang, Jie and Zhu, Ye and Zhang, Ruimao and Li, Zhen and Zhanng, Lingyan and Ma, Wanling and Wan, Xiang and Luo, Ping}, booktitle = {Advances in Neural Information Processing Systems}, editor = {S. Koyejo and S. Mohamed and A. Agarwal and D. Belgrave and K. Cho and A. Oh}, pages = {36722--36732}, publisher = {Curran Associates, Inc.}, title = {AMOS: A Large-Scale Abdominal Multi-Organ Benchmark for Versatile Medical Image Segmentation}, url = {https://proceedings.neurips.cc/paper_files/paper/2022/file/ee604e1bedbd069d9fc9328b7b9584be-Paper-Datasets_and_Benchmarks.pdf}, volume = {35}, year = {2022} }

Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
Iris Dumeur; Iris Dumeur; Silvia Valero; Silvia Valero; Jordi Inglada; Jordi Inglada (2024). Unlabeled Sentinel 2 time series dataset (training, T30TUVU): Self-Supervised Spatio-Temporal Representation Learning of Satellite Image Time Series [Dataset]. http://doi.org/10.5281/zenodo.7892410
Organization logo

Unlabeled Sentinel 2 time series dataset (training, T30TUVU): Self-Supervised Spatio-Temporal Representation Learning of Satellite Image Time Series

Explore at:
zipAvailable download formats
Dataset updated
Oct 3, 2024
Dataset provided by
Zenodohttp://zenodo.org/
Authors
Iris Dumeur; Iris Dumeur; Silvia Valero; Silvia Valero; Jordi Inglada; Jordi Inglada
License

Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically

Description

This is a part of the unlabeled Sentinel 2 (S2) L2A dataset composed of patch time series acquired over France used to pretrain U-BARN. For further details, see section IV.A of the pre-print article "Self-Supervised Spatio-Temporal Representation Learning Of Satellite Image Time Series" available here. Each patch is constituted of the 10 bands [B2,B3,B4,B5,B6,B7,B8,B8A,B11,B12] and the three masks ['CLM_R1', 'EDG_R1', 'SAT_R1']. The global dataset is composed of two disjoint datasets: training (9 tiles) and validation dataset (4 tiles).

In this repo, only data from the S2 tile T30UVU are available. To download the full pretraining dataset, see: 10.5281/zenodo.7891924

Dataset nameS2 tilesROI sizeTemporal extent
Train

T30TXT,T30TYQ,T30TYS,T30UVU,

T31TDJ,T31TDL,T31TFN,T31TGJ,T31UEP

1024*10242018-2020
ValT30TYR,T30UWU,T31TEK,T31UER256*2562016-2019
Search
Clear search
Close search
Google apps
Main menu