100+ datasets found
  1. m

    COVID-19 & Normal CT Segmentation Dataset

    • data.mendeley.com
    Updated Nov 27, 2023
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Arvin Arian (2023). COVID-19 & Normal CT Segmentation Dataset [Dataset]. http://doi.org/10.17632/pfmgfpwnmm.2
    Explore at:
    Dataset updated
    Nov 27, 2023
    Authors
    Arvin Arian
    License

    Attribution-NonCommercial 3.0 (CC BY-NC 3.0)https://creativecommons.org/licenses/by-nc/3.0/
    License information was derived automatically

    Description

    This dataset includes CT data and segmentation masks from patients diagnosed with COVID-19, as well as data from subjects without the infection.

    This study is approved under the ethical approval codes of IR.TUMS.IKHC.REC.1399.255 and IR.TUMS.VCR.REC.1399.488 at Tehran University of Medical Sciences.

    The code for loading the dataset and running an AI model is available on: https://github.com/SamanSotoudeh/COVID19-segmentation

    Please use the following citations:

    1- Arian, Arvin; Mehrabinejad, Mohammad-Mehdi; Zoorpaikar, Mostafa; Hasanzadeh, Navid; Sotoudeh-Paima, Saman; Kolahi, Shahriar; Gity, Masoumeh; Soltanian-Zadeh, "Accuracy of Artificial Intelligence CT Quantification in Predicting COVID-19 Subjects’ Prognosis" PLoS ONE (2023).

    2- Sotoudeh-Paima, Saman, et al. "A Multi-centric Evaluation of Deep Learning Models for Segmentation of COVID-19 Lung Lesions on Chest CT Scans." Iranian Journal of Radiology 19.4 (2022).

    3- Hasanzadeh, Navid, et al. "Segmentation of COVID-19 Infections on CT: Comparison of four UNet-based networks." 2020 27th National and 5th International Iranian Conference on Biomedical Engineering (ICBME). IEEE, 2020.

  2. s

    Common Objects Segmentation Dataset

    • hmn.shaip.com
    • sv.shaip.com
    • +6more
    json
    Updated Dec 25, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Shaip (2024). Common Objects Segmentation Dataset [Dataset]. https://hmn.shaip.com/offerings/specific-object-contour-segmentation-datasets/
    Explore at:
    jsonAvailable download formats
    Dataset updated
    Dec 25, 2024
    Dataset authored and provided by
    Shaip
    License

    CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
    License information was derived automatically

    Description

    Cov Khoom Siv Feem Ntau Segmentation Dataset ua haujlwm rau kev lag luam e-lag luam thiab kev lom zem pom kev lag luam nrog ntau cov duab sau hauv internet, muaj cov kev daws teeb meem xws li 800 × 600 txog 4160 × 3120. Cov ntaub ntawv no suav nrog ntau qhov sib txawv ntawm cov xwm txheej niaj hnub thiab cov khoom, suav nrog ntau tus neeg, tsiaj txhu thiab cov rooj tog zaum. segmentation.

  3. KSSD2025 - Kidney Stone Segmentation Dataset

    • kaggle.com
    zip
    Updated Apr 14, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Murillo Bouzon (2025). KSSD2025 - Kidney Stone Segmentation Dataset [Dataset]. https://www.kaggle.com/datasets/murillobouzon/kssd2025-kidney-stone-segmentation-dataset
    Explore at:
    zip(82320796 bytes)Available download formats
    Dataset updated
    Apr 14, 2025
    Authors
    Murillo Bouzon
    Description

    KSSD2025 – CT Kidney Stone Segmentation Dataset

    A High-Quality Annotated Dataset for Deep Learning-Based Kidney Stone Segmentation

    📌 Overview

    KSSD2025 is a dataset of axial CT images with expert-annotated kidney stone segmentation masks, created to support deep learning research in medical image segmentation. It is derived from the public dataset by Islam et al. (2022), which contains CT images with different kidney conditions. KSSD2025 focuses exclusively on kidney stone cases, offering precise ground-truth masks for developing and benchmarking AI-based segmentation models.

    🎈 Description

    This dataset presents a carefully refined subset of the original "CT Kidney Dataset: Normal-Cyst-Tumor and Stone" by Islam et al., comprising only axial CT images that exhibit kidney stones. Out of 12,446 images in the original collection, 838 images were selected for manual annotation based on the presence of stones and the axial orientation, which offers better anatomical context for segmentation tasks.

    To ensure high-quality ground-truth segmentation, a three-step preprocessing pipeline was applied:

    1) Thresholding: Pixel intensity thresholding at 150 was used to isolate high-density structures, which often correspond to kidney stones.
    
    2) Connected Component Filtering: Regions larger than 300 pixels were discarded to remove bones and other non-stone structures.
    
    3) Manual Refinement: Remaining artifacts were removed and stone regions refined in collaboration with specialists in urology and radiology.
    

    Each image in the dataset is paired with a binary mask that precisely delineates kidney stone regions, making it ideal for training and evaluating deep learning models in tasks like medical image segmentation and object detection.

    📊 Dataset Details Total Annotated Images: 838 View: Axial Annotations: Binary segmentation masks (kidney stone regions) Image Format: TIF Size: 305.38 MB Source Dataset: CT KIDNEY DATASET: Normal-Cyst-Tumor and Stone Annotation Method: Semi-automatic (thresholding + connected components) followed by expert manual refinement

    🔍 Use Cases ✔️ Deep Learning-Based Kidney Stone Segmentation ✔️ AI-Powered Medical Imaging Tools ✔️ Benchmarking Medical Image Segmentation Models ✔️ Educational Applications in Radiology and Urology

    🔬 Research Potential

    KSSD2025 addresses the scarcity of annotated kidney stone segmentation datasets. By offering pixel-level annotations, it opens new opportunities for developing robust segmentation models and AI-assisted diagnostic systems in urology.

    ⚖️ License

    Datafiles © Nazmul Islam

    🏫 Institutions Involved

    • Centro Universitário FEI
    • Hospital Universitário da Universidade de São Paulo Based on original dataset by: Islam MN, Hasan M, Hossain M, Alam M, Rabiul G, Uddin MZ, Soylu A. Vision transformer and explainable transfer learning models for auto detection of kidney cyst, stone and tumor from CT-radiography. Scientific Reports. 2022.

    📢 Citation

    If you use this dataset in your research, please cite:

    Islam MN, Hasan M, Hossain M, Alam M, Rabiul G, Uddin MZ, Soylu A. Vision transformer and explainable transfer learning models for auto detection of kidney cyst, stone and tumor from CT-radiography. Scientific Reports. 2022.

    M. F. Bouzon et al., "KSSD2025: A New Annotated Dataset for Automatic Kidney Stone Segmentation and Evaluation with Modified U-Net Based Deep Learning Models," in IEEE Access, doi: 10.1109/ACCESS.2025.3610027

    🙏 If you find this dataset helpful, please give it an upvote and share your feedback. Thank you! 😊

  4. R

    Normal Multi Aperio Semantic Dataset

    • universe.roboflow.com
    zip
    Updated Apr 12, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    FYPSEMANTIC (2023). Normal Multi Aperio Semantic Dataset [Dataset]. https://universe.roboflow.com/fypsemantic/normal-multi-aperio-semantic
    Explore at:
    zipAvailable download formats
    Dataset updated
    Apr 12, 2023
    Dataset authored and provided by
    FYPSEMANTIC
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Variables measured
    Mitotic Cells Masks
    Description

    Normal Multi Aperio Semantic

    ## Overview
    
    Normal Multi Aperio Semantic is a dataset for semantic segmentation tasks - it contains Mitotic Cells annotations for 865 images.
    
    ## Getting Started
    
    You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
    
      ## License
    
      This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).
    
  5. BUS-UCLM: Breast Ultrasound Dataset

    • kaggle.com
    zip
    Updated Jul 6, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Orvile (2025). BUS-UCLM: Breast Ultrasound Dataset [Dataset]. https://www.kaggle.com/datasets/orvile/bus-uclm-breast-ultrasound-dataset/suggestions
    Explore at:
    zip(672995684 bytes)Available download formats
    Dataset updated
    Jul 6, 2025
    Authors
    Orvile
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    BUS-UCLM: Breast Ultrasound Lesion Segmentation Dataset 🎗️

    Overview 🔬

    The BUS-UCLM dataset is a collection of breast ultrasound images from 38 patients, specifically designed for lesion segmentation research. It comprises a total of 683 images categorized into benign (174), malignant (90), and normal (419) cases. The images were acquired using a Siemens ACUSON S2000TM Ultrasound System between 2022 and 2023. Ground truth segmentation masks are provided as separate RGB files.

    Dataset Contents 📁

    The dataset contains:

    • Images: 683 breast ultrasound images with a mix of benign, malignant, and normal cases.
    • Segmentation Masks: Corresponding RGB segmentation masks for each image. Green indicates benign lesions, red indicates malignant lesions, and black represents the background or normal breast tissue.

    Annotation Details 📌

    The ground truth for lesion segmentation is provided in separate files as RGB images. The color coding is as follows:

    • Green: Benign lesions
    • Red: Malignant lesions
    • Black: Background or normal breast tissue

    These annotations were created by expert radiologists, ensuring high-quality ground truth for training and evaluation.

    Potential Uses 💡

    This dataset is a valuable resource for research in:

    • Breast Cancer Diagnosis: Developing and evaluating models for distinguishing between benign, malignant, and normal breast tissue in ultrasound images. 🧠
    • Lesion Detection: Training algorithms to automatically detect the presence of lesions in breast ultrasound images. 👀
    • Medical Image Segmentation: Creating and testing segmentation models to precisely delineate lesion boundaries. ✂️
    • Health Care Applications: Contributing to the development of computer-aided diagnosis systems for improved breast cancer screening and diagnosis. 👩‍⚕️
    • Computer Vision: Providing a real-world medical imaging dataset for advancing computer vision techniques. 🖼️

    Citation ✍️

    Please cite this dataset as follows:

    Vallez, Noelia; Bueno, Gloria; Deniz, Oscar; Rienda, Miguel Angel; Pastor, Carlos (2024), “BUS-UCLM: Breast ultrasound lesion segmentation dataset”, Mendeley Data, V1, doi: 10.17632/7fvgj4jsp7.1

    Contributors 🧑‍🤝‍🧑

    • Noelia Vallez
    • Gloria Bueno
    • Oscar Deniz
    • Miguel Angel Rienda
    • Carlos Pastor

    Institutions 🏢

    • Universidad de Castilla-La Mancha
    • Hospital General de Ciudad Real

    Categories 🏷️

    Breast Cancer, Image Segmentation, Object Detection, Ultrasound, Breast Ultrasonography, Instance Segmentation

    Funding 💰

    This research was supported by:

    • Ministerio de Ciencia, Innovación y Universidades (PID2021-127567NB-I00)
    • European Union NextGenerationEU/PRTR

    License 📜

    This dataset is licensed under the Creative Commons Attribution 4.0 International License (CC BY 4.0). This license allows for the sharing and adaptation of the material for any purpose, even commercially, as long as appropriate credit is given to the authors.

    Please consider upvoting this dataset if you find it useful! 👍

  6. R

    Kidney Normal Dataset

    • universe.roboflow.com
    zip
    Updated Feb 23, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    KidneyDisease (2025). Kidney Normal Dataset [Dataset]. https://universe.roboflow.com/kidneydisease/kidney-normal
    Explore at:
    zipAvailable download formats
    Dataset updated
    Feb 23, 2025
    Dataset authored and provided by
    KidneyDisease
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Variables measured
    Kidney Polygons
    Description

    Kidney Normal

    ## Overview
    
    Kidney Normal is a dataset for instance segmentation tasks - it contains Kidney annotations for 1,025 images.
    
    ## Getting Started
    
    You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
    
      ## License
    
      This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).
    
  7. z

    The Colorectal_Cancer_IHC_CISH_HE_Epithelium_Segmentation dataset

    • zenodo.org
    • dataverse.azure.uit.no
    • +1more
    txt
    Updated Feb 4, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Henrik Sahlin Pettersen; Henrik Sahlin Pettersen; Erik Nesje Wiik; Erik Nesje Wiik (2025). The Colorectal_Cancer_IHC_CISH_HE_Epithelium_Segmentation dataset [Dataset]. http://doi.org/10.18710/digqgq
    Explore at:
    txtAvailable download formats
    Dataset updated
    Feb 4, 2025
    Dataset provided by
    Zenodo
    Authors
    Henrik Sahlin Pettersen; Henrik Sahlin Pettersen; Erik Nesje Wiik; Erik Nesje Wiik
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Dataset Overview

    One of the largest (colorectal) epithelium segmentation datasets to date, featuring H&E and IHC images with pathologist-annotated segmentations. (full dataset can be downloaded at https://dataverse.no/dataset.xhtml?persistentId=doi:10.18710/DIGQGQ).

    With over 500,000 epithelium annotations, it represents a significant advance in scale:

    Dataset Contents

    The collection includes Tissue Microarray (TMA) cores from 100 patients, featuring both normal colorectal mucosa and cancer tissue. Each patient has:

    • 3 normal colorectal mucosa cores
    • 3 colorectal cancer cores
    • Each core includes original image and corresponding epithelium segmentation mask

    Imaging Markers

    All images are approximately 10,000 x 10,000 pixels at 40X magnification. Each marker includes matched pairs of JPG images and PNG segmentation masks:

    1. H&E stained cores (552 image/mask pairs)

    2. Immunohistochemistry for 13 proteins:

      • E-Cadherin (558 pairs)
      • Vimentin (559 pairs)
      • Smooth Muscle Actin (SMA) (558 pairs)
      • Ki-67 (557 pairs)
      • SMAD3 (546 pairs)
      • MACC1 (531 pairs)
      • LASP1 (535 pairs)
      • CD44 (526 pairs)
      • NAIP (553 pairs)
      • KLF5 (553 pairs)
      • FSCN1 (553 pairs)
      • CTNND1 (556 pairs)
      • KRAS (547 pairs)
    3. ISH stains:

      • miR-143 (2,195 pairs)
      • miR-145 (2,189 pairs)
      • ISH Positive Control (U6 snRNA) (555 pairs)
      • ISH Negative Controls (Scrambled probe) (556 pairs)

    Total: 13,179 image/mask pairs (26,358 files)

    Data Format

    • High-resolution histopathological images
    • Full epithelium segmentation masks for both normal and cancer tissue
    • Quantitative measurements (SPSS datafile) for selected markers
    • All annotations validated by pathologists

    Applications

    Ideal for AI training in:

    • Colorectal cancer analysis
    • Epithelium segmentation
    • Digital pathology
    • Deep learning-based image analysis

    Dataset use must adhere to the CC0 licence and cited as: Pettersen, Henrik Sahlin; Wiik, Erik Nesje, 2025, "The Colorectal_Cancer_IHC_CISH_HE_Epithelium_Segmentation dataset", https://doi.org/10.18710/DIGQGQ, DataverseNO, V1

  8. DUT Sewer3D Semantic Segmentation (S3DSS) Dataset

    • kaggle.com
    zip
    Updated Mar 8, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    liminghao123 (2024). DUT Sewer3D Semantic Segmentation (S3DSS) Dataset [Dataset]. https://www.kaggle.com/datasets/liminghao123/dut-sewer3d-semantic-segmentation-s3dss-dataset
    Explore at:
    zip(19982492968 bytes)Available download formats
    Dataset updated
    Mar 8, 2024
    Authors
    liminghao123
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Welcome to DUT Sewer3D Semantic Segmentation Dataset !

    Motivation and Background

    Urban sewer pipelines, as the critical guarantors of urban resilience and sustainable development, undertake the task of sewage disposal and flood prevention. However, in many countries, the most municipal sewer systems have been in service for 60 to 100 years, with the worst condition rating (D+) evaluated by ASCE.

    As laser scanning is fast becoming the state-of-the-art inspection technique for underground sewers, semantic segmentation of pipeline point clouds is an essential intermediate step for pipeline condition assessment and digital twinning. Currently, similar to other building structures, the scarcity of real-world point clouds has hindered the application of deep learning techniques for automated sewer pipeline semantic segmentation.

    Goal

    We provided a high-quality, realistic, semantically-rich public dataset named "**Sewer3D Semantic Segmentation**" (S3DSS), including 800 synthetic scans and 500 real-world scans, for point cloud semantic segmentation in sewer pipeline domain, for which there are no public datasets in the past. S3DSS contains over 917 million points with 8 categories of common sewer defects. We hope it can be a starting point for benchmarking developed approaches to promote deep learning research on point cloud of sewer pipeline defects.

    Content and Annotations

    The two sub-datasets were obtained in the following way.

    The real point cloud data were captured in laboratory scenarios using a FARO Focus S laser scanner. We used two prototype reinforced concrete sewer pipes to create most of the defect scenes. However, for misalign and displace defects that are difficult to operate with concrete pipes, we used two steel pipes which were well-designed to simulate. A total of 500 real scans were collected.

    The synthetic point cloud data were obtained by our automated synthetic data generator in Unity3D. The introduction to the synthetic point cloud data generation methodology can be found in our paper. We generated 800 scans of sewer defect scenes. If you need more data, please contact Minghao Li (liminghao@dlut.edu.cn). In S3DSS, 8 common defect classes are used which includes:

    • Normal
    • Spalling
    • Blockage
    • Corrosion
    • Misalignment
    • Deposit
    • Displacement
    • Rubber Ring

    Acknowledgements

    This work was supported by the National Key R & D Program of China (Grant No. 2022YFC3801000) and the National Natural Science Foundation of China (Grant No. 52479118). We also thank Haurum et al. for sharing their great work "Sewer Defect Classification using Synthetic Point Clouds" as a reference for this work.

    Citation

    • Please cite this paper if you find this dataset useful:

    【M. Li, X. Feng, Z. Wu, J. Bai, F. Yang, Game engine-driven synthetic point cloud generation method for LiDAR-based defect detection in sewers, Tunnelling and Underground Space Technology 163 (2025) 106755. https://doi.org/10.1016/j.tust.2025.106755.】

    【Z. Wu, M. Li, Y. Han, X. Feng, Semantic segmentation of 3D point cloud for sewer defect detection using an integrated global and local deep learning network, Measurement 253 (2025) 117434. https://doi.org/10.1016/j.measurement.2025.117434.】

    • Previous Publications:

    M. Li, X. Feng, Q. Hu, 3D laser point cloud-based geometric digital twin for condition assessment of large diameter pipelines. Tunnelling and Underground Space Technology 142 (2023) 105430.

  9. Data from: FISBe: A real-world benchmark dataset for instance segmentation...

    • zenodo.org
    • data.niaid.nih.gov
    • +1more
    bin, json +3
    Updated Apr 2, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Lisa Mais; Lisa Mais; Peter Hirsch; Peter Hirsch; Claire Managan; Claire Managan; Ramya Kandarpa; Josef Lorenz Rumberger; Josef Lorenz Rumberger; Annika Reinke; Annika Reinke; Lena Maier-Hein; Lena Maier-Hein; Gudrun Ihrke; Gudrun Ihrke; Dagmar Kainmueller; Dagmar Kainmueller; Ramya Kandarpa (2024). FISBe: A real-world benchmark dataset for instance segmentation of long-range thin filamentous structures [Dataset]. http://doi.org/10.5281/zenodo.10875063
    Explore at:
    zip, text/x-python, bin, json, txtAvailable download formats
    Dataset updated
    Apr 2, 2024
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Lisa Mais; Lisa Mais; Peter Hirsch; Peter Hirsch; Claire Managan; Claire Managan; Ramya Kandarpa; Josef Lorenz Rumberger; Josef Lorenz Rumberger; Annika Reinke; Annika Reinke; Lena Maier-Hein; Lena Maier-Hein; Gudrun Ihrke; Gudrun Ihrke; Dagmar Kainmueller; Dagmar Kainmueller; Ramya Kandarpa
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Time period covered
    Feb 26, 2024
    Description

    General

    For more details and the most up-to-date information please consult our project page: https://kainmueller-lab.github.io/fisbe.

    Summary

    • A new dataset for neuron instance segmentation in 3d multicolor light microscopy data of fruit fly brains
      • 30 completely labeled (segmented) images
      • 71 partly labeled images
      • altogether comprising ∼600 expert-labeled neuron instances (labeling a single neuron takes between 30-60 min on average, yet a difficult one can take up to 4 hours)
    • To the best of our knowledge, the first real-world benchmark dataset for instance segmentation of long thin filamentous objects
    • A set of metrics and a novel ranking score for respective meaningful method benchmarking
    • An evaluation of three baseline methods in terms of the above metrics and score

    Abstract

    Instance segmentation of neurons in volumetric light microscopy images of nervous systems enables groundbreaking research in neuroscience by facilitating joint functional and morphological analyses of neural circuits at cellular resolution. Yet said multi-neuron light microscopy data exhibits extremely challenging properties for the task of instance segmentation: Individual neurons have long-ranging, thin filamentous and widely branching morphologies, multiple neurons are tightly inter-weaved, and partial volume effects, uneven illumination and noise inherent to light microscopy severely impede local disentangling as well as long-range tracing of individual neurons. These properties reflect a current key challenge in machine learning research, namely to effectively capture long-range dependencies in the data. While respective methodological research is buzzing, to date methods are typically benchmarked on synthetic datasets. To address this gap, we release the FlyLight Instance Segmentation Benchmark (FISBe) dataset, the first publicly available multi-neuron light microscopy dataset with pixel-wise annotations. In addition, we define a set of instance segmentation metrics for benchmarking that we designed to be meaningful with regard to downstream analyses. Lastly, we provide three baselines to kick off a competition that we envision to both advance the field of machine learning regarding methodology for capturing long-range data dependencies, and facilitate scientific discovery in basic neuroscience.

    Dataset documentation:

    We provide a detailed documentation of our dataset, following the Datasheet for Datasets questionnaire:

    >> FISBe Datasheet

    Our dataset originates from the FlyLight project, where the authors released a large image collection of nervous systems of ~74,000 flies, available for download under CC BY 4.0 license.

    Files

    • fisbe_v1.0_{completely,partly}.zip
      • contains the image and ground truth segmentation data; there is one zarr file per sample, see below for more information on how to access zarr files.
    • fisbe_v1.0_mips.zip
      • maximum intensity projections of all samples, for convenience.
    • sample_list_per_split.txt
      • a simple list of all samples and the subset they are in, for convenience.
    • view_data.py
      • a simple python script to visualize samples, see below for more information on how to use it.
    • dim_neurons_val_and_test_sets.json
      • a list of instance ids per sample that are considered to be of low intensity/dim; can be used for extended evaluation.
    • Readme.md
      • general information

    How to work with the image files

    Each sample consists of a single 3d MCFO image of neurons of the fruit fly.
    For each image, we provide a pixel-wise instance segmentation for all separable neurons.
    Each sample is stored as a separate zarr file (zarr is a file storage format for chunked, compressed, N-dimensional arrays based on an open-source specification.").
    The image data ("raw") and the segmentation ("gt_instances") are stored as two arrays within a single zarr file.
    The segmentation mask for each neuron is stored in a separate channel.
    The order of dimensions is CZYX.

    We recommend to work in a virtual environment, e.g., by using conda:

    conda create -y -n flylight-env -c conda-forge python=3.9
    conda activate flylight-env

    How to open zarr files

    1. Install the python zarr package:
      pip install zarr
    2. Opened a zarr file with:

      import zarr
      raw = zarr.open(
      seg = zarr.open(

      # optional:
      import numpy as np
      raw_np = np.array(raw)

    Zarr arrays are read lazily on-demand.
    Many functions that expect numpy arrays also work with zarr arrays.
    Optionally, the arrays can also explicitly be converted to numpy arrays.

    How to view zarr image files

    We recommend to use napari to view the image data.

    1. Install napari:
      pip install "napari[all]"
    2. Save the following Python script:

      import zarr, sys, napari

      raw = zarr.load(sys.argv[1], mode='r', path="volumes/raw")
      gts = zarr.load(sys.argv[1], mode='r', path="volumes/gt_instances")

      viewer = napari.Viewer(ndisplay=3)
      for idx, gt in enumerate(gts):
      viewer.add_labels(
      gt, rendering='translucent', blending='additive', name=f'gt_{idx}')
      viewer.add_image(raw[0], colormap="red", name='raw_r', blending='additive')
      viewer.add_image(raw[1], colormap="green", name='raw_g', blending='additive')
      viewer.add_image(raw[2], colormap="blue", name='raw_b', blending='additive')
      napari.run()

    3. Execute:
      python view_data.py 

    Metrics

    • S: Average of avF1 and C
    • avF1: Average F1 Score
    • C: Average ground truth coverage
    • clDice_TP: Average true positives clDice
    • FS: Number of false splits
    • FM: Number of false merges
    • tp: Relative number of true positives

    For more information on our selected metrics and formal definitions please see our paper.

    Baseline

    To showcase the FISBe dataset together with our selection of metrics, we provide evaluation results for three baseline methods, namely PatchPerPix (ppp), Flood Filling Networks (FFN) and a non-learnt application-specific color clustering from Duan et al..
    For detailed information on the methods and the quantitative results please see our paper.

    License

    The FlyLight Instance Segmentation Benchmark (FISBe) dataset is licensed under the Creative Commons Attribution 4.0 International (CC BY 4.0) license.

    Citation

    If you use FISBe in your research, please use the following BibTeX entry:

    @misc{mais2024fisbe,
     title =    {FISBe: A real-world benchmark dataset for instance
             segmentation of long-range thin filamentous structures},
     author =    {Lisa Mais and Peter Hirsch and Claire Managan and Ramya
             Kandarpa and Josef Lorenz Rumberger and Annika Reinke and Lena
             Maier-Hein and Gudrun Ihrke and Dagmar Kainmueller},
     year =     2024,
     eprint =    {2404.00130},
     archivePrefix ={arXiv},
     primaryClass = {cs.CV}
    }

    Acknowledgments

    We thank Aljoscha Nern for providing unpublished MCFO images as well as Geoffrey W. Meissner and the entire FlyLight Project Team for valuable
    discussions.
    P.H., L.M. and D.K. were supported by the HHMI Janelia Visiting Scientist Program.
    This work was co-funded by Helmholtz Imaging.

    Changelog

    There have been no changes to the dataset so far.
    All future change will be listed on the changelog page.

    Contributing

    If you would like to contribute, have encountered any issues or have any suggestions, please open an issue for the FISBe dataset in the accompanying github repository.

    All contributions are welcome!

  10. Dog Segmentation Dataset

    • kaggle.com
    zip
    Updated Mar 31, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Santhoshkumar (2023). Dog Segmentation Dataset [Dataset]. https://www.kaggle.com/datasets/santhoshkumarv/dog-segmentation-dataset
    Explore at:
    zip(5252057 bytes)Available download formats
    Dataset updated
    Mar 31, 2023
    Authors
    Santhoshkumar
    Description

    A dog segmentation dataset created manually typically involves the following steps:

    Image selection: Selecting a set of images that include dogs in various poses and backgrounds.

    Image labeling: Manually labeling the dogs in each image using a labeling tool, where each dog is segmented and assigned a unique label.

    Image annotation: Annotating the labeled images with the corresponding segmentation masks, where the dog region is assigned a value of 1 and the background region is assigned a value of 0.

    Dataset splitting: Splitting the annotated dataset into training, validation, and test sets.

    Dataset format: Saving the annotated dataset in a format suitable for use in machine learning frameworks such as TensorFlow or PyTorch.

    Dataset characteristics: The dataset may have varying image sizes and resolutions, different dog breeds, backgrounds, lighting conditions, and other variations that are typical of natural images.

    Dataset size: The size of the dataset can vary, but it should be large enough to provide a sufficient amount of training data for deep learning models.

    Dataset availability: The dataset may be made publicly available for research and educational purposes.

    Overall, a manually created dog segmentation dataset provides a high-quality training data for deep learning models and is essential for developing robust segmentation models.

  11. d

    Synthetic image data and annotation (bounding box, segmentation, keypoint,...

    • datarade.ai
    Updated Nov 28, 2021
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Mirage (2021). Synthetic image data and annotation (bounding box, segmentation, keypoint, depth, normals) [Dataset]. https://datarade.ai/data-products/synthetic-image-data-and-annotation-bounding-box-segmentati-mirage
    Explore at:
    Dataset updated
    Nov 28, 2021
    Dataset authored and provided by
    Mirage
    Area covered
    New Zealand, South Sudan, Lesotho, British Indian Ocean Territory, Cameroon, Croatia, India, Liberia, Japan, Norway
    Description

    Synthetic image data is generated on 3D game engines ready to use, fully annotated (bounding box, segmentation, keypoint, depth, normal) without any errors. Synthetic data - Solves cold start problems - Reduces development time and costs - Enables more experimentation - Covers edge cases - Removes privacy concerns - Improves existing dataset performance

  12. R

    Green Nir Normal Dataset

    • universe.roboflow.com
    zip
    Updated Feb 25, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    detection segmentation and classification (2025). Green Nir Normal Dataset [Dataset]. https://universe.roboflow.com/detection-segmentation-and-classification/green-nir-normal-dhhrv/dataset/1
    Explore at:
    zipAvailable download formats
    Dataset updated
    Feb 25, 2025
    Dataset authored and provided by
    detection segmentation and classification
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Variables measured
    Bunch AFXa Polygons
    Description

    GREEN NIR NORMAL

    ## Overview
    
    GREEN NIR NORMAL is a dataset for instance segmentation tasks - it contains Bunch AFXa annotations for 1,865 images.
    
    ## Getting Started
    
    You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
    
      ## License
    
      This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).
    
  13. Dataset split settings.

    • plos.figshare.com
    xls
    Updated Mar 21, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    You Xue; Xinya Chen; Pei Liu; Xiaoyi Lv (2024). Dataset split settings. [Dataset]. http://doi.org/10.1371/journal.pone.0299392.t001
    Explore at:
    xlsAvailable download formats
    Dataset updated
    Mar 21, 2024
    Dataset provided by
    PLOShttp://plos.org/
    Authors
    You Xue; Xinya Chen; Pei Liu; Xiaoyi Lv
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Skin cancer is one of the most common malignant tumors worldwide, and early detection is crucial for improving its cure rate. In the field of medical imaging, accurate segmentation of lesion areas within skin images is essential for precise diagnosis and effective treatment. Due to the capacity of deep learning models to conduct adaptive feature learning through end-to-end training, they have been widely applied in medical image segmentation tasks. However, challenges such as boundary ambiguity between normal skin and lesion areas, significant variations in the size and shape of lesion areas, and different types of lesions in different samples pose significant obstacles to skin lesion segmentation. Therefore, this study introduces a novel network model called HDS-Net (Hybrid Dynamic Sparse Network), aiming to address the challenges of boundary ambiguity and variations in lesion areas in skin image segmentation. Specifically, the proposed hybrid encoder can effectively extract local feature information and integrate it with global features. Additionally, a dynamic sparse attention mechanism is introduced, mitigating the impact of irrelevant redundancies on segmentation performance by precisely controlling the sparsity ratio. Experimental results on multiple public datasets demonstrate a significant improvement in Dice coefficients, reaching 0.914, 0.857, and 0.898, respectively.

  14. u

    Chest CT Segmentation Dataset

    • unidata.pro
    nii
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Unidata L.L.C-FZ, Chest CT Segmentation Dataset [Dataset]. https://unidata.pro/datasets/chest-ct/
    Explore at:
    niiAvailable download formats
    Dataset authored and provided by
    Unidata L.L.C-FZ
    Description

    Dataset with CT scans includes over 1,000 studies that highlight various pathologies such as cancer, emphysema, hydrothorax, and etc

  15. s

    Indoor Objects Segmentation Dataset

    • shaip.com
    • hmn.shaip.com
    • +3more
    json
    Updated Nov 26, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Shaip (2024). Indoor Objects Segmentation Dataset [Dataset]. https://www.shaip.com/offerings/environment-scene-segmentation-datasets/
    Explore at:
    jsonAvailable download formats
    Dataset updated
    Nov 26, 2024
    Dataset authored and provided by
    Shaip
    License

    CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
    License information was derived automatically

    Description

    The Indoor Objects Segmentation Dataset serves the advertisement, gaming, and visual entertainment sectors, offering high-resolution images ranging from 1024 × 1024 to 3024 × 4032. This dataset includes over 50 types of common indoor objects and architectural elements, such as furniture and room structures, annotated for instance, semantic, and contour segmentation.

  16. Population average atlas for BundleSeg

    • zenodo.org
    zip
    Updated Sep 4, 2025
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Francois Rheault; Francois Rheault (2025). Population average atlas for BundleSeg [Dataset]. http://doi.org/10.5281/zenodo.10103446
    Explore at:
    zipAvailable download formats
    Dataset updated
    Sep 4, 2025
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Francois Rheault; Francois Rheault
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Multi-atlas bundle segmentation

    This data is made to be used with the following script:
    https://github.com/scilus/scilpy/blob/master/scripts/scil_tractogram_segment_with_bundleseg.py

    Or the following Nextflow pipeline:
    https://github.com/scilus/rbx_flow

    Etienne St-Onge, Kurt Schilling, Francois Rheault, "BundleSeg: A versatile, reliable and reproducible approach to whitte matter bundle segmentation.", arXiv, 2308.10958 (2023)

    Rheault, François. "Analyse et reconstruction de faisceaux de la matière blanche." Computer Science (Université de Sherbrooke) (2020), https://savoirs.usherbrooke.ca/handle/11143/17255

    Usage
    Here is an example (for more details use `scil_tractogram_segment_with_bundleseg.py -h`) :

    antsRegistrationSyNQuick.sh -d 3 -f ${T1} -m mni_masked.nii.gz -t a -n 4
    scil_tractogram_segment_with_bundleseg.py ${TRACTOGRAM} config_fss_1.json atlas/*/ output0GenericAffine.mat --out_dir ${OUTPUT_DIR}/ --log_level DEBUG --minimal_vote 0.4 --processes 8 --seed 0 --inverse -f

    To facilitate interpretation, all endpoints were uniformized head/tail. To see, which side of a bundle is head or tail, you can load the atlas bundle into the software MI-Brain

    Notes on bundles
    - AC and PC were added mostly in case the atlas is used for lesion-mapping or figures. Likely, segmentation won't produce good results. This is mostly due to difficult tracking for these bundles.
    - The CC are split for each lobe. However, for technical consideration, the frontal portion was split in two to facilitate clustering and segmentation. For the same reason, the portion fanning to the pre/post central gyri were separated.
    - The streamlines present in the CC are homotopic, Recobundles will allow for variation and thus lead to 'some' heterotopy. However, it is expected that the results will be mostly homotopic.
    - CG has 3 possible endpoint locations. However, the full extent of the tail is difficult to track and is often missing.
    - FPT and POPT should terminate in the pons. However, to fully capture candidate streamlines and improve segmentation quality even streamlines reaching down the brainstem are selected.
    - PYT should reach down the brainstem. For similar reasons to the FPT/POPT, streamlines ending in the pons are selected. Otherwise, fanning is affected and bundles is too skinny.
    - OR_ML will most likely have difficulty capturing the full ML. However, this is often due to difficult tracking.
    - The cerebellum is often cut due to acquisition FOV. In such a case, all projection bundles will be more difficult to recognize and most cerebellum bundles will be missing (ICP, MCP, SCP).

    See Mosaic of bundles here.

    Acronym
    AC - Anterior commisure
    AF - Arcuate fasciculus
    CC_Fr_1 - Corpus callosum, Frontal lobe (most anterior part)
    CC_Fr_2 - Corpus callosum, Frontal lobe (most posterior part)
    CC_Oc - Corpus callosum, Occipital lobe
    CC_Pa - Corpus callosum, Parietal lobe
    CC_Pr_Po - Corpus callosum, Pre/Post central gyri
    CC_Te - Corpus callosum, Temporal lobe
    CG - Cingulum
    FAT - Frontal aslant tract
    FPT - Fronto-pontine tract
    FX - Fornix
    ICP - Inferior cerebellar peduncle
    IFOF - Inferior fronto-occipital fasciculus
    ILF - Inferior longitudinal fasciculus
    MCP - Middle cerebellar peduncle
    MdLF - Middle longitudinal fascicle
    OR_ML - Optic radiation and Meyer's loop
    PC - Posterior commisure
    POPT - parieto-occipito pontine tract
    PYT - Pyramidal tract
    SCP - Superior cerebellar peduncle
    SLF - Superior longitudinal fasciculus
    UF - Uncinate fasciculus

  17. Oily & Dry Skin Dataset (ROI)

    • kaggle.com
    zip
    Updated Feb 20, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Manith Marapperuma (2024). Oily & Dry Skin Dataset (ROI) [Dataset]. https://www.kaggle.com/datasets/manithj/oily-and-dry-skin-dataset
    Explore at:
    zip(118484915 bytes)Available download formats
    Dataset updated
    Feb 20, 2024
    Authors
    Manith Marapperuma
    License

    Apache License, v2.0https://www.apache.org/licenses/LICENSE-2.0
    License information was derived automatically

    Description

    Image Segmentation is a crucial task in computer vision that involves dividing an image into meaningful regions or segments. These segments can correspond to objects, boundaries, or other relevant parts of the image. One common approach for image segmentation is the use of Region of Interest (ROI) techniques.

    1. What Is Image Segmentation?

      • Image segmentation aims to partition an image into distinct regions based on certain criteria. These regions can be homogeneous in terms of color, texture, or other visual properties.
      • It plays a vital role in various applications, including object recognition, medical imaging, autonomous vehicles, and more.
      • Techniques for image segmentation include thresholding, edge-based methods, clustering, and deep learning-based approaches.
    2. Region of Interest (ROI) in Image Segmentation:

      • The concept of ROI refers to identifying specific areas within an image that are of particular interest or relevance.
      • In medical imaging, for instance, ROI might correspond to a tumor, blood vessel, or abnormal tissue.
      • By segmenting the ROI, we can focus our analysis on the critical regions, leading to more accurate results.
    3. Skin Classification Using Image Segmentation:

      • Skin classification involves identifying skin regions within an image.
      • In dermatology, skin lesion segmentation is essential for diagnosing conditions like melanoma or psoriasis.
      • Image segmentation helps isolate the skin area, making it easier to analyze and detect anomalies.
    4. Challenges in Skin Segmentation:

      • Skin tones can vary significantly across individuals due to factors like ethnicity, lighting conditions, and camera settings.
      • Robust skin segmentation algorithms must account for these variations.
      • Deep learning models, such as convolutional neural networks (CNNs), have shown promising results in skin segmentation tasks.
    5. Applications of Skin Segmentation:

      • Dermatology: Detecting skin diseases, assessing lesions, and monitoring treatment progress.
      • Cosmetics: Virtual makeup application, skin tone matching, and beauty filters.
      • Computer Graphics: Realistic rendering of human characters in video games and movies.
      • Biometrics: Facial recognition systems rely on accurate skin segmentation.
  18. d

    Customer Segmentation (Normalized)

    • search.dataone.org
    Updated Oct 29, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Anez, Diomar; Anez, Dimar (2025). Customer Segmentation (Normalized) [Dataset]. http://doi.org/10.7910/DVN/1RLQBY
    Explore at:
    Dataset updated
    Oct 29, 2025
    Dataset provided by
    Harvard Dataverse
    Authors
    Anez, Diomar; Anez, Dimar
    Description

    This dataset provides processed and normalized/standardized indices for the management tool 'Customer Segmentation', including the closely related concept of Market Segmentation. Derived from five distinct raw data sources, these indices are specifically designed for comparative longitudinal analysis, enabling the examination of trends and relationships across different empirical domains (web search, literature, academic publishing, and executive adoption). The data presented here represent transformed versions of the original source data, aimed at achieving metric comparability. Users requiring the unprocessed source data should consult the corresponding Customer Segmentation dataset in the Management Tool Source Data (Raw Extracts) Dataverse. Data Files and Processing Methodologies: Google Trends File (Prefix: GT_): Normalized Relative Search Interest (RSI) Input Data: Native monthly RSI values from Google Trends (Jan 2004 - Jan 2025) for the query "customer segmentation" + "market segmentation" + "customer segmentation marketing". Processing: None. Utilizes the original base-100 normalized Google Trends index. Output Metric: Monthly Normalized RSI (Base 100). Frequency: Monthly. Google Books Ngram Viewer File (Prefix: GB_): Normalized Relative Frequency Input Data: Annual relative frequency values from Google Books Ngram Viewer (1950-2022, English corpus, no smoothing) for the query Customer Segmentation + Market Segmentation. Processing: Annual relative frequency series normalized (peak year = 100). Output Metric: Annual Normalized Relative Frequency Index (Base 100). Frequency: Annual. Crossref.org File (Prefix: CR_): Normalized Relative Publication Share Index Input Data: Absolute monthly publication counts matching Customer Segmentation-related keywords [("customer segmentation" OR ...) AND (...) - see raw data for full query] in titles/abstracts (1950-2025), alongside total monthly Crossref publications. Deduplicated via DOIs. Processing: Monthly relative share calculated (Segmentation Count / Total Count). Monthly relative share series normalized (peak month's share = 100). Output Metric: Monthly Normalized Relative Publication Share Index (Base 100). Frequency: Monthly. Bain & Co. Survey - Usability File (Prefix: BU_): Normalized Usability Index Input Data: Original usability percentages (%) from Bain surveys for specific years: Customer Segmentation (1999, 2000, 2002, 2004, 2006, 2008, 2010, 2012, 2014, 2017). Note: Not reported in 2022 survey data. Processing: Normalization: Original usability percentages normalized relative to its historical peak (Max % = 100). Output Metric: Biennial Estimated Normalized Usability Index (Base 100 relative to historical peak). Frequency: Biennial (Approx.). Bain & Co. Survey - Satisfaction File (Prefix: BS_): Standardized Satisfaction Index Input Data: Original average satisfaction scores (1-5 scale) from Bain surveys for specific years: Customer Segmentation (1999-2017). Note: Not reported in 2022 survey data. Processing: Standardization (Z-scores): Using Z = (X - 3.0) / 0.891609. Index Scale Transformation: Index = 50 + (Z * 22). Output Metric: Biennial Standardized Satisfaction Index (Center=50, Range?[1,100]). Frequency: Biennial (Approx.). File Naming Convention: Files generally follow the pattern: PREFIX_Tool_Processed.csv or similar, where the PREFIX indicates the data source (GT_, GB_, CR_, BU_, BS_). Consult the parent Dataverse description (Management Tool Comparative Indices) for general context and the methodological disclaimer. For original extraction details (specific keywords, URLs, etc.), refer to the corresponding Customer Segmentation dataset in the Raw Extracts Dataverse. Comprehensive project documentation provides full details on all processing steps.

  19. HOWS-CL-25: Household Objects Within Simulation Dataset for Continual...

    • zenodo.org
    • opendatalab.com
    • +1more
    bin, zip
    Updated Oct 21, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Markus Knauer; Markus Knauer; Maximilian Denninger; Maximilian Denninger; Rudolph Triebel; Rudolph Triebel (2022). HOWS-CL-25: Household Objects Within Simulation Dataset for Continual Learning [Dataset]. http://doi.org/10.5281/zenodo.7189434
    Explore at:
    bin, zipAvailable download formats
    Dataset updated
    Oct 21, 2022
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Markus Knauer; Markus Knauer; Maximilian Denninger; Maximilian Denninger; Rudolph Triebel; Rudolph Triebel
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    HOWS-CL-25 (Household Objects Within Simulation dataset for Continual Learning) is a synthetic dataset especially designed for object classification on mobile robots operating in a changing environment (like a household), where it is important to learn new, never seen objects on the fly.
    This dataset can also be used for other learning use-cases, like instance segmentation or depth estimation.
    Or where household objects or continual learning are of interest.

    Our dataset contains 150,795 unique synthetic images using 25 different household categories with 925 3D models in total. For each of those categories, we generated about 6000 RGB images. In addition, we also provide a corresponding depth, segmentation, and normal image.

    The dataset was created with BlenderProc [Denninger et al. (2019)], a procedural pipeline to generate images for deep learning.
    This tool created a virtual room with randomly textured floors, walls, and a light source with randomly chosen light intensity and color. After that, a 3D model is placed in the resulting room. This object gets customized by randomly assigning materials, including different textures, to achieve a diverse dataset. Moreover, each object might be deformed with a random
    displacement texture.
    We use 774 3D models from the ShapeNet dataset [A. X. Chang et al. (2015)] and the other models from various internet sites. Please note that we had to manually fix and filter most of the models with Blender before using them in the pipeline!

    For continual learning (CL), we provide two different loading schemes:
    - Five sequences with five categories each
    - Twelve sequences with three categories in the first and two in the other sequences.

    In addition to the RGB, depth, segmentation, and normal images, we also provide the calculated features of the RGB images (by ResNet50) as used in our RECALL paper.
    In those two loading schemes, ten percent of the images are used for validation, where we ensure that an object instance is either in the training or the validation set, not in both. This avoids learning to recognize certain instances by heart.

    We recommend using those loading schemes to compare your approach with others.

    Here we provide three files for download:
    - HOWS_CL_25.zip [124GB]: This is the original dataset with the RGB, depth, segmentation, and normal images, as well as the loading schemes. It is divided into three archive parts. To open the dataset, please ensure to download all three parts.
    - HOWS_CL_25_hdf5_features.zip [2.5GB]: This only contains the calculated features from the RGB input by a ResNet50 in a .hdf5 file. Download this if you want to use the dataset for learning and/or want to compare your approach to our RECALL approach (where we used the same features).
    - README.md: Some additional explanation.

    For further information and code examples, please have a look at our website: https://github.com/DLR-RM/RECALL.

  20. Experimental environment information.

    • plos.figshare.com
    xls
    Updated Feb 22, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Huazhe Wang; Li Ma (2024). Experimental environment information. [Dataset]. http://doi.org/10.1371/journal.pone.0296031.t001
    Explore at:
    xlsAvailable download formats
    Dataset updated
    Feb 22, 2024
    Dataset provided by
    PLOShttp://plos.org/
    Authors
    Huazhe Wang; Li Ma
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    In the realm of digital image applications, image processing technology occupies a pivotal position, with image segmentation serving as a foundational component. As the digital image application domain expands across industries, the conventional segmentation techniques increasingly challenge to cater to modern demands. To address this gap, this paper introduces an MCMC-based image segmentation algorithm based on the Markov Random Field (MRF) model, marking a significant stride in the field. The novelty of this research lies in its method that capitalizes on domain information in pixel space, amplifying the local segmentation precision of image segmentation algorithms. Further innovation is manifested in the development of an adaptive segmentation image denoising algorithm based on MCMC sampling. This algorithm not only elevates image segmentation outcomes, but also proficiently denoises the image. In the experimental results, MRF-MCMC achieves better segmentation performance, with an average segmentation accuracy of 94.26% in Lena images, significantly superior to other common image segmentation algorithms. In addition, the study proposes that the denoising model outperforms other algorithms in peak signal-to-noise ratio and structural similarity in environments with noise standard deviations of 15, 25, and 50. In essence, these experimental findings affirm the efficacy of this study, opening avenues for refining digital image segmentation methodologies.

Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
Arvin Arian (2023). COVID-19 & Normal CT Segmentation Dataset [Dataset]. http://doi.org/10.17632/pfmgfpwnmm.2

COVID-19 & Normal CT Segmentation Dataset

Explore at:
Dataset updated
Nov 27, 2023
Authors
Arvin Arian
License

Attribution-NonCommercial 3.0 (CC BY-NC 3.0)https://creativecommons.org/licenses/by-nc/3.0/
License information was derived automatically

Description

This dataset includes CT data and segmentation masks from patients diagnosed with COVID-19, as well as data from subjects without the infection.

This study is approved under the ethical approval codes of IR.TUMS.IKHC.REC.1399.255 and IR.TUMS.VCR.REC.1399.488 at Tehran University of Medical Sciences.

The code for loading the dataset and running an AI model is available on: https://github.com/SamanSotoudeh/COVID19-segmentation

Please use the following citations:

1- Arian, Arvin; Mehrabinejad, Mohammad-Mehdi; Zoorpaikar, Mostafa; Hasanzadeh, Navid; Sotoudeh-Paima, Saman; Kolahi, Shahriar; Gity, Masoumeh; Soltanian-Zadeh, "Accuracy of Artificial Intelligence CT Quantification in Predicting COVID-19 Subjects’ Prognosis" PLoS ONE (2023).

2- Sotoudeh-Paima, Saman, et al. "A Multi-centric Evaluation of Deep Learning Models for Segmentation of COVID-19 Lung Lesions on Chest CT Scans." Iranian Journal of Radiology 19.4 (2022).

3- Hasanzadeh, Navid, et al. "Segmentation of COVID-19 Infections on CT: Comparison of four UNet-based networks." 2020 27th National and 5th International Iranian Conference on Biomedical Engineering (ICBME). IEEE, 2020.

Search
Clear search
Close search
Google apps
Main menu