100+ datasets found
  1. i

    Medical Imaging Datasets for Multimodal Disease Detection and Diagnosis...

    • ieee-dataport.org
    Updated Aug 10, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Nur Rusyidah Azri (2024). Medical Imaging Datasets for Multimodal Disease Detection and Diagnosis Research [Dataset]. https://ieee-dataport.org/documents/medical-imaging-datasets-multimodal-disease-detection-and-diagnosis-research
    Explore at:
    Dataset updated
    Aug 10, 2024
    Authors
    Nur Rusyidah Azri
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    brain MRI

  2. The MultiCaRe Dataset: A Multimodal Case Report Dataset with Clinical Cases,...

    • zenodo.org
    bin, csv, zip
    Updated Jan 5, 2024
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Mauro Nievas Offidani; Mauro Nievas Offidani; Claudio Delrieux; Claudio Delrieux (2024). The MultiCaRe Dataset: A Multimodal Case Report Dataset with Clinical Cases, Labeled Images and Captions from Open Access PMC Articles [Dataset]. http://doi.org/10.5281/zenodo.10079370
    Explore at:
    zip, bin, csvAvailable download formats
    Dataset updated
    Jan 5, 2024
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Mauro Nievas Offidani; Mauro Nievas Offidani; Claudio Delrieux; Claudio Delrieux
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    The dataset contains multi-modal data from over 75,000 open access and de-identified case reports, including metadata, clinical cases, image captions and more than 130,000 images. Images and clinical cases belong to different medical specialties, such as oncology, cardiology, surgery and pathology. The structure of the dataset allows to easily map images with their corresponding article metadata, clinical case, captions and image labels. Details of the data structure can be found in the file data_dictionary.csv.

    Almost 100,000 patients and almost 400,000 medical doctors and researchers were involved in the creation of the articles included in this dataset. The citation data of each article can be found in the metadata.parquet file.

    Refer to the examples showcased in this GitHub repository to understand how to optimize the use of this dataset.

    For a detailed insight about the contents of this dataset, please refer to this data article published in Data In Brief.

  3. h

    Multimodal ground truth datasets for abdominal medical image registration...

    • heidata.uni-heidelberg.de
    zip
    Updated Feb 23, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Frank Zöllner; Frank Zöllner (2023). Multimodal ground truth datasets for abdominal medical image registration [data] [Dataset]. http://doi.org/10.11588/DATA/ICSFUS
    Explore at:
    zip(3796777237), zip(27228993659), zip(2968034134)Available download formats
    Dataset updated
    Feb 23, 2023
    Dataset provided by
    heiDATA
    Authors
    Frank Zöllner; Frank Zöllner
    License

    Attribution-NonCommercial 4.0 (CC BY-NC 4.0)https://creativecommons.org/licenses/by-nc/4.0/
    License information was derived automatically

    Dataset funded by
    BMBF
    Description

    Sparsity of annotated data is a major limitation in medical image processing tasks such as registration. Registered multimodal image data are essential for the diagnosis of medical conditions and the success of interventional medical procedures. To overcome the shortage of data, we present a method that allows the generation of annotated multimodal 4D datasets. We use a CycleGAN network architecture to generate multimodal synthetic data from the 4D extended cardiac–torso (XCAT) phantom and real patient data. Organ masks are provided by the XCAT phantom; therefore, the generated dataset can serve as ground truth for image segmentation and registration. Compared to real patient data, the synthetic data showed good agreement regarding the image voxel intensity distribution and the noise characteristics. The generated T1-weighted magnetic resonance imaging, computed tomography (CT), and cone beam CT images are inherently co-registered.

  4. d

    Pixta AI | Imagery Data | Global | High volume | Annotation and Labelling...

    • datarade.ai
    .json, .xml, .csv
    Updated Jul 19, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Pixta AI (2023). Pixta AI | Imagery Data | Global | High volume | Annotation and Labelling Services Provided | Multimodal Medical Images OTS Datasets for AI and ML [Dataset]. https://datarade.ai/data-products/multimodal-medical-image-ots-datasets-pixta-ai
    Explore at:
    .json, .xml, .csvAvailable download formats
    Dataset updated
    Jul 19, 2023
    Dataset authored and provided by
    Pixta AI
    Area covered
    Guernsey, Serbia, Haiti, Malaysia, Montenegro, Pitcairn, Uruguay, French Polynesia, Lebanon, Maldives
    Description
    1. Overview This dataset is a collection of multimodal high quality image sets of medical data that are ready to use for optimizing the accuracy of computer vision models. All of the contents are sourced from Pixta AI's partner network with high quality & full data compliance.

    2. Data subject The datasets consist of various models

    3. X-ray datasets

    4. CT datasets

    5. MRI datasets

    6. Mammography datasets

    7. Segmentation datasets

    8. Classification datasets

    9. Regression datasets

    10. Use case The dataset could be used for various Healthcare & Medical models:

    11. Medical Image Analysis

    12. Remote Diagnosis

    13. Medical Record Keeping ... Each data set is supported by both AI and expert doctors review process to ensure labelling consistency and accuracy. Contact us for more custom datasets.

    14. About PIXTA PIXTASTOCK is the largest Asian-featured stock platform providing data, contents, tools and services since 2005. PIXTA experiences 15 years of integrating advanced AI technology in managing, curating, processing over 100M visual materials and serving global leading brands for their creative and data demands. Visit us at https://www.pixta.ai/ or contact via our email admin.bi@pixta.co.jp.

  5. i

    Multimodal Text Medical Image Dataset

    • ieee-dataport.org
    Updated May 17, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Xiang zhe (2025). Multimodal Text Medical Image Dataset [Dataset]. https://ieee-dataport.org/documents/multimodal-text-medical-image-dataset
    Explore at:
    Dataset updated
    May 17, 2025
    Authors
    Xiang zhe
    Description

    The Dataset is a meticulously curated high-quality dataset specifically designed for semantic-guided image fusion in the medical domain. This dataset aims to facilitate advanced research and development in multimodal medical image analysis by providing a comprehensive collection of images from various imaging modalities.

  6. Z

    MultiCaRe: An open-source clinical case dataset for medical image...

    • data.niaid.nih.gov
    • explore.openaire.eu
    Updated Mar 9, 2025
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Nievas Offidani, Mauro (2025). MultiCaRe: An open-source clinical case dataset for medical image classification and multimodal AI applications [Dataset]. https://data.niaid.nih.gov/resources?id=zenodo_10079369
    Explore at:
    Dataset updated
    Mar 9, 2025
    Dataset authored and provided by
    Nievas Offidani, Mauro
    License

    Attribution-NonCommercial-ShareAlike 4.0 (CC BY-NC-SA 4.0)https://creativecommons.org/licenses/by-nc-sa/4.0/
    License information was derived automatically

    Description

    The dataset contains multi-modal data from over 70,000 open access and de-identified case reports, including metadata, clinical cases, image captions and more than 130,000 images. Images and clinical cases belong to different medical specialties, such as oncology, cardiology, surgery and pathology. The structure of the dataset allows to easily map images with their corresponding article metadata, clinical case, captions and image labels. Details of the data structure can be found in the file data_dictionary.csv.

    More than 90,000 patients and 280,000 medical doctors and researchers were involved in the creation of the articles included in this dataset. The citation data of each article can be found in the metadata.parquet file.

    Refer to the examples showcased in this GitHub repository to understand how to optimize the use of this dataset.The license of the dataset as a whole is CC BY-NC-SA. However, its individual contents may have less restrictive license types (CC BY, CC BY-NC, CC0). For instance, regarding image filess, 66K of them are CC BY, 32K are CC BY-NC-SA, 32K are CC BY-NC, and 20 of them are CC0.

  7. M

    OLD-INSPECT: A Multimodal Dataset for Pulmonary Embolism Diagnosis and...

    • stanfordaimi.azurewebsites.net
    Updated May 30, 2024
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Microsoft Research (2024). OLD-INSPECT: A Multimodal Dataset for Pulmonary Embolism Diagnosis and Prognosis [Dataset]. https://stanfordaimi.azurewebsites.net/datasets/318f3464-c4b6-4006-9856-6f48ba40ad67
    Explore at:
    Dataset updated
    May 30, 2024
    Dataset authored and provided by
    Microsoft Research
    License

    https://aimistanford-web-api.azurewebsites.net/licenses/f1f352a6-243f-4905-8e00-389edbca9e83/viewhttps://aimistanford-web-api.azurewebsites.net/licenses/f1f352a6-243f-4905-8e00-389edbca9e83/view

    Description

    Synthesizing information from various data sources plays a crucial role in the practice of modern medicine. Current applications of artificial intelligence in medicine often focus on single-modality data due to a lack of publicly available, multimodal medical datasets. To address this limitation, we introduce INSPECT, which contains de-identified longitudinal records from a large cohort of pulmonary embolism (PE) patients, along with ground truth labels for multiple outcomes. INSPECT contains data from 19,438 patients, including CT images, sections of radiology reports, and structured electronic health record (EHR) data (including demographics, diagnoses, procedures, and vitals). Using our provided dataset, we develop and release a benchmark for evaluating several baseline modeling approaches on a variety of important PE related tasks. We evaluate image-only, EHR-only, and fused models. Trained models and the de-identified dataset are made available for non-commercial use under a data use agreement. To the best our knowledge, INSPECT is the largest multimodal dataset for enabling reproducible research on strategies for integrating 3D medical imaging and EHR data. NOTE: this is the first part of release due to PHI review. This release has 20078 CT scans, 21,266 impression sections and the EHR modality data will be uploaded to Stanford Redivis website (https://redivis.com/Stanford)

  8. h

    Medical-Multimodal-EN-TH

    • huggingface.co
    Updated Jul 3, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    ZomBitX64 (2025). Medical-Multimodal-EN-TH [Dataset]. https://huggingface.co/datasets/ZombitX64/Medical-Multimodal-EN-TH
    Explore at:
    Dataset updated
    Jul 3, 2025
    Dataset authored and provided by
    ZomBitX64
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    HealthGPTVL-Translation Medical-Multimodal-EN-TH

    This dataset is a bilingual (English-Thai) medical multimodal evaluation dataset containing medical images with corresponding question-answer pairs for visual question answering and translation tasks.

      Dataset Details
    
    
    
    
    
      Dataset Description
    

    This dataset contains 17,047 medical image-text pairs designed for multimodal medical AI evaluation. It includes medical images from various imaging modalities (MRI, CT, X-Ray… See the full description on the dataset page: https://huggingface.co/datasets/ZombitX64/Medical-Multimodal-EN-TH.

  9. f

    VGG16 Metrics, Average AUC = 0.9987707721217087.

    • plos.figshare.com
    xls
    Updated Dec 13, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Craig Macfadyen; Ajay Duraiswamy; David Harris-Birtill (2023). VGG16 Metrics, Average AUC = 0.9987707721217087. [Dataset]. http://doi.org/10.1371/journal.pdig.0000191.t005
    Explore at:
    xlsAvailable download formats
    Dataset updated
    Dec 13, 2023
    Dataset provided by
    PLOS Digital Health
    Authors
    Craig Macfadyen; Ajay Duraiswamy; David Harris-Birtill
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Algorithms that classify hyper-scale multi-modal datasets, comprising of millions of images, into constituent modality types can help researchers quickly retrieve and classify diagnostic imaging data, accelerating clinical outcomes. This research aims to demonstrate that a deep neural network that is trained on a hyper-scale dataset (4.5 million images) composed of heterogeneous multi-modal data can be used to obtain significant modality classification accuracy (96%). By combining 102 medical imaging datasets, a dataset of 4.5 million images was created. A ResNet-50, ResNet-18, and VGG16 were trained to classify these images by the imaging modality used to capture them (Computed Tomography (CT), Magnetic Resonance Imaging (MRI), Positron Emission Tomography (PET), and X-ray) across many body locations. The classification accuracy of the models was then tested on unseen data. The best performing model achieved classification accuracy of 96% on unseen data, which is on-par, or exceeds the accuracy of more complex implementations using EfficientNets or Vision Transformers (ViTs). The model achieved a balanced accuracy of 86%. This research shows it is possible to train Deep Learning (DL) Convolutional Neural Networks (CNNs) with hyper-scale multimodal datasets, composed of millions of images. Such models can find use in real-world applications with volumes of image data in the hyper-scale range, such as medical imaging repositories, or national healthcare institutions. Further research can expand this classification capability to include 3D-scans.

  10. h

    Medical_Multimodal_Evaluation_Data

    • huggingface.co
    Updated Jun 1, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    FreedomAI (2025). Medical_Multimodal_Evaluation_Data [Dataset]. https://huggingface.co/datasets/FreedomIntelligence/Medical_Multimodal_Evaluation_Data
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Jun 1, 2025
    Dataset authored and provided by
    FreedomAI
    License

    Apache License, v2.0https://www.apache.org/licenses/LICENSE-2.0
    License information was derived automatically

    Description

    Evaluation Guide

    This dataset is used to evaluate medical multimodal LLMs, as used in HuatuoGPT-Vision. It includes benchmarks such as VQA-RAD, SLAKE, PathVQA, PMC-VQA, OmniMedVQA, and MMMU-Medical-Tracks.
    To get started:

    Download the dataset and extract the images.zip file.
    Find evaluation code on our GitHub: HuatuoGPT-Vision.

    This open-source release aims to simplify the evaluation of medical multimodal capabilities in large models. Please cite the relevant benchmark… See the full description on the dataset page: https://huggingface.co/datasets/FreedomIntelligence/Medical_Multimodal_Evaluation_Data.

  11. p

    A multimodal dental dataset facilitating machine learning research and...

    • physionet.org
    Updated Oct 11, 2024
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Wenjing Liu; Yunyou Huang; Suqin Tang (2024). A multimodal dental dataset facilitating machine learning research and clinic services [Dataset]. http://doi.org/10.13026/h1tt-fc69
    Explore at:
    Dataset updated
    Oct 11, 2024
    Authors
    Wenjing Liu; Yunyou Huang; Suqin Tang
    License

    https://github.com/MIT-LCP/license-and-dua/tree/master/draftshttps://github.com/MIT-LCP/license-and-dua/tree/master/drafts

    Description

    Oral diseases affect nearly 3.5 billion people, with the majority residing in low- and middle-income countries. Due to limited healthcare resources, many individuals are unable to access proper oral healthcare services. Image-based machine learning technology is one of the most promising approaches to improving oral healthcare services and reducing patient costs. Openly accessible datasets play a crucial role in facilitating the development of machine learning techniques. However, existing dental datasets have limitations such as a scarcity of Cone Beam Computed Tomography (CBCT) data, lack of matched multi-modal data, and insufficient complexity and diversity of the data. This project addresses these challenges by providing a dataset that includes 329 CBCT images from 169 patients, multi-modal data with matching modalities, and images representing various oral health conditions.

  12. Z

    Data from: MedMNIST Classification Decathlon: A Lightweight AutoML Benchmark...

    • data.niaid.nih.gov
    • explore.openaire.eu
    Updated Apr 19, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Jiancheng Yang (2023). MedMNIST Classification Decathlon: A Lightweight AutoML Benchmark for Medical Image Analysis [Dataset]. https://data.niaid.nih.gov/resources?id=zenodo_4269851
    Explore at:
    Dataset updated
    Apr 19, 2023
    Dataset provided by
    Rui Shi
    Bingbing Ni
    Jiancheng Yang
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    This data repository for MedMNIST v1 is out of date! Please check the latest version of MedMNIST v2.

    Abstract

    We present MedMNIST, a collection of 10 pre-processed medical open datasets. MedMNIST is standardized to perform classification tasks on lightweight 28x28 images, which requires no background knowledge. Covering the primary data modalities in medical image analysis, it is diverse on data scale (from 100 to 100,000) and tasks (binary/multi-class, ordinal regression and multi-label). MedMNIST could be used for educational purpose, rapid prototyping, multi-modal machine learning or AutoML in medical image analysis. Moreover, MedMNIST Classification Decathlon is designed to benchmark AutoML algorithms on all 10 datasets; We have compared several baseline methods, including open-source or commercial AutoML tools. The datasets, evaluation code and baseline methods for MedMNIST are publicly available at https://medmnist.github.io/.

    Please note that this dataset is NOT intended for clinical use.

    We recommend our official code to download, parse and use the MedMNIST dataset:

    pip install medmnist

    Citation and Licenses

    If you find this project useful, please cite our ISBI'21 paper as: Jiancheng Yang, Rui Shi, Bingbing Ni. "MedMNIST Classification Decathlon: A Lightweight AutoML Benchmark for Medical Image Analysis," arXiv preprint arXiv:2010.14925, 2020.

    or using bibtex: @article{medmnist, title={MedMNIST Classification Decathlon: A Lightweight AutoML Benchmark for Medical Image Analysis}, author={Yang, Jiancheng and Shi, Rui and Ni, Bingbing}, journal={arXiv preprint arXiv:2010.14925}, year={2020} }

    Besides, please cite the corresponding paper if you use any subset of MedMNIST. Each subset uses the same license as that of the source dataset.

    PathMNIST

    Jakob Nikolas Kather, Johannes Krisam, et al., "Predicting survival from colorectal cancer histology slides using deep learning: A retrospective multicenter study," PLOS Medicine, vol. 16, no. 1, pp. 1–22, 01 2019.

    License: CC BY 4.0

    ChestMNIST

    Xiaosong Wang, Yifan Peng, et al., "Chestx-ray8: Hospital-scale chest x-ray database and benchmarks on weakly-supervised classification and localization of common thorax diseases," in CVPR, 2017, pp. 3462–3471.

    License: CC0 1.0

    DermaMNIST

    Philipp Tschandl, Cliff Rosendahl, and Harald Kittler, "The ham10000 dataset, a large collection of multisource dermatoscopic images of common pigmented skin lesions," Scientific data, vol. 5, pp. 180161, 2018.

    Noel Codella, Veronica Rotemberg, Philipp Tschandl, M. Emre Celebi, Stephen Dusza, David Gutman, Brian Helba, Aadi Kalloo, Konstantinos Liopyris, Michael Marchetti, Harald Kittler, and Allan Halpern: “Skin Lesion Analysis Toward Melanoma Detection 2018: A Challenge Hosted by the International Skin Imaging Collaboration (ISIC)”, 2018; arXiv:1902.03368.

    License: CC BY-NC 4.0

    OCTMNIST/PneumoniaMNIST

    Daniel S. Kermany, Michael Goldbaum, et al., "Identifying medical diagnoses and treatable diseases by image-based deep learning," Cell, vol. 172, no. 5, pp. 1122 – 1131.e9, 2018.

    License: CC BY 4.0

    RetinaMNIST

    DeepDR Diabetic Retinopathy Image Dataset (DeepDRiD), "The 2nd diabetic retinopathy – grading and image quality estimation challenge," https://isbi.deepdr.org/data.html, 2020.

    License: CC BY 4.0

    BreastMNIST

    Walid Al-Dhabyani, Mohammed Gomaa, Hussien Khaled, and Aly Fahmy, "Dataset of breast ultrasound images," Data in Brief, vol. 28, pp. 104863, 2020.

    License: CC BY 4.0

    OrganMNIST_{Axial,Coronal,Sagittal}

    Patrick Bilic, Patrick Ferdinand Christ, et al., "The liver tumor segmentation benchmark (lits)," arXiv preprint arXiv:1901.04056, 2019.

    Xuanang Xu, Fugen Zhou, et al., "Efficient multiple organ localization in ct image using 3d region proposal network," IEEE Transactions on Medical Imaging, vol. 38, no. 8, pp. 1885–1898, 2019.

    License: CC BY 4.0

  13. f

    Data from: MMDental - A multimodal dataset of tooth CBCT images with expert...

    • springernature.figshare.com
    bin
    Updated Jul 10, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Chengkai Wang; Yifang Zhang; Chengyu Wu; Xingliang Huang; Liuxi Wu; Yitong Wang; Yiting Lu; Xiang Feng; Yaqi Wang (2025). MMDental - A multimodal dataset of tooth CBCT images with expert medical records [Dataset]. http://doi.org/10.6084/m9.figshare.28505276.v1
    Explore at:
    binAvailable download formats
    Dataset updated
    Jul 10, 2025
    Dataset provided by
    figshare
    Authors
    Chengkai Wang; Yifang Zhang; Chengyu Wu; Xingliang Huang; Liuxi Wu; Yitong Wang; Yiting Lu; Xiang Feng; Yaqi Wang
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    In the rapidly evolving field of dental intelligent healthcare, where Artificial Intelligence (AI) plays a pivotal role, the demand for multimodal datasets is critical. Existing public datasets are primarily composed of single-modal data, predominantly dental radiographs or scans, which limits the development of AI-driven applications for intelligent dental treatment. In this paper, we collect a MultiModal Dental (MMDental) dataset to address this gap. MMDental comprises data from 660 patients, including 3D Cone-beam Computed Tomography (CBCT) images and corresponding detailed expert medical records with initial diagnoses and follow-up documentation. All CBCT scans are conducted under the guidance of professional physicians, and all patient records are reviewed by senior doctors. To the best of our knowledge, this is the first and largest dataset containing 3D CBCT images of teeth with corresponding medical records. Furthermore, we provide a comprehensive analysis of the dataset by exploring patient demographics, prevalence of various dental conditions, and the disease distribution across age groups. We believe this work will be beneficial for further advancements in dental intelligent treatment.

  14. h

    MedTrinity-25M

    • huggingface.co
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    UCSC-VLAA, MedTrinity-25M [Dataset]. https://huggingface.co/datasets/UCSC-VLAA/MedTrinity-25M
    Explore at:
    Dataset authored and provided by
    UCSC-VLAA
    Description

    Tutorial of using Medtrinity-25M

    MedTrinity-25M, a comprehensive, large-scale multimodal dataset for medicine, covering over 25 million images across 10 modalities, with multigranular annotations for more than 65 diseases. These enriched annotations encompass both global textual information, such as disease/lesion type, modality, region-specific descriptions, and inter-regional relationships, as well as detailed local annotations for regions of interest (ROIs), including bounding… See the full description on the dataset page: https://huggingface.co/datasets/UCSC-VLAA/MedTrinity-25M.

  15. p

    Data from: Multimodal Clinical Monitoring in the Emergency Department...

    • physionet.org
    Updated Mar 3, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Aman Kansal; Emma Chen; Tom Jin; Pranav Rajpurkar; David Kim (2025). Multimodal Clinical Monitoring in the Emergency Department (MC-MED) [Dataset]. http://doi.org/10.13026/jz99-4j81
    Explore at:
    Dataset updated
    Mar 3, 2025
    Authors
    Aman Kansal; Emma Chen; Tom Jin; Pranav Rajpurkar; David Kim
    License

    https://github.com/MIT-LCP/license-and-dua/tree/master/draftshttps://github.com/MIT-LCP/license-and-dua/tree/master/drafts

    Description

    Emergency department (ED) patients often present with undiagnosed complaints, and can exhibit rapidly evolving physiology. Therefore, data from continuous physiologic monitoring, in addition to the electronic health record, is essential to understand the acute course of illness and responses to interventions. The complexity of ED care and the large amount of unstructured multimodal data it produces has limited the accessibility of detailed ED data for research. We release Multimodal Clinical Monitoring in the Emergency Department (MC-MED), a comprehensive, multimodal, and de-identified clinical and physiological dataset. MC-MED includes 118,385 adult ED visits to an academic medical center from 2020 to 2022. Data include continuously monitored vital signs, physiologic waveforms (electrocardiogram, photoplethysmogram, respiration), patient demographics, medical histories, orders, medication administrations, laboratory and imaging results, and visit outcomes. MC-MED is the first dataset to combine detailed physiologic monitoring with clinical events and outcomes for a large, diverse ED population.

  16. MedMultiPoints

    • huggingface.co
    Updated Jan 3, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Simula Metropolitan Center for Digital Engineering (SimulaMet) (2025). MedMultiPoints [Dataset]. https://huggingface.co/datasets/SimulaMet/MedMultiPoints
    Explore at:
    Dataset updated
    Jan 3, 2025
    Dataset provided by
    Simula Metropolitan Center for Digital Engineering
    Authors
    Simula Metropolitan Center for Digital Engineering (SimulaMet)
    Description

    🩺 MedMultiPoints: A Multimodal Dataset for Object Detection, Localization, and Counting in Medical Imaging

    📫 For queries, contact: sushant@simula.no

      Dataset Summary
    

    MedMultiPoints is a curated, multimodal medical imaging dataset designed for multi-task learning in the medical domain—spanning object detection, localization, and counting tasks. It integrates data from endoscopic and microscopic modalities, reflecting real-world clinical diversity. The dataset is… See the full description on the dataset page: https://huggingface.co/datasets/SimulaMet/MedMultiPoints.

  17. ROCOv2: Radiology Objects in COntext Version 2, An Updated Multimodal Image...

    • zenodo.org
    csv, zip
    Updated Oct 9, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Johannes Rückert; Johannes Rückert; Louise Bloch; Louise Bloch; Raphael Brüngel; Raphael Brüngel; Ahmad Idrissi-Yaghir; Ahmad Idrissi-Yaghir; Henning Schäfer; Henning Schäfer; Cynthia S. Schmidt; Cynthia S. Schmidt; Sven Koitka; Sven Koitka; Obioma Pelka; Asma Ben Abacha; Asma Ben Abacha; Alba Garcia Seco de Herrera; Alba Garcia Seco de Herrera; Henning Müller; Henning Müller; Peter A. Horn; Felix Nensa; Felix Nensa; Christoph M. Friedrich; Christoph M. Friedrich; Obioma Pelka; Peter A. Horn (2024). ROCOv2: Radiology Objects in COntext Version 2, An Updated Multimodal Image Dataset [Dataset]. http://doi.org/10.5281/zenodo.10821435
    Explore at:
    csv, zipAvailable download formats
    Dataset updated
    Oct 9, 2024
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Johannes Rückert; Johannes Rückert; Louise Bloch; Louise Bloch; Raphael Brüngel; Raphael Brüngel; Ahmad Idrissi-Yaghir; Ahmad Idrissi-Yaghir; Henning Schäfer; Henning Schäfer; Cynthia S. Schmidt; Cynthia S. Schmidt; Sven Koitka; Sven Koitka; Obioma Pelka; Asma Ben Abacha; Asma Ben Abacha; Alba Garcia Seco de Herrera; Alba Garcia Seco de Herrera; Henning Müller; Henning Müller; Peter A. Horn; Felix Nensa; Felix Nensa; Christoph M. Friedrich; Christoph M. Friedrich; Obioma Pelka; Peter A. Horn
    License

    Attribution-NonCommercial 4.0 (CC BY-NC 4.0)https://creativecommons.org/licenses/by-nc/4.0/
    License information was derived automatically

    Time period covered
    Nov 10, 2023
    Description

    Recent advances in deep learning techniques have enabled the development of systems for automatic analysis of medical images. These systems often require large amounts of training data with high quality labels, which is difficult and time consuming to generate.

    Here, we introduce Radiology Object in COntext Version 2 (ROCOv2), a multimodal dataset consisting of radiological images and associated medical concepts and captions extracted from the PubMed Open Access subset. Concepts for clinical modality, anatomy (X-ray), and directionality (X-ray) were manually curated and additionally evaluated by a radiologist. Unlike MIMIC-CXR, ROCOv2 includes seven different clinical modalities.

    It is an updated version of the ROCO dataset published in 2018, and includes 35,705 new images added to PubMed since 2018, as well as manually curated medical concepts for modality, body region (X-ray) and directionality (X-ray). The dataset consists of 79,789 images and has been used, with minor modifications, in the concept detection and caption prediction tasks of ImageCLEFmedical 2023. The participants had access to the training and validation sets after signing a user agreement.

    The dataset is suitable for training image annotation models based on image-caption pairs, or for multi-label image classification using the UMLS concepts provided with each image, e.g., to build systems to support structured medical reporting.

    Additional possible use cases for the ROCOv2 dataset include the pre-training of models for the medical domain, and the evaluation evaluation of deep learning models for multi-task learning.

  18. h

    ROCO-radiology

    • huggingface.co
    Updated Nov 14, 2024
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Ronan L.M. (2024). ROCO-radiology [Dataset]. https://huggingface.co/datasets/eltorio/ROCO-radiology
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Nov 14, 2024
    Authors
    Ronan L.M.
    License

    MIT Licensehttps://opensource.org/licenses/MIT
    License information was derived automatically

    Description

    The "ROCO-radiology" dataset is derived from the Radiology Objects in COntext (ROCO) dataset, a large-scale medical and multimodal imaging collection. The language used is primarily English, and it covers the domain of medical imaging, specifically radiology. We only modified the dataset by choosing only for radiology dataset and convert the image into PIL Object. For further details and citation, pleaser refer to original author.… See the full description on the dataset page: https://huggingface.co/datasets/eltorio/ROCO-radiology.

  19. INSPECT EHR

    • redivis.com
    application/jsonl +7
    Updated Apr 19, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Shah Lab (2025). INSPECT EHR [Dataset]. http://doi.org/10.57761/ak51-d519
    Explore at:
    parquet, sas, csv, arrow, stata, avro, application/jsonl, spssAvailable download formats
    Dataset updated
    Apr 19, 2025
    Dataset provided by
    Redivis Inc.
    Authors
    Shah Lab
    Description

    Abstract

    The INSPECT dataset (Integrating Numerous Sources for Prognostic Evaluation of Clinical Timelines) contains de-identified longitudinal electronic health records (EHRs) from a large cohort of pulmonary embolism (PE) patients, along with ground truth labels for multiple outcomes. It includes 19,390 patients EHRs linked to 23,248 CTPA studies with paired radiology impressions.

    Methodology

    https://redivis.com/fileUploads/282601b3-2c4b-4de2-a84c-742037a916cd%3E" alt="inspect-logo.png">

    1. Overview

    INSPECT is a large-scale 3D multimodal medical imaging dataset:

    • 19,390 patients
    • 23,248 CT scans
    • 225+ million clinical events
    • 3 linked modalities

    %3C!-- --%3E

    2. CT Scans + Radiology Impression Notes

    Imaging data are available for download from the Stanford AIMI Center.

    3. EHR Data

    EHR data is sourced from Stanford’s STARR-OMOP database. Data are standardized in the OMOP CDM schema and are fully de-identified. Complete technical details are included in the paper, but key highlights:

    • Dates are jittered within patient to conceal real dates (but preserve deltas between dates)
    • Data for patients %3E= 90 years old are removed
    • Data for minors %3C18 are removed
    • Unstructured text fields not mappable to OMOP standard concepts are redacted
    • All clinical note text is redacted
    • HIV test result are redacted.
    • Provider names and NPIs are redacted

    %3C!-- --%3E

    Please see our Github repo to obtain code for loading the dataset, including a full data preprocessing pipeline for reproducibility, and running a set of pretrained baseline models

    Usage

    Access to the INSPECT dataset requires the following:

    • Verified Affiliation (Academic, Government, Industry Research Lab). Please use your verified email address when applying, do not use gmail or personal emails.
    • Encryption Verification / Attestation for Data Storage
    • Signing the terms of the INSPECT Data Set License 1.0
    • Providing a short description of your intended research use of INSPECT
    • CITI Training

    %3C!-- --%3E

    **These data must remain on your encrypted machine. Redistribution of data is FORBIDDEN and will result in immediate termination of access privileges. **

    IMPORTANT NOTES:

    • Our policy on derived works aligns with PhysioNet's guidelines, requiring that these artifacts be hosted on Redivis. If you create derived research artifacts based on INSPECT EHR (such as additional annotations or synthetic data), please contact us to discuss hosting arrangements.
    • Sending INSPECT data over a non-HIPAA-compliant API is a violation of the DUA.

    %3C!-- --%3E

    Please allow 7-10 business days to process applications.

  20. G

    Clinical Notes Image Matching

    • gomask.ai
    csv
    Updated Jul 12, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    GoMask.ai (2025). Clinical Notes Image Matching [Dataset]. https://gomask.ai/marketplace/datasets/clinical-notes-image-matching
    Explore at:
    csv(Unknown)Available download formats
    Dataset updated
    Jul 12, 2025
    Dataset provided by
    GoMask.ai
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Variables measured
    note_id, image_id, modality, body_part, diagnosis, note_date, note_text, image_date, image_type, patient_id, and 3 more
    Description

    This dataset provides a comprehensive mapping between de-identified clinical notes and their corresponding diagnostic images, enabling advanced research in multi-modal AI for healthcare. Each entry includes rich metadata for both text and imaging, supporting tasks such as automated diagnosis, cross-modal retrieval, and explainable AI in clinical settings.

Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
Nur Rusyidah Azri (2024). Medical Imaging Datasets for Multimodal Disease Detection and Diagnosis Research [Dataset]. https://ieee-dataport.org/documents/medical-imaging-datasets-multimodal-disease-detection-and-diagnosis-research

Medical Imaging Datasets for Multimodal Disease Detection and Diagnosis Research

Explore at:
Dataset updated
Aug 10, 2024
Authors
Nur Rusyidah Azri
License

Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically

Description

brain MRI

Search
Clear search
Close search
Google apps
Main menu