4 datasets found
  1. p

    Visual Question Answering evaluation dataset for MIMIC CXR

    • physionet.org
    Updated Jan 28, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Timo Kohlberger; Charles Lau; Tom Pollard; Andrew Sellergren; Atilla Kiraly; Fayaz Jamil (2025). Visual Question Answering evaluation dataset for MIMIC CXR [Dataset]. http://doi.org/10.13026/cvsk-ny21
    Explore at:
    Dataset updated
    Jan 28, 2025
    Authors
    Timo Kohlberger; Charles Lau; Tom Pollard; Andrew Sellergren; Atilla Kiraly; Fayaz Jamil
    License

    https://github.com/MIT-LCP/license-and-dua/tree/master/draftshttps://github.com/MIT-LCP/license-and-dua/tree/master/drafts

    Description

    MIMIC CXR [1] is a large publicly available dataset of chest radiographs in DICOM format with free-text radiology reports. In addition, labels for the presence of 12 different chest-related pathologies, as well as of any support devices, and overall normal/abnormal status were made available via the MIMIC Chest X-ray JPG (MIMIC-CXR-JPG) [2] labels, which were generated using the CheXpert and NegBio algorithms.

    Based on these labels, we created a visual question answering dataset comprising 224 questions for 48 cases from the official test set, and 111 questions for 23 validation cases. A majority (68%) of the questions are close-ended (answerable with yes or no), and focus on the presence of one out of 15 chest pathologies, or any support device, or generically on any abnormality, whereas the remaining open-ended questions inquire about the location, size, severity or type of a pathology/device, if present in the specific case, indicated by the MIMIC-CXR-JPG labels.

    For each question and case we also provide a reference answer, which was authored by a board-certified radiologist (with 17 years of post-residency experience) based on the chest X-ray and original radiology report

  2. h

    MIMIC-CXR-files

    • huggingface.co
    Updated Mar 23, 2025
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Pinxin Liu (2025). MIMIC-CXR-files [Dataset]. https://huggingface.co/datasets/pliu23/MIMIC-CXR-files
    Explore at:
    Dataset updated
    Mar 23, 2025
    Authors
    Pinxin Liu
    Description

    This will download the MIMIC-CXR files. To download the full data. You can go to MIMIC-CXR. At the end of the page, you will see the structure of the folder. files/ p1/ xxx.jpg p2/ p2/ IMAGE_FILENMAESLICENCES.txt This repo will present all the needed image data under the fils directory. For other data, you still need to request them from the website.

  3. p

    Data from: CheXmask Database: a large-scale dataset of anatomical...

    • physionet.org
    Updated Mar 1, 2024
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Nicolas Gaggion; Candelaria Mosquera; Martina Aineseder; Lucas Mansilla; Diego Milone; Enzo Ferrante (2024). CheXmask Database: a large-scale dataset of anatomical segmentation masks for chest x-ray images [Dataset]. http://doi.org/10.13026/pgag-by42
    Explore at:
    Dataset updated
    Mar 1, 2024
    Authors
    Nicolas Gaggion; Candelaria Mosquera; Martina Aineseder; Lucas Mansilla; Diego Milone; Enzo Ferrante
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    The CheXmask Database presents a comprehensive, uniformly annotated collection of chest radiographs, constructed from five public databases: ChestX-ray8, Chexpert, MIMIC-CXR-JPG, Padchest and VinDr-CXR. The database aggregates 657,566 anatomical segmentation masks derived from images which have been processed using the HybridGNet model to ensure consistent, high-quality segmentation. To confirm the quality of the segmentations, we include in this database individual Reverse Classification Accuracy (RCA) scores for each of the segmentation masks. This dataset is intended to catalyze further innovation and refinement in the field of semantic chest X-ray analysis, offering a significant resource for researchers in the medical imaging domain.

  4. Dataset for Detection and Segmentation of the Radiographic Features of...

    • zenodo.org
    png, text/x-python +1
    Updated Dec 14, 2024
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Viacheslav Danilov; Viacheslav Danilov (2024). Dataset for Detection and Segmentation of the Radiographic Features of Pulmonary Edema [Dataset]. http://doi.org/10.5281/zenodo.8390417
    Explore at:
    zip, png, text/x-pythonAvailable download formats
    Dataset updated
    Dec 14, 2024
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Viacheslav Danilov; Viacheslav Danilov
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Objectives: This comprehensive dataset is well suited for training, evaluating, and using machine learning models to detect, segment, and analyze radiological features associated with pulmonary edema in chest X-ray images.

    Description: This dataset consists of a collection of chest X-rays extracted from the MIMIC database, carefully collected at the Beth Israel Deaconess Medical Center. In total, it comprises 1000 chest X-rays obtained from 741 patients with features suggestive of edema. These X-rays were carefully selected for manual annotation. The annotations are rich and detailed, covering specific radiological features commonly associated with pulmonary edema, including cephalization, Kerley lines, pleural effusions, bat wings, and infiltrates. The dataset includes a wide variety of radiological features, with a total of 4263 annotations (Table 1). Furthermore, each chest radiograph is thoughtfully assigned a severity category, categorizing it as "no edema", "vascular congestion", "interstitial edema", or "alveolar edema".

    Annotation Method: The annotation process was meticulously performed by a highly qualified clinician with over 10 years of radiology experience, utilizing both frontal and lateral views for each chest X-ray study. Cephalization and Kerley lines were delineated using polylines, while other features were delineated using binary masks. This methodological approach was carefully chosen to provide a comprehensive data set that would ensure accuracy in subsequent analyses and label assignments.

    Notably, all features are represented as bounding boxes, meticulously defined by their respective upper-left (x1; y1) and lower-right (x2; y2) corners. In addition, selected features are provided with masks encoded in base 64 format. To facilitate seamless decoding, we provide a conversion script called "mask_converter.py" that allows the transformation of encoded masks into a versatile numpy array format. This feature improves the usability of the dataset for precise analysis and deep learning applications.

    Datasets:

    1. SLY dataset: The dataset contains chest X-ray images labeled by clinicians, including both stacked frontal and lateral images. We obtained this dataset by annotating it on the Supervisely platform, and it is stored in JSON and PNG formats.
    2. Source dataset: The dataset is a transformed version of the SLY dataset. In this dataset, all annotations are consolidated into a single spreadsheet, and only frontal view images are represented.
    3. Processed dataset: The dataset focuses exclusively on the lung area for analysis, as other areas surrounding the lung typically contain extraneous information that clinicians do not use in their decision-making process.
    4. COCO dataset: A collection of subsets prepared in the COCO format and suitable for training and testing. It includes subsets for each feature and for all features evaluated in this study.

    Access to the Study: Further information about this study, including curated source code, dataset details, and trained models, can be accessed through the following repositories:

    Table 1. Summary of annotated radiological features and severity labels

    Radiological feature

    Number of objects

    Severity label

    Number of cases

    Cephalization

    1656

    No edema

    21

    Kerley line

    609

    Vascular congestion

    74

    Pleural effusion

    317

    Interstitial edema

    51

    Bat wing

    1604

    Alveolar edema

    595

    Infiltrate

    77

    TOTAL

    4263

    TOTAL

    741

  5. Not seeing a result you expected?
    Learn how you can add new datasets to our index.

Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
Timo Kohlberger; Charles Lau; Tom Pollard; Andrew Sellergren; Atilla Kiraly; Fayaz Jamil (2025). Visual Question Answering evaluation dataset for MIMIC CXR [Dataset]. http://doi.org/10.13026/cvsk-ny21

Visual Question Answering evaluation dataset for MIMIC CXR

Explore at:
Dataset updated
Jan 28, 2025
Authors
Timo Kohlberger; Charles Lau; Tom Pollard; Andrew Sellergren; Atilla Kiraly; Fayaz Jamil
License

https://github.com/MIT-LCP/license-and-dua/tree/master/draftshttps://github.com/MIT-LCP/license-and-dua/tree/master/drafts

Description

MIMIC CXR [1] is a large publicly available dataset of chest radiographs in DICOM format with free-text radiology reports. In addition, labels for the presence of 12 different chest-related pathologies, as well as of any support devices, and overall normal/abnormal status were made available via the MIMIC Chest X-ray JPG (MIMIC-CXR-JPG) [2] labels, which were generated using the CheXpert and NegBio algorithms.

Based on these labels, we created a visual question answering dataset comprising 224 questions for 48 cases from the official test set, and 111 questions for 23 validation cases. A majority (68%) of the questions are close-ended (answerable with yes or no), and focus on the presence of one out of 15 chest pathologies, or any support device, or generically on any abnormality, whereas the remaining open-ended questions inquire about the location, size, severity or type of a pathology/device, if present in the specific case, indicated by the MIMIC-CXR-JPG labels.

For each question and case we also provide a reference answer, which was authored by a board-certified radiologist (with 17 years of post-residency experience) based on the chest X-ray and original radiology report

Search
Clear search
Close search
Google apps
Main menu