100+ datasets found
  1. t

    Robust CLIP-Based Detector for Exposing Diffusion Model-Generated Images -...

    • service.tib.eu
    • resodate.org
    Updated Dec 2, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    (2024). Robust CLIP-Based Detector for Exposing Diffusion Model-Generated Images - Dataset - LDM [Dataset]. https://service.tib.eu/ldmservice/dataset/robust-clip-based-detector-for-exposing-diffusion-model-generated-images
    Explore at:
    Dataset updated
    Dec 2, 2024
    Description

    Diffusion models (DMs) have revolutionized image generation, producing high-quality images with applications spanning various fields. However, their ability to create hyper-realistic images poses significant challenges in distinguishing between real and synthetic content, raising concerns about digital authenticity and potential misuse in creating deepfakes. This work introduces a robust detection framework that integrates image and text features extracted by CLIP model with a Multilayer Perceptron (MLP) classifier.

  2. Stable Diffusion 1.5 Images

    • kaggle.com
    zip
    Updated Nov 17, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Dev K11212 (2023). Stable Diffusion 1.5 Images [Dataset]. https://www.kaggle.com/datasets/devk11212/stable-diffusion-1-5-images
    Explore at:
    zip(4379938 bytes)Available download formats
    Dataset updated
    Nov 17, 2023
    Authors
    Dev K11212
    License

    MIT Licensehttps://opensource.org/licenses/MIT
    License information was derived automatically

    Description

    Dataset

    This dataset was created by Dev K11212

    Released under MIT

    Contents

  3. z

    Data from: Synthbuster: Towards Detection of Diffusion Model Generated...

    • zenodo.org
    • data.europa.eu
    zip
    Updated Nov 2, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Quentin Bammey; Quentin Bammey (2023). Synthbuster: Towards Detection of Diffusion Model Generated Images [Dataset]. http://doi.org/10.5281/zenodo.10066460
    Explore at:
    zipAvailable download formats
    Dataset updated
    Nov 2, 2023
    Dataset provided by
    IEEE Open Journal of Signal Processing
    Authors
    Quentin Bammey; Quentin Bammey
    License

    Attribution-NonCommercial-ShareAlike 4.0 (CC BY-NC-SA 4.0)https://creativecommons.org/licenses/by-nc-sa/4.0/
    License information was derived automatically

    Time period covered
    Sep 6, 2023
    Description

    Dataset described in the paper "Synthbuster: Towards Detection of Diffusion Model Generated Images" (Quentin Bammey, 2023, Open Journal of Signal Processing)

    This dataset contains synthetic, AI-generated images from 9 different models:

    • DALL·E 2
    • DALL·E 3
    • Adobe Firefly
    • Midjourney v5
    • Stable Diffusion 1.3
    • Stable Diffusion 1.4
    • Stable Diffusion 2
    • Stable Diffusion XL
    • Glide

    1000 images were generated per model. The images are loosely based on raise-1k images (Dang-Nguyen, Duc-Tien, et al. "Raise: A raw images dataset for digital image forensics." Proceedings of the 6th ACM multimedia systems conference. 2015.). For each image of the raise-1k dataset, a description was generated using the Midjourney /describe function and CLIP interrogator (https://github.com/pharmapsychotic/clip-interrogator/). Each of these prompts was manually edited to produce results as photorealistic as possible and remove living persons and artists names.

    In addition to this, parameters were randomly selected within reasonable values for methods requiring so.

    The prompts and parameters used for each method can be found in the `prompts.csv` file.

    This dataset can be used to evaluate AI-generated image detection methods. We recommend matching the generated images with the real Raise-1k images, to evaluate whether the methods can distinguish the two of them. Raise-1k images are not included in the dataset, they can be downloaded separately at (http://loki.disi.unitn.it/RAISE/download.html).

    None of the images suffered degradations such as JPEG compression or resampling, which leaves room to add your own degradations to test robustness to various transformation in a controlled manner.

  4. t

    Data from: DreamBooth: Fine Tuning Text-to-Image Diffusion Models for...

    • service.tib.eu
    Updated Dec 2, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    (2024). DreamBooth: Fine Tuning Text-to-Image Diffusion Models for Subject-Driven Generation [Dataset]. https://service.tib.eu/ldmservice/dataset/dreambooth--fine-tuning-text-to-image-diffusion-models-for-subject-driven-generation
    Explore at:
    Dataset updated
    Dec 2, 2024
    Description

    A dataset for subject-driven generation, containing 30 subjects, including objects and live subjects/pets.

  5. r

    Text-to-Image Diffusion Models

    • resodate.org
    • service.tib.eu
    Updated Dec 16, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Jindong Gu; Volker Tresp; Yao Qin (2024). Text-to-Image Diffusion Models [Dataset]. https://resodate.org/resources/aHR0cHM6Ly9zZXJ2aWNlLnRpYi5ldS9sZG1zZXJ2aWNlL2RhdGFzZXQvdGV4dC10by1pbWFnZS1kaWZmdXNpb24tbW9kZWxz
    Explore at:
    Dataset updated
    Dec 16, 2024
    Dataset provided by
    Leibniz Data Manager
    Authors
    Jindong Gu; Volker Tresp; Yao Qin
    Description

    The dataset used for text-to-image diffusion models, including Bluefire, Paintings, 3D, and Origami styles.

  6. R

    Stable Diffusion Images Dataset

    • universe.roboflow.com
    zip
    Updated Nov 24, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    planes (2024). Stable Diffusion Images Dataset [Dataset]. https://universe.roboflow.com/planes-zmdv1/stable-diffusion-images
    Explore at:
    zipAvailable download formats
    Dataset updated
    Nov 24, 2024
    Dataset authored and provided by
    planes
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Variables measured
    Planes Yyvq Bounding Boxes
    Description

    Stable Diffusion Images

    ## Overview
    
    Stable Diffusion Images is a dataset for object detection tasks - it contains Planes Yyvq annotations for 350 images.
    
    ## Getting Started
    
    You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.
    
      ## License
    
      This dataset is available under the [CC BY 4.0 license](https://creativecommons.org/licenses/CC BY 4.0).
    
  7. Synthetic Image Dataset of Five Object Classes Generated Using Stable...

    • figshare.com
    pdf
    Updated Jul 24, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Gurpreet Singh (2025). Synthetic Image Dataset of Five Object Classes Generated Using Stable Diffusion XL [Dataset]. http://doi.org/10.6084/m9.figshare.29640548.v1
    Explore at:
    pdfAvailable download formats
    Dataset updated
    Jul 24, 2025
    Dataset provided by
    figshare
    Figsharehttp://figshare.com/
    Authors
    Gurpreet Singh
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    This dataset contains 500 synthetic images generated via prompt-based text-to-image diffusion modeling using Stable Diffusion XL. Each image belongs to one of five classes: cat, dog, horse, car, and tree.Gurpreet, S. (2025). Synthetic Image Dataset of Five Object Classes Generated Using Stable Diffusion XL [Data set]. Zenodo. https://doi.org/10.5281/zenodo.16414387

  8. Stable ImageNet-1K

    • kaggle.com
    zip
    Updated Sep 8, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Vitaliy Kinakh (2022). Stable ImageNet-1K [Dataset]. https://www.kaggle.com/datasets/vitaliykinakh/stable-imagenet1k
    Explore at:
    zip(10519983320 bytes)Available download formats
    Dataset updated
    Sep 8, 2022
    Authors
    Vitaliy Kinakh
    License

    https://creativecommons.org/publicdomain/zero/1.0/https://creativecommons.org/publicdomain/zero/1.0/

    Description

    This dataset consists of images generated by Stable Diffusion v1.4 from diffusers library. 100 images per class. The prompt a photo of {class}, realistic, high quality was used, 50 sampling steps and 7.5 classifier guidance. Each image is 512x512 pixels.

  9. h

    Stable-Diffusion-Prompts

    • huggingface.co
    • opendatalab.com
    Updated Oct 8, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Gustavo (2022). Stable-Diffusion-Prompts [Dataset]. https://huggingface.co/datasets/Gustavosta/Stable-Diffusion-Prompts
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Oct 8, 2022
    Authors
    Gustavo
    License

    https://choosealicense.com/licenses/unknown/https://choosealicense.com/licenses/unknown/

    Description

    Stable Diffusion Dataset

    This is a set of about 80,000 prompts filtered and extracted from the image finder for Stable Diffusion: "Lexica.art". It was a little difficult to extract the data, since the search engine still doesn't have a public API without being protected by cloudflare. If you want to test the model with a demo, you can go to: "spaces/Gustavosta/MagicPrompt-Stable-Diffusion". If you want to see the model, go to: "Gustavosta/MagicPrompt-Stable-Diffusion".

  10. h

    stable-diffusion-xl-base-1.0-images

    • huggingface.co
    Updated Jun 27, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    dubm (2024). stable-diffusion-xl-base-1.0-images [Dataset]. https://huggingface.co/datasets/dubm/stable-diffusion-xl-base-1.0-images
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Jun 27, 2024
    Authors
    dubm
    Description

    dubm/stable-diffusion-xl-base-1.0-images dataset hosted on Hugging Face and contributed by the HF Datasets community

  11. D

    Data from: Efficient Diffusion Model for Image Restoration by Residual...

    • researchdata.ntu.edu.sg
    Updated Oct 4, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Zongsheng Yue; Zongsheng Yue; Jianyi Wang; Jianyi Wang; Chen Change Loy; Chen Change Loy (2024). Efficient Diffusion Model for Image Restoration by Residual Shifting [Dataset]. http://doi.org/10.21979/N9/VYPJ0O
    Explore at:
    Dataset updated
    Oct 4, 2024
    Dataset provided by
    DR-NTU (Data)
    Authors
    Zongsheng Yue; Zongsheng Yue; Jianyi Wang; Jianyi Wang; Chen Change Loy; Chen Change Loy
    License

    https://researchdata.ntu.edu.sg/api/datasets/:persistentId/versions/1.0/customlicense?persistentId=doi:10.21979/N9/VYPJ0Ohttps://researchdata.ntu.edu.sg/api/datasets/:persistentId/versions/1.0/customlicense?persistentId=doi:10.21979/N9/VYPJ0O

    Description

    While diffusion-based image restoration (IR) methods have achieved remarkable success, they are still limited by the low inference speed attributed to the necessity of executing hundreds or even thousands of sampling steps. Existing acceleration sampling techniques, though seeking to expedite the process, inevitably sacrifice performance to some extent, resulting in over-blurry restored outcomes. To address this issue, this study proposes a novel and efficient diffusion model for IR that significantly reduces the required number of diffusion steps. Our method avoids the need for post-acceleration during inference, thereby avoiding the associated performance deterioration. Specifically, our proposed method establishes a Markov chain that facilitates the transitions between the high-quality and low-quality images by shifting their residuals, substantially improving the transition efficiency. A carefully formulated noise schedule is devised to flexibly control the shifting speed and the noise strength during the diffusion process. Extensive experimental evaluations demonstrate that the proposed method achieves superior or comparable performance to current state-of-the-art methods on three classical IR tasks, namely image super-resolution, image inpainting, and blind face restoration, even only with four sampling steps.

  12. Stable Diffusion Face Dataset

    • kaggle.com
    Updated Apr 23, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Mohannad Ayman Salah (2024). Stable Diffusion Face Dataset [Dataset]. https://www.kaggle.com/datasets/mohannadaymansalah/stable-diffusion-dataaaaaaaaa
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Apr 23, 2024
    Dataset provided by
    Kagglehttp://kaggle.com/
    Authors
    Mohannad Ayman Salah
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    About the images:

    Fake Ai generated Human faces using Stable Diffusion 1.5, 2.1, and SDXL 1.0 checkpoint. The main objective was to generate photos that were as realistic as possible, without any specific style, focusing mainly on the face.

    Fake Ai generated Human faces

    • Images in 512x512px resolution were generated using SD 1.5;
    • Images in 768x768px resolution were generated using SD 2.1;
    • Images in 1024x1024px resolution were generated using SD XL 1.0;

    More details on the images and the process of creating the images in the readme file.

    The data is not mine, the data is taken from a GitHub repository to a user named: tobecwb Repo link: https://github.com/tobecwb/stable-diffusion-face-dataset

  13. r

    Data from: Prompting4Debugging: Red-Teaming Text-to-Image Diffusion Models...

    • resodate.org
    • service.tib.eu
    Updated Dec 2, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Zhi-Yi Chin; Chieh-Ming Jiang; Ching-Chun Huang; Pin-Yu Chen; Wei-Chen Chiu (2024). Prompting4Debugging: Red-Teaming Text-to-Image Diffusion Models by Finding Problematic Prompts [Dataset]. https://resodate.org/resources/aHR0cHM6Ly9zZXJ2aWNlLnRpYi5ldS9sZG1zZXJ2aWNlL2RhdGFzZXQvcHJvbXB0aW5nNGRlYnVnZ2luZy0tcmVkLXRlYW1pbmctdGV4dC10by1pbWFnZS1kaWZmdXNpb24tbW9kZWxzLWJ5LWZpbmRpbmctcHJvYmxlbWF0aWMtcHJvbXB0cw==
    Explore at:
    Dataset updated
    Dec 2, 2024
    Dataset provided by
    Leibniz Data Manager
    Authors
    Zhi-Yi Chin; Chieh-Ming Jiang; Ching-Chun Huang; Pin-Yu Chen; Wei-Chen Chiu
    Description

    Text-to-image diffusion models, e.g. Stable Diffusion (SD), lately have shown remarkable ability in high-quality content generation, and become one of the representatives for the recent wave of transformative AI.

  14. t

    eDiff-I: Text-to-image diffusion models with an ensemble of expert denoisers...

    • service.tib.eu
    Updated Dec 16, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    (2024). eDiff-I: Text-to-image diffusion models with an ensemble of expert denoisers - Dataset - LDM [Dataset]. https://service.tib.eu/ldmservice/dataset/ediff-i--text-to-image-diffusion-models-with-an-ensemble-of-expert-denoisers
    Explore at:
    Dataset updated
    Dec 16, 2024
    Description

    Text-to-image diffusion models with an ensemble of expert denoisers.

  15. DrawBench

    • opendatalab.com
    zip
    Updated Jan 1, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Google Research (2022). DrawBench [Dataset]. https://opendatalab.com/OpenDataLab/DrawBench
    Explore at:
    zipAvailable download formats
    Dataset updated
    Jan 1, 2022
    Dataset provided by
    Google Research
    Googlehttp://google.com/
    Description

    We present Imagen, a text-to-image diffusion model with an unprecedented degree of photorealism and a deep level of language understanding. Imagen builds on the power of large transformer language models in understanding text and hinges on the strength of diffusion models in high-fidelity image generation. Our key discovery is that generic large language models (e.g. T5), pretrained on text-only corpora, are surprisingly effective at encoding text for image synthesis: increasing the size of the language model in Imagen boosts both sample fidelity and imagetext alignment much more than increasing the size of the image diffusion model. Imagen achieves a new state-of-the-art FID score of 7.27 on the COCO dataset, without ever training on COCO, and human raters find Imagen samples to be on par with the COCO data itself in image-text alignment. To assess text-to-image models in greater depth, we introduce DrawBench, a comprehensive and challenging benchmark for text-to-image models. With DrawBench, we compare Imagen with recent methods including VQ-GAN+CLIP, Latent Diffusion Models, GLIDE and DALL-E 2, and find that human raters prefer Imagen over other models in side-byside comparisons, both in terms of sample quality and image-text alignment.

  16. m

    CivitAI AI Models

    • modelslab.com
    Updated Aug 17, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    CivitAI Community (2025). CivitAI AI Models [Dataset]. https://modelslab.com/civitai
    Explore at:
    Dataset updated
    Aug 17, 2025
    Dataset provided by
    ModelsLab
    Authors
    CivitAI Community
    Description

    Collection of 100,000+ CivitAI models including Stable Diffusion checkpoints, LoRA models, Textual Inversions, and VAE models for text-to-image generation

  17. A sample set of original and generated images using the 14 cluster model.

    • figshare.com
    zip
    Updated Jul 17, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Naoaki ONO (2025). A sample set of original and generated images using the 14 cluster model. [Dataset]. http://doi.org/10.6084/m9.figshare.29588849.v1
    Explore at:
    zipAvailable download formats
    Dataset updated
    Jul 17, 2025
    Dataset provided by
    figshare
    Figsharehttp://figshare.com/
    Authors
    Naoaki ONO
    License

    CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
    License information was derived automatically

    Description

    Generative image models have revolutionized artificial intelligence by enabling the synthesis of high-quality, realistic images. These models utilize deep learning techniques to learn complex data distributions and generate novel images that closely resemble the training dataset. Recent advancements, particularly in diffusion models, have led to remarkable improvements in image fidelity, diversity, and controllability. In this work, we investigate the application of a conditional latent diffusion model in the healthcare domain. Specifically, we trained a latent diffusion model using unlabeled histopathology images. Initially, these images were embedded into a lower-dimensional latent space using a Vector Quantized Generative Adversarial Network (VQ-GAN). Subsequently, a diffusion process was applied within this latent space, and clustering was performed on the resulting latent features. The clustering results were then used as a conditioning mechanism for the diffusion model, enabling conditional image generation. Finally, we determined the optimal number of clusters using cluster validation metrics and assessed the quality of the synthetic images through quantitative methods. To enhance the interpretability of the synthetic image generation process, expert input was incorporated into the cluster assignments.

  18. f

    Trained model parameters.

    • figshare.com
    zip
    Updated Jul 17, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Naoaki ONO (2025). Trained model parameters. [Dataset]. http://doi.org/10.6084/m9.figshare.29588807.v1
    Explore at:
    zipAvailable download formats
    Dataset updated
    Jul 17, 2025
    Dataset provided by
    figshare
    Authors
    Naoaki ONO
    License

    CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
    License information was derived automatically

    Description

    Generative image models have revolutionized artificial intelligence by enabling the synthesis of high-quality, realistic images. These models utilize deep learning techniques to learn complex data distributions and generate novel images that closely resemble the training dataset. Recent advancements, particularly in diffusion models, have led to remarkable improvements in image fidelity, diversity, and controllability. In this work, we investigate the application of a conditional latent diffusion model in the healthcare domain. Specifically, we trained a latent diffusion model using unlabeled histopathology images. Initially, these images were embedded into a lower-dimensional latent space using a Vector Quantized Generative Adversarial Network (VQ-GAN). Subsequently, a diffusion process was applied within this latent space, and clustering was performed on the resulting latent features. The clustering results were then used as a conditioning mechanism for the diffusion model, enabling conditional image generation. Finally, we determined the optimal number of clusters using cluster validation metrics and assessed the quality of the synthetic images through quantitative methods. To enhance the interpretability of the synthetic image generation process, expert input was incorporated into the cluster assignments.

  19. h

    stable-diffusion-2-1-without-images

    • huggingface.co
    Updated Jul 12, 2023
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    gryffindor (2023). stable-diffusion-2-1-without-images [Dataset]. https://huggingface.co/datasets/gryffindor-ISWS/stable-diffusion-2-1-without-images
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Jul 12, 2023
    Dataset authored and provided by
    gryffindor
    License

    https://choosealicense.com/licenses/gpl-3.0/https://choosealicense.com/licenses/gpl-3.0/

    Description

    gryffindor-ISWS/stable-diffusion-2-1-without-images dataset hosted on Hugging Face and contributed by the HF Datasets community

  20. D

    Data from: Arbitrary-steps Image Super-resolution via Diffusion Inversion

    • researchdata.ntu.edu.sg
    Updated Mar 11, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Zongsheng Yue; Zongsheng Yue; Kang Liao; Kang Liao; Chen Change Loy; Chen Change Loy (2025). Arbitrary-steps Image Super-resolution via Diffusion Inversion [Dataset]. http://doi.org/10.21979/N9/SZJQME
    Explore at:
    Dataset updated
    Mar 11, 2025
    Dataset provided by
    DR-NTU (Data)
    Authors
    Zongsheng Yue; Zongsheng Yue; Kang Liao; Kang Liao; Chen Change Loy; Chen Change Loy
    License

    https://researchdata.ntu.edu.sg/api/datasets/:persistentId/versions/1.0/customlicense?persistentId=doi:10.21979/N9/SZJQMEhttps://researchdata.ntu.edu.sg/api/datasets/:persistentId/versions/1.0/customlicense?persistentId=doi:10.21979/N9/SZJQME

    Dataset funded by
    RIE2020 Industry Alignment Fund– Industry Collaboration Projects (IAF-ICP) Funding Initiative
    Description

    This study presents a new image super-resolution (SR) technique based on diffusion inversion, aiming at harnessing the rich image priors encapsulated in large pre-trained diffusion models to improve SR performance. We design a Partial noise Prediction strategy to construct an intermediate state of the diffusion model, which serves as the starting sampling point. Central to our approach is a deep noise predictor to estimate the optimal noise maps for the forward diffusion process. Once trained, this noise predictor can be used to initialize the sampling process partially along the diffusion trajectory, generating the desirable high-resolution result. Compared to existing approaches, our method offers a flexible and efficient sampling mechanism that supports an arbitrary number of sampling steps, ranging from one to five. Even with a single sampling step, our method demonstrates superior or comparable performance to recent state-of-the-art approaches.

Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
(2024). Robust CLIP-Based Detector for Exposing Diffusion Model-Generated Images - Dataset - LDM [Dataset]. https://service.tib.eu/ldmservice/dataset/robust-clip-based-detector-for-exposing-diffusion-model-generated-images

Robust CLIP-Based Detector for Exposing Diffusion Model-Generated Images - Dataset - LDM

Explore at:
Dataset updated
Dec 2, 2024
Description

Diffusion models (DMs) have revolutionized image generation, producing high-quality images with applications spanning various fields. However, their ability to create hyper-realistic images poses significant challenges in distinguishing between real and synthetic content, raising concerns about digital authenticity and potential misuse in creating deepfakes. This work introduces a robust detection framework that integrates image and text features extracted by CLIP model with a Multilayer Perceptron (MLP) classifier.

Search
Clear search
Close search
Google apps
Main menu