100+ datasets found
  1. T

    segment_anything

    • tensorflow.org
    Updated Dec 11, 2024
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    (2024). segment_anything [Dataset]. https://www.tensorflow.org/datasets/catalog/segment_anything
    Explore at:
    Dataset updated
    Dec 11, 2024
    Description

    SA-1B Download

    Segment Anything 1 Billion (SA-1B) is a dataset designed for training general-purpose object segmentation models from open world images. The dataset was introduced in the paper "Segment Anything".

    The SA-1B dataset consists of 11M diverse, high-resolution, licensed, and privacy-protecting images and 1.1B mask annotations. Masks are given in the COCO run-length encoding (RLE) format, and do not have classes.

    The license is custom. Please, read the full terms and conditions on https://ai.facebook.com/datasets/segment-anything-downloads.

    All the features are in the original dataset except image.content (content of the image).

    You can decode segmentation masks with:

    import tensorflow_datasets as tfds
    
    pycocotools = tfds.core.lazy_imports.pycocotools
    
    ds = tfds.load('segment_anything', split='train')
    for example in tfds.as_numpy(ds):
     segmentation = example['annotations']['segmentation']
     for counts, size in zip(segmentation['counts'], segmentation['size']):
      encoded_mask = {'size': size, 'counts': counts}
      mask = pycocotools.decode(encoded_mask) # np.array(dtype=uint8) mask
      ...
    

    To use this dataset:

    import tensorflow_datasets as tfds
    
    ds = tfds.load('segment_anything', split='train')
    for ex in ds.take(4):
     print(ex)
    

    See the guide for more informations on tensorflow_datasets.

  2. Data from: Segment Anything Model (SAM)

    • hub.arcgis.com
    • uneca.africageoportal.com
    Updated Apr 17, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Esri (2023). Segment Anything Model (SAM) [Dataset]. https://hub.arcgis.com/content/9b67b441f29f4ce6810979f5f0667ebe
    Explore at:
    Dataset updated
    Apr 17, 2023
    Dataset authored and provided by
    Esrihttp://esri.com/
    Area covered
    Description

    Segmentation models perform a pixel-wise classification by classifying the pixels into different classes. The classified pixels correspond to different objects or regions in the image. These models have a wide variety of use cases across multiple domains. When used with satellite and aerial imagery, these models can help to identify features such as building footprints, roads, water bodies, crop fields, etc.Generally, every segmentation model needs to be trained from scratch using a dataset labeled with the objects of interest. This can be an arduous and time-consuming task. Meta's Segment Anything Model (SAM) is aimed at creating a foundational model that can be used to segment (as the name suggests) anything using zero-shot learning and generalize across domains without additional training. SAM is trained on the Segment Anything 1-Billion mask dataset (SA-1B) which comprises a diverse set of 11 million images and over 1 billion masks. This makes the model highly robust in identifying object boundaries and differentiating between various objects across domains, even though it might have never seen them before. Use this model to extract masks of various objects in any image.Using the modelFollow the guide to use the model. Before using this model, ensure that the supported deep learning libraries are installed. For more details, check Deep Learning Libraries Installer for ArcGIS. Fine-tuning the modelThis model can be fine-tuned using SamLoRA architecture in ArcGIS. Follow the guide and refer to this sample notebook to fine-tune this model.Input8-bit, 3-band imagery.OutputFeature class containing masks of various objects in the image.Applicable geographiesThe model is expected to work globally.Model architectureThis model is based on the open-source Segment Anything Model (SAM) by Meta.Training dataThis model has been trained on the Segment Anything 1-Billion mask dataset (SA-1B) which comprises a diverse set of 11 million images and over 1 billion masks.Sample resultsHere are a few results from the model.

  3. segment-anything-fb

    • kaggle.com
    zip
    Updated Jan 9, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Jinttt (2024). segment-anything-fb [Dataset]. https://www.kaggle.com/datasets/jinttt/segment-anything-fb
    Explore at:
    zip(19167964 bytes)Available download formats
    Dataset updated
    Jan 9, 2024
    Authors
    Jinttt
    License

    Apache License, v2.0https://www.apache.org/licenses/LICENSE-2.0
    License information was derived automatically

    Description

    Dataset

    This dataset was created by Jinttt

    Released under Apache 2.0

    Contents

  4. h

    Enhancing-Segment-Anything-Model-with-Prioritized-Memory-For-Efficient-Image-Embeddings...

    • huggingface.co
    Updated Apr 1, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    pandey (2025). Enhancing-Segment-Anything-Model-with-Prioritized-Memory-For-Efficient-Image-Embeddings [Dataset]. https://huggingface.co/datasets/vinit000/Enhancing-Segment-Anything-Model-with-Prioritized-Memory-For-Efficient-Image-Embeddings
    Explore at:
    Dataset updated
    Apr 1, 2025
    Authors
    pandey
    License

    Apache License, v2.0https://www.apache.org/licenses/LICENSE-2.0
    License information was derived automatically

    Description

    Segment Anything Model (SAM) with Prioritized Memory Overview The Segment Anything Model (SAM) by Meta is a state-of-the-art image segmentation model leveraging vision transformers. However, it suffers from high memory usage and computational inefficiencies. Our research introduces a prioritized memory mechanism to enhance SAM’s performance while optimizing resource consumption. Methodology We propose a structured memory hierarchy to efficiently manage image embeddings and self-attention… See the full description on the dataset page: https://huggingface.co/datasets/vinit000/Enhancing-Segment-Anything-Model-with-Prioritized-Memory-For-Efficient-Image-Embeddings.

  5. Forgery Detection: Segment Anything with SAM2

    • kaggle.com
    zip
    Updated Nov 26, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Jirka Borovec (2025). Forgery Detection: Segment Anything with SAM2 [Dataset]. https://www.kaggle.com/datasets/jirkaborovec/forgery-segment-sam2-npz
    Explore at:
    zip(40460512 bytes)Available download formats
    Dataset updated
    Nov 26, 2025
    Authors
    Jirka Borovec
    License

    https://creativecommons.org/publicdomain/zero/1.0/https://creativecommons.org/publicdomain/zero/1.0/

    Description

    The Segment Anything Model (SAM) and its successor, SAM2, are highly influential foundation models in the field of computer vision, specifically designed for promptable image segmentation.

    What SAM Is

    • Foundation Model for Segmentation: SAM is a large, general-purpose model trained on a massive dataset to perform image segmentation. Its goal is to segment any object in any image, even those it hasn't seen before.
    • Prompt-Based Segmentation: Its core feature is the ability to generate a high-quality segmentation mask based on a minimal user "prompt." These prompts can include:
      • Clicks: The user clicks a point on an object, and SAM generates the mask for that entire object.
      • Bounding Boxes: The user draws a box around an area, and SAM refines it into a precise mask.
      • Text or Scribbles: (In more advanced versions like SAM2) The model can generate masks based on more complex inputs.

    Context in the Forgery Detection Competition

    In the context of the Kaggle competition, the use of SAM/SAM2 is for Candidate Generation:

    • The Problem: The competition, "Recod.ai/LUC - Scientific Image Forgery Detection," requires models to not only detect a forgery (like a copy-move manipulation in a biomedical image) but also to segment the manipulated region at the pixel level.
    • SAM/SAM2's Role: SAM is a powerful tool for generating segmentation masks for any prominent object or region in an image. The notebook you linked, "Forgery using SAM2 for Candidate Generation," leverages this ability to:
      1. Run SAM2 over the forged images to automatically generate a large collection of general segmentation masks (the "candidates").
      2. The idea is that the actual forged regions, which often involve copying and pasting distinct objects or image patches, are likely to be contained within or closely align with one of these automatically generated SAM/SAM2 masks.
    • Aiding Forgery Detection: By using SAM2, the notebook transforms the problem from searching for a forgery anywhere in the image to classifying and refining a smaller, more manageable set of high-quality object masks. This can significantly simplify the task for the final forgery-specific model.
  6. t

    Task-Aware Low-Rank Adaptation of Segment Anything Model - Dataset - LDM

    • service.tib.eu
    • resodate.org
    Updated Dec 2, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    (2024). Task-Aware Low-Rank Adaptation of Segment Anything Model - Dataset - LDM [Dataset]. https://service.tib.eu/ldmservice/dataset/task-aware-low-rank-adaptation-of-segment-anything-model
    Explore at:
    Dataset updated
    Dec 2, 2024
    Description

    The Segment Anything Model (SAM) has been proven to be a powerful foundation model for image segmentation tasks, which is an important task in computer vision. However, the transfer of its rich semantic information to multiple different downstream tasks remains unexplored. In this paper, we propose the Task-Aware Low-Rank Adaptation (TA-LoRA) method, which enables SAM to work as a foundation model for multi-task learning.

  7. r

    Data from: Robustness of SAM: Segment Anything under corruptions and beyond

    • resodate.org
    • service.tib.eu
    Updated Dec 3, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Lv Tang; Haoke Xiao; Bo Li (2024). Robustness of SAM: Segment Anything under corruptions and beyond [Dataset]. https://resodate.org/resources/aHR0cHM6Ly9zZXJ2aWNlLnRpYi5ldS9sZG1zZXJ2aWNlL2RhdGFzZXQvcm9idXN0bmVzcy1vZi1zYW0tLXNlZ21lbnQtYW55dGhpbmctdW5kZXItY29ycnVwdGlvbnMtYW5kLWJleW9uZA==
    Explore at:
    Dataset updated
    Dec 3, 2024
    Dataset provided by
    Leibniz Data Manager
    Authors
    Lv Tang; Haoke Xiao; Bo Li
    Description

    This work investigates the robustness of SAM to corruptions and adversarial attacks.

  8. t

    Segment Anything Model (SAM) for Medical Images - Dataset - LDM

    • service.tib.eu
    Updated Dec 16, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    (2024). Segment Anything Model (SAM) for Medical Images - Dataset - LDM [Dataset]. https://service.tib.eu/ldmservice/dataset/segment-anything-model--sam--for-medical-images
    Explore at:
    Dataset updated
    Dec 16, 2024
    Description

    Three publicly available medical imaging datasets: Breast Ultrasound Scan Images (BUSI), CVC-ClinicDB, and ISIC-2016.

  9. Segment Anything Model - v2

    • kaggle.com
    zip
    Updated Aug 10, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Mr. Ri and Ms. Tique (2024). Segment Anything Model - v2 [Dataset]. https://www.kaggle.com/datasets/mrriandmstique/segment-anything-model-v2
    Explore at:
    zip(1451090638 bytes)Available download formats
    Dataset updated
    Aug 10, 2024
    Authors
    Mr. Ri and Ms. Tique
    License

    https://creativecommons.org/publicdomain/zero/1.0/https://creativecommons.org/publicdomain/zero/1.0/

    Description

    Dataset

    This dataset was created by Mr. Ri and Ms. Tique

    Released under CC0: Public Domain

    Contents

  10. D

    EdgeSAM

    • researchdata.ntu.edu.sg
    pdf
    Updated Sep 12, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Chen Change Loy; Chen Change Loy (2024). EdgeSAM [Dataset]. http://doi.org/10.21979/N9/KF8798
    Explore at:
    pdf(14457216)Available download formats
    Dataset updated
    Sep 12, 2024
    Dataset provided by
    DR-NTU (Data)
    Authors
    Chen Change Loy; Chen Change Loy
    License

    https://researchdata.ntu.edu.sg/api/datasets/:persistentId/versions/2.0/customlicense?persistentId=doi:10.21979/N9/KF8798https://researchdata.ntu.edu.sg/api/datasets/:persistentId/versions/2.0/customlicense?persistentId=doi:10.21979/N9/KF8798

    Description

    We present EdgeSAM, an accelerated variant of the Segment Anything Model (SAM), optimized for efficient execution on edge devices with minimal compromise in performance. Our approach involves distilling the original ViT-based SAM image encoder into a purely CNN-based architecture, better suited for edge devices. We carefully benchmark various distillation strategies and demonstrate that taskagnostic encoder distillation fails to capture the full knowledge embodied in SAM. To overcome this bottleneck, we include both the prompt encoder and mask decoder in the distillation process, with box and point prompts in the loop, so that the distilled model can accurately capture the intricate dynamics between user input and mask generation. To mitigate dataset bias issues stemming from point prompt distillation, we incorporate a lightweight module within the encoder. As a result, EdgeSAM achieves a 37-fold speed increase compared to the original SAM, and it also outperforms MobileSAM/EfficientSAM, being over 7 times as fast when deployed on edge devices while enhancing the mIoUs on COCO and LVIS by 2.3/1.5 and 3.1/1.6, respectively. It is also the first SAM variant that can run at over 30 FPS on an iPhone 14.

  11. t

    GoodSAM: Bridging Domain and Capacity Gaps via Segment Anything Model -...

    • service.tib.eu
    Updated Dec 2, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    (2024). GoodSAM: Bridging Domain and Capacity Gaps via Segment Anything Model - Dataset - LDM [Dataset]. https://service.tib.eu/ldmservice/dataset/goodsam--bridging-domain-and-capacity-gaps-via-segment-anything-model
    Explore at:
    Dataset updated
    Dec 2, 2024
    Description

    This paper tackles a novel problem: how to transfer knowledge from the emerging Segment Anything Model (SAM) to learn a compact panoramic semantic segmentation model, i.e., student, without requiring any labeled data.

  12. PointPrompt: A Visual Prompting Dataset based on the Segment Anything Model

    • zenodo.org
    Updated Aug 4, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Jorge Quesada; Jorge Quesada; Zoe Fowler; Zoe Fowler; Mohammad Alotaibi; Mohit Prabhushankar; Mohit Prabhushankar; Ghassan AlRegib; Ghassan AlRegib; Mohammad Alotaibi (2024). PointPrompt: A Visual Prompting Dataset based on the Segment Anything Model [Dataset]. http://doi.org/10.5281/zenodo.11580815
    Explore at:
    Dataset updated
    Aug 4, 2024
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Jorge Quesada; Jorge Quesada; Zoe Fowler; Zoe Fowler; Mohammad Alotaibi; Mohit Prabhushankar; Mohit Prabhushankar; Ghassan AlRegib; Ghassan AlRegib; Mohammad Alotaibi
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Each folder in 'Prompting data.zip' corresponds to a single category (Bird, Cat, Bus etc), and each of these contain folders corresponding to a single participant (st1, st2 etc). Each participant folder should contain 5 subfolders:

    • 'masks' contains the binary masks produced for each image, in the format a_b_mask.png, where 'a' corresponds to the image number (0 to 399) and 'b' indexes through timestamps in the prompting process
    • 'points' contains the inclusion and exclusion points formatted as a_green.npy and a_red.npy respectively, where 'a' corresponds to the image number. Each of these files is a list of lists corresponding to the prompted points at each timestep. The outer list is of size (t,), where 't' is the number of timesteps for that image, an each inner list is fo size (n,2), where 'n' is the number of points at a given timestep
    • 'scores' contains the scores at each timestep for every image (mIoU)
    • 'sorts' contains sorted timestamp indexes, going from max to min based on the score
    • 'eachround' indicates which timesteps belong to each of the two rounds (if they exist). Each file contains a list of lenght t (number of timestamps) where values of 0 corresponds to timestamps that belong to the first round and values of 1 correspond to timestamps that belong to the second round

    Quick usage:

    -To get the best (highes score) mask for a given image : masks[sorts[0]]
    -To get the best set of prompts for that image : green[sorts[0]] and red[sorts[0]]
    -To get which round produced the highest score in that image : eachround[sorts[0]]

    The codebase associated with this work can be found at this Github.

    Please refer to our lab-wide github for more information regarding the code associated with our other papers.

  13. r

    Radiation Oncology Segment Anything Model (SAM) Dataset

    • resodate.org
    • service.tib.eu
    Updated Dec 3, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Lian Zhang; Zhengliang Liu; Lu Zhang; Zihao Wu; Xiaowei Yu; Jason Holmes; Hongying Feng; Haixing Dai; Xiang Li; Quanzheng Li; Dajiang Zhu; Tianming Liu; Wei Liu (2024). Radiation Oncology Segment Anything Model (SAM) Dataset [Dataset]. https://resodate.org/resources/aHR0cHM6Ly9zZXJ2aWNlLnRpYi5ldS9sZG1zZXJ2aWNlL2RhdGFzZXQvcmFkaWF0aW9uLW9uY29sb2d5LXNlZ21lbnQtYW55dGhpbmctbW9kZWwtLXNhbS0tZGF0YXNldA==
    Explore at:
    Dataset updated
    Dec 3, 2024
    Dataset provided by
    Leibniz Data Manager
    Authors
    Lian Zhang; Zhengliang Liu; Lu Zhang; Zihao Wu; Xiaowei Yu; Jason Holmes; Hongying Feng; Haixing Dai; Xiang Li; Quanzheng Li; Dajiang Zhu; Tianming Liu; Wei Liu
    Description

    The dataset used in this study for evaluating the performance of the Segment Anything Model (SAM) in clinical radiotherapy.

  14. Brain MRI segmentation Dataset Annotated with SAM

    • kaggle.com
    zip
    Updated May 21, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    nessim ben abbes (2023). Brain MRI segmentation Dataset Annotated with SAM [Dataset]. https://www.kaggle.com/datasets/nessimbenabbes/brain-mri-segmentation-dataset-annotated-with-sam
    Explore at:
    zip(6503189 bytes)Available download formats
    Dataset updated
    May 21, 2023
    Authors
    nessim ben abbes
    License

    Attribution-NonCommercial-ShareAlike 4.0 (CC BY-NC-SA 4.0)https://creativecommons.org/licenses/by-nc-sa/4.0/
    License information was derived automatically

    Description

    Description This dataset is a comprehensive collection of Brain Magnetic Resonance Imaging (MRI) scans, meticulously annotated with the Segment Anything Model (SAM). The data is stored in a CSV file format for easy access and manipulation.

    Content The dataset contains MRI scans of the brain, each of which is annotated with SAM. The annotations provide detailed information about the segmentation of various structures present in brain scans. The dataset is designed to aid in developing and validating algorithms for automatic brain structure segmentation.

  15. segment-anything-2

    • kaggle.com
    zip
    Updated Jul 30, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    emma123@bccto.cc (2024). segment-anything-2 [Dataset]. https://www.kaggle.com/datasets/emma123bcctocc/segment-anything-2
    Explore at:
    zip(1496138809 bytes)Available download formats
    Dataset updated
    Jul 30, 2024
    Authors
    emma123@bccto.cc
    Description
  16. Segmentation Living Image data

    • figshare.com
    zip
    Updated Oct 9, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    qianxiang yao (2023). Segmentation Living Image data [Dataset]. http://doi.org/10.6084/m9.figshare.24270154.v3
    Explore at:
    zipAvailable download formats
    Dataset updated
    Oct 9, 2023
    Dataset provided by
    figshare
    Figsharehttp://figshare.com/
    Authors
    qianxiang yao
    License

    https://www.apache.org/licenses/LICENSE-2.0.htmlhttps://www.apache.org/licenses/LICENSE-2.0.html

    Description

    This study introduces the concept of "structural beauty" as an objective computational approach for evaluating the aesthetic appeal of images. Through the utilization of the Segment anything model (SAM), we propose a method that leverages recursive segmentation to extract finer-grained substructures. Additionally, by reconstructing the hierarchical structure, we obtain a more accurate representation of substructure quantity and hierarchy. This approach reproduces and extends our previous research, allowing for the simultaneous assessment of Livingness in full-color images without the need for grayscale conversion or separate computations for foreground and background Livingness. Furthermore, the application of our method to the Scenic or Not dataset, a repository of subjective scenic ratings, demonstrates a high degree of consistency with subjective ratings in the 0-6 score range. This underscores that structural beauty is not solely a subjective perception, but a quantifiable attribute accessible through objective computation. Through our case studies, we have arrived at three significant conclusions. 1) our method demonstrates the capability to accurately segment meaningful objects, including trees, buildings, and windows, as well as abstract substructures within paintings. 2) we observed that the clarity of an image impacts our computational results; clearer images tend to yield higher Livingness scores. However, for equally blurry images, Livingness does not exhibit a significant reduction, aligning with human visual perception. 3) our approach fundamentally differs from methods employing Convolutional Neural Networks (CNNs) for predicting image scores. Our method not only provides computational results but also offers transparency and interpretability, positioning it as a novel avenue in the realm of Explainable AI (XAI).

  17. Data from: Segmentation evaluation.

    • plos.figshare.com
    • datasetcatalog.nlm.nih.gov
    xls
    Updated Oct 30, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Haoyu Wu; Clare Flynn; Carole Hall; Christian Che-Castaldo; Dimitris Samaras; Mathew Schwaller; Heather J. Lynch (2024). Segmentation evaluation. [Dataset]. http://doi.org/10.1371/journal.pone.0311038.t002
    Explore at:
    xlsAvailable download formats
    Dataset updated
    Oct 30, 2024
    Dataset provided by
    PLOShttp://plos.org/
    Authors
    Haoyu Wu; Clare Flynn; Carole Hall; Christian Che-Castaldo; Dimitris Samaras; Mathew Schwaller; Heather J. Lynch
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Evaluation of the the Segment Anything Model (SAM) for penguin colony segmentation using mean intersection over union (mIoU), difference in perimeter to area ratio (PAR), area error, and accuracy (i.e. panels a-c in Figs 3 and 4 vs. ground truth). 95% confidence intervals are shown. An up (down) arrow indicates a measure where a larger (smaller) number is preferred.

  18. r

    Real-Time All-Purpose Segment Anything Model

    • resodate.org
    • service.tib.eu
    Updated Dec 16, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Shilin Xu; Haobo Yuan; Qingyu Shi; Lu Qi; Jingbo Wang; Yibo Yang; Yining Li; Kai Chen; Yunhai Tong (2024). Real-Time All-Purpose Segment Anything Model [Dataset]. https://resodate.org/resources/aHR0cHM6Ly9zZXJ2aWNlLnRpYi5ldS9sZG1zZXJ2aWNlL2RhdGFzZXQvcmVhbC10aW1lLWFsbC1wdXJwb3NlLXNlZ21lbnQtYW55dGhpbmctbW9kZWw=
    Explore at:
    Dataset updated
    Dec 16, 2024
    Dataset provided by
    Leibniz Data Manager
    Authors
    Shilin Xu; Haobo Yuan; Qingyu Shi; Lu Qi; Jingbo Wang; Yibo Yang; Yining Li; Kai Chen; Yunhai Tong
    Description

    Advanced by transformer architecture, vision foundation models (VFMs) achieve remarkable progress in performance and generalization ability. Segment Anything Model (SAM) is one remarkable model that can achieve generalized segmentation. However, most VFMs cannot run in real-time, which makes it difficult to transfer them into several products.

  19. I

    auto_annotate

    • app.ikomia.ai
    Updated Jan 20, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Ikomia (2024). auto_annotate [Dataset]. https://app.ikomia.ai/hub/algorithms/auto_annotate/
    Explore at:
    Dataset updated
    Jan 20, 2024
    Dataset authored and provided by
    Ikomia
    License

    Apache License, v2.0https://www.apache.org/licenses/LICENSE-2.0
    License information was derived automatically

    Description

    Auto-annotate images with GroundingDINO and SAM models Auto-annotate images using a text prompt. GroundingDINO is employed for object detection (bounding boxes), followed by MobileSAM or SAM for segmentation. The annotations are then saved in both Pascal VOC format and COCO format....

  20. z

    Bugzz lightyears: To Semantic Segmentation and Bug-yond!

    • zenodo.org
    Updated Oct 26, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Maeen Alikarrar; Faizan Kazi; Maeen Alikarrar; Faizan Kazi (2024). Bugzz lightyears: To Semantic Segmentation and Bug-yond! [Dataset]. http://doi.org/10.5281/zenodo.13995425
    Explore at:
    Dataset updated
    Oct 26, 2024
    Dataset provided by
    zeonodo
    Authors
    Maeen Alikarrar; Faizan Kazi; Maeen Alikarrar; Faizan Kazi
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Dataset Title: Bugzz lightyears: To Semantic Segmentation and Bug-yond!

    Description:

    This dataset comprises a collection of real and robotic toy bugs designed for a small-scale semantic segmentation project. Each bug has been captured six times from various angles, ensuring comprehensive coverage of their features and details. The dataset serves as a valuable resource for exploring semantic segmentation techniques and evaluating machine learning models.

    Dataset Details:

    • Images: Each bug is represented by six images taken from different perspectives, facilitating robust segmentation and analysis.
    • Segmentation: The dataset has been meticulously segmented using Label Studio in conjunction with the SAM (Segment Anything Model), enabling precise delineation of each bug from the background.
    • Diversity: The collection includes a variety of bugs, both real and robotic, providing a unique blend for training and testing segmentation models.

    Usage: This toy dataset is ideal for researchers and developers interested in:

    • Experimenting with semantic segmentation algorithms.
    • Developing and refining computer vision models for object detection and segmentation.
    • Educational purposes in machine learning and computer vision courses.

    License: This dataset is made available under [specify license type, e.g., CC BY 4.0], allowing for both academic and commercial use, with proper attribution to the creator.

Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
(2024). segment_anything [Dataset]. https://www.tensorflow.org/datasets/catalog/segment_anything

segment_anything

Explore at:
102 scholarly articles cite this dataset (View in Google Scholar)
Dataset updated
Dec 11, 2024
Description

SA-1B Download

Segment Anything 1 Billion (SA-1B) is a dataset designed for training general-purpose object segmentation models from open world images. The dataset was introduced in the paper "Segment Anything".

The SA-1B dataset consists of 11M diverse, high-resolution, licensed, and privacy-protecting images and 1.1B mask annotations. Masks are given in the COCO run-length encoding (RLE) format, and do not have classes.

The license is custom. Please, read the full terms and conditions on https://ai.facebook.com/datasets/segment-anything-downloads.

All the features are in the original dataset except image.content (content of the image).

You can decode segmentation masks with:

import tensorflow_datasets as tfds

pycocotools = tfds.core.lazy_imports.pycocotools

ds = tfds.load('segment_anything', split='train')
for example in tfds.as_numpy(ds):
 segmentation = example['annotations']['segmentation']
 for counts, size in zip(segmentation['counts'], segmentation['size']):
  encoded_mask = {'size': size, 'counts': counts}
  mask = pycocotools.decode(encoded_mask) # np.array(dtype=uint8) mask
  ...

To use this dataset:

import tensorflow_datasets as tfds

ds = tfds.load('segment_anything', split='train')
for ex in ds.take(4):
 print(ex)

See the guide for more informations on tensorflow_datasets.

Search
Clear search
Close search
Google apps
Main menu