20 datasets found
  1. Z

    COCO dataset and neural network weights for micro-FTIR particle detection on...

    • data.niaid.nih.gov
    Updated Aug 13, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Schowing, Thibault (2024). COCO dataset and neural network weights for micro-FTIR particle detection on filters. [Dataset]. https://data.niaid.nih.gov/resources?id=zenodo_10839526
    Explore at:
    Dataset updated
    Aug 13, 2024
    Dataset authored and provided by
    Schowing, Thibault
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    The IMPTOX project has received funding from the EU's H2020 framework programme for research and innovation under grant agreement n. 965173. Imptox is part of the European MNP cluster on human health.

    More information about the project here.

    Description: This repository includes the trained weights and a custom COCO-formatted dataset used for developing and testing a Faster R-CNN R_50_FPN_3x object detector, specifically designed to identify particles in micro-FTIR filter images.

    Contents:

    Weights File (neuralNetWeights_V3.pth):

    Format: .pth

    Description: This file contains the trained weights for a Faster R-CNN model with a ResNet-50 backbone and a Feature Pyramid Network (FPN), trained for 3x schedule. These weights are specifically tuned for detecting particles in micro-FTIR filter images.

    Custom COCO Dataset (uFTIR_curated_square.v5-uftir_curated_square_2024-03-14.coco-segmentation.zip):

    Format: .zip

    Description: This zip archive contains a custom COCO-formatted dataset, including JPEG images and their corresponding annotation file. The dataset consists of images of micro-FTIR filters with annotated particles.

    Contents:

    Images: JPEG format images of micro-FTIR filters.

    Annotations: A JSON file in COCO format providing detailed annotations of the particles in the images.

    Management: The dataset can be managed and manipulated using the Pycocotools library, facilitating easy integration with existing COCO tools and workflows.

    Applications: The provided weights and dataset are intended for researchers and practitioners in the field of microscopy and particle detection. The dataset and model can be used for further training, validation, and fine-tuning of object detection models in similar domains.

    Usage Notes:

    The neuralNetWeights_V3.pth file should be loaded into a PyTorch model compatible with the Faster R-CNN architecture, such as Detectron2.

    The contents of uFTIR_curated_square.v5-uftir_curated_square_2024-03-14.coco-segmentation.zip should be extracted and can be used with any COCO-compatible object detection framework for training and evaluation purposes.

    Code can be found on the related Github repository.

  2. SPEECH-COCO

    • zenodo.org
    • data.niaid.nih.gov
    xz, zip
    Updated Nov 24, 2020
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    William N. Havard; William N. Havard; Laurent Besacier; Laurent Besacier (2020). SPEECH-COCO [Dataset]. http://doi.org/10.5281/zenodo.4282267
    Explore at:
    zip, xzAvailable download formats
    Dataset updated
    Nov 24, 2020
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    William N. Havard; William N. Havard; Laurent Besacier; Laurent Besacier
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    SpeechCoco

    Introduction

    Our corpus is an extension of the MS COCO image recognition and captioning dataset. MS COCO comprises images paired with a set of five captions. Yet, it does not include any speech. Therefore, we used Voxygen's text-to-speech system to synthesise the available captions.

    The addition of speech as a new modality enables MSCOCO to be used for researches in the field of language acquisition, unsupervised term discovery, keyword spotting, or semantic embedding using speech and vision.

    Our corpus is licensed under a Creative Commons Attribution 4.0 License.

    Data Set

    • This corpus contains 616,767 spoken captions from MSCOCO's val2014 and train2014 subsets (respectively 414,113 for train2014 and 202,654 for val2014).

    • We used 8 different voices. 4 of them have a British accent (Paul, Bronwen, Judith, and Elizabeth) and the 4 others have an American accent (Phil, Bruce, Amanda, Jenny).

    • In order to make the captions sound more natural, we used SOX tempo command, enabling us to change the speed without changing the pitch. 1/3 of the captions are 10% slower than the original pace, 1/3 are 10% faster. The last third of the captions was kept untouched.

    • We also modified approximately 30% of the original captions and added disfluencies such as "um", "uh", "er" so that the captions would sound more natural.

    • Each WAV file is paired with a JSON file containing various information: timecode of each word in the caption, name of the speaker, name of the WAV file, etc. The JSON files have the following data structure:

    {
      "duration": float,
      "speaker": string,
      "synthesisedCaption": string,
      "timecode": list,
      "speed": float,
      "wavFilename": string,
      "captionID": int,
      "imgID": int,
      "disfluency": list
    }
    • On average, each caption comprises 10.79 tokens, disfluencies included. The WAV files are on average 3.52 seconds long.

    Repository

    The repository is organized as follows:

    • CORPUS-MSCOCO (~75GB once decompressed)

      • train2014/ : folder contains 413,915 captions

        • json/

        • wav/

        • translations/

          • train_en_ja.txt

          • train_translate.sqlite3

        • train_2014.sqlite3

      • val2014/ : folder contains 202,520 captions

        • json/

        • wav/

        • translations/

          • train_en_ja.txt

          • train_translate.sqlite3

        • val_2014.sqlite3

      • speechcoco_API/

        • speechcoco/

          • _init_.py

          • speechcoco.py

        • setup.py

    Filenames

    .wav files contain the spoken version of a caption

    .json files contain all the metadata of a given WAV file

    .sqlite3 files are SQLite databases containing all the information contained in the JSON files

    We adopted the following naming convention for both the WAV and JSON files:

    imageID_captionID_Speaker_DisfluencyPosition_Speed[.wav/.json]

    Script

    We created a script called speechcoco.py in order to handle the metadata and allow the user to easily find captions according to specific filters. The script uses the *.db files.

    Features:

    • Aggregate all the information in the JSON files into a single SQLite database

    • Find captions according to specific filters (name, gender and nationality of the speaker, disfluency position, speed, duration, and words in the caption). The script automatically builds the SQLite query. The user can also provide his own SQLite query.

    The following Python code returns all the captions spoken by a male with an American accent for which the speed was slowed down by 10% and that contain "keys" at any position

    # create SpeechCoco object
    db = SpeechCoco(train_2014.sqlite3, train_translate.sqlite3, verbose=True)
    
    # filter captions (returns Caption Objects)
    captions = db.filterCaptions(gender="Male", nationality="US", speed=0.9, text='%keys%')
    for caption in captions:
      print('
    {}\t{}\t{}\t{}\t{}\t{}\t\t{}'.format(caption.imageID,
                             caption.captionID,
                             caption.speaker.name,
                             caption.speaker.nationality,
                             caption.speed,
                             caption.filename,
                             caption.text))
    ...
    298817   26763  Phil  0.9   298817_26763_Phil_None_0-9.wav     A group of turkeys with bushes in the background.
    108505   147972 Phil  0.9   108505_147972_Phil_Middle_0-9.wav        Person using a, um, slider cell phone with blue backlit keys.
    258289   154380 Bruce  0.9   258289_154380_Bruce_None_0-9.wav        Some donkeys and sheep are in their green pens .
    545312   201303 Phil  0.9   545312_201303_Phil_None_0-9.wav     A man walking next to a couple of donkeys.
    ...
    • Find all the captions belonging to a specific image

    captions = db.getImgCaptions(298817)
    for caption in captions:
      print('
    {}'.format(caption.text))
    Birds wondering through grassy ground next to bushes.
    A flock of turkeys are making their way up a hill.
    Um, ah. Two wild turkeys in a field walking around.
    Four wild turkeys and some bushes trees and weeds.
    A group of turkeys with bushes in the background.
    • Parse the timecodes and have them structured

    input:

    ...
    [1926.3068, "SYL", ""],
    [1926.3068, "SEPR", " "],
    [1926.3068, "WORD", "white"],
    [1926.3068, "PHO", "w"],
    [2050.7955, "PHO", "ai"],
    [2144.6591, "PHO", "t"],
    [2179.3182, "SYL", ""],
    [2179.3182, "SEPR", " "]
    ...

    output:

    print(caption.timecode.parse())
    ...
    {
    'begin': 1926.3068,
    'end': 2179.3182,
    'syllable': [{'begin': 1926.3068,
           'end': 2179.3182,
           'phoneme': [{'begin': 1926.3068,
                  'end': 2050.7955,
                  'value': 'w'},
                 {'begin': 2050.7955,
                  'end': 2144.6591,
                  'value': 'ai'},
                 {'begin': 2144.6591,
                  'end': 2179.3182,
                  'value': 't'}],
           'value': 'wait'}],
    'value': 'white'
    },
    ...
    • Convert the timecodes to Praat TextGrid files

    caption.timecode.toTextgrid(outputDir, level=3)
    • Get the words, syllables and phonemes between n seconds/milliseconds

    The following Python code returns all the words between 0.2 and 0.6 seconds for which at least 50% of the word's total length is within the specified interval

    pprint(caption.getWords(0.20, 0.60, seconds=True, level=1, olapthr=50))
    ...
    404537   827239 Bruce  US   0.9   404537_827239_Bruce_None_0-9.wav        Eyeglasses, a cellphone, some keys and other pocket items are all laid out on the cloth. .
    [
      {
        'begin': 0.0,
        'end': 0.7202778,
        'overlapPercentage': 55.53412863758955,
        'word': 'eyeglasses'
      }
    ]
     ...
    • Get the translations of the selected captions

    As for now, only japanese translations are available. We also used Kytea to tokenize and tag the captions translated with Google Translate

    captions = db.getImgCaptions(298817)
    for caption in captions:
      print('
    {}'.format(caption.text))
    
      # Get translations and POS
      print('\tja_google: {}'.format(db.getTranslation(caption.captionID, "ja_google")))
      print('\t\tja_google_tokens: {}'.format(db.getTokens(caption.captionID, "ja_google")))
      print('\t\tja_google_pos: {}'.format(db.getPOS(caption.captionID, "ja_google")))
      print('\tja_excite: {}'.format(db.getTranslation(caption.captionID, "ja_excite")))

      Birds wondering through grassy ground next to bushes.
      ja_google: 鳥は茂みの下に茂った地面を抱えています。
        ja_google_tokens: 鳥 は 茂み の 下 に 茂 っ た 地面 を 抱え て い ま す 。
        ja_google_pos: 鳥/名詞/とり は/助詞/は 茂み/名詞/しげみ の/助詞/の 下/名詞/した に/助詞/に

  3. f

    IoU and HD of UNets family when the optimizer is set as Adam. lr: learning...

    • plos.figshare.com
    xls
    Updated Jun 11, 2025
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Bang-Chuan Chen; Chung-Yi Shen; Jyh-Wen Chai; Ren-Hung Hwang; Wei-Chuan Chiang; Chi-Hsiang Chou; Wei-Min Liu (2025). IoU and HD of UNets family when the optimizer is set as Adam. lr: learning rate. The best-performing results are highlighted using bold font. [Dataset]. http://doi.org/10.1371/journal.pone.0323692.t002
    Explore at:
    xlsAvailable download formats
    Dataset updated
    Jun 11, 2025
    Dataset provided by
    PLOS ONE
    Authors
    Bang-Chuan Chen; Chung-Yi Shen; Jyh-Wen Chai; Ren-Hung Hwang; Wei-Chuan Chiang; Chi-Hsiang Chou; Wei-Min Liu
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    IoU and HD of UNets family when the optimizer is set as Adam. lr: learning rate. The best-performing results are highlighted using bold font.

  4. f

    numoverseg and nummiss of three BTS models. The best-performing results are...

    • plos.figshare.com
    xls
    Updated Jun 11, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Bang-Chuan Chen; Chung-Yi Shen; Jyh-Wen Chai; Ren-Hung Hwang; Wei-Chuan Chiang; Chi-Hsiang Chou; Wei-Min Liu (2025). numoverseg and nummiss of three BTS models. The best-performing results are highlighted using bold font. lr: learning rate. [Dataset]. http://doi.org/10.1371/journal.pone.0323692.t003
    Explore at:
    xlsAvailable download formats
    Dataset updated
    Jun 11, 2025
    Dataset provided by
    PLOS ONE
    Authors
    Bang-Chuan Chen; Chung-Yi Shen; Jyh-Wen Chai; Ren-Hung Hwang; Wei-Chuan Chiang; Chi-Hsiang Chou; Wei-Min Liu
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    numoverseg and nummiss of three BTS models. The best-performing results are highlighted using bold font. lr: learning rate.

  5. O

    Conceptual Captions

    • opendatalab.com
    • paperswithcode.com
    zip
    Updated Mar 24, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Google Research (2023). Conceptual Captions [Dataset]. https://opendatalab.com/OpenDataLab/Conceptual_Captions
    Explore at:
    zipAvailable download formats
    Dataset updated
    Mar 24, 2023
    Dataset provided by
    Google Research
    License

    https://github.com/google-research-datasets/conceptual-captions/blob/master/LICENSEhttps://github.com/google-research-datasets/conceptual-captions/blob/master/LICENSE

    Description

    Automatic image captioning is the task of producing a natural-language utterance (usually a sentence) that correctly reflects the visual content of an image. Up to this point, the resource most used for this task was the MS-COCO dataset, containing around 120,000 images and 5-way image-caption annotations (produced by paid annotators). Google's Conceptual Captions dataset has more than 3 million images, paired with natural-language captions. In contrast with the curated style of the MS-COCO images, Conceptual Captions images and their raw descriptions are harvested from the web, and therefore represent a wider variety of styles. The raw descriptions are harvested from the Alt-text HTML attribute associated with web images. The authors developed an automatic pipeline that extracts, filters, and transforms candidate image/caption pairs, with the goal of achieving a balance of cleanliness, informativeness, fluency, and learnability of the resulting captions.

  6. h

    laion-coco-aesthetic

    • huggingface.co
    Updated Feb 15, 2019
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Guangyi Liu (2019). laion-coco-aesthetic [Dataset]. https://huggingface.co/datasets/guangyil/laion-coco-aesthetic
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Feb 15, 2019
    Authors
    Guangyi Liu
    License

    Apache License, v2.0https://www.apache.org/licenses/LICENSE-2.0
    License information was derived automatically

    Description

    LAION COCO with aesthetic score and watermark score

    This dataset contains 10% samples of the LAION-COCO dataset filtered by some text rules (remove url, special tokens, etc.), and image rules (image size > 384x384, aesthetic score>4.75 and watermark probability<0.5). There are total 8,563,753 data instances in this dataset. And the corresponding aesthetic score and watermark score are also included. Noted: watermark score in the table means the probability of the existence of the… See the full description on the dataset page: https://huggingface.co/datasets/guangyil/laion-coco-aesthetic.

  7. C

    Carbon Block Report

    • datainsightsmarket.com
    doc, pdf, ppt
    Updated Jun 27, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Data Insights Market (2025). Carbon Block Report [Dataset]. https://www.datainsightsmarket.com/reports/carbon-block-1117085
    Explore at:
    ppt, pdf, docAvailable download formats
    Dataset updated
    Jun 27, 2025
    Dataset authored and provided by
    Data Insights Market
    License

    https://www.datainsightsmarket.com/privacy-policyhttps://www.datainsightsmarket.com/privacy-policy

    Time period covered
    2025 - 2033
    Area covered
    Global
    Variables measured
    Market Size
    Description

    The global carbon block filter market, valued at $1671.2 million in 2025, is projected to experience steady growth, driven by increasing demand for clean and safe drinking water. The compound annual growth rate (CAGR) of 3.4% from 2025 to 2033 indicates a consistent expansion, fueled by rising concerns about waterborne contaminants and the increasing adoption of point-of-use (POU) and point-of-entry (POE) water filtration systems in both residential and commercial settings. Key drivers include the growing awareness of water contamination issues, stricter government regulations regarding water quality, and the rising disposable incomes in developing economies leading to increased demand for improved water filtration solutions. Furthermore, advancements in carbon block filter technology, such as the development of more efficient and long-lasting filters, are contributing to market growth. The market is segmented by type (e.g., granular activated carbon, carbon block), application (residential, commercial), and region. Competitive landscape analysis reveals key players such as Marmon, Multipure, and others actively innovating and expanding their product portfolios to cater to evolving consumer preferences. Despite the positive growth outlook, certain challenges exist. Fluctuations in raw material prices, especially activated carbon, can impact production costs and profitability. Furthermore, the presence of numerous smaller players creates a competitive market environment. However, the increasing emphasis on sustainable water management practices and the growing adoption of water purification technologies are expected to offset these restraints and drive long-term market expansion. The market is geographically diverse, with North America and Europe currently holding significant market shares. However, developing economies in Asia-Pacific and other regions are expected to witness rapid growth due to increasing urbanization and rising awareness of water quality issues. The forecast period of 2025-2033 promises continued expansion, driven by technological advancements and a growing global focus on water safety and purity.

  8. f

    IoU and HD of UNets family when the optimizer is set as Adagrad. The...

    • plos.figshare.com
    xls
    Updated Jun 11, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Bang-Chuan Chen; Chung-Yi Shen; Jyh-Wen Chai; Ren-Hung Hwang; Wei-Chuan Chiang; Chi-Hsiang Chou; Wei-Min Liu (2025). IoU and HD of UNets family when the optimizer is set as Adagrad. The best-performing results are highlighted using bold font. [Dataset]. http://doi.org/10.1371/journal.pone.0323692.t001
    Explore at:
    xlsAvailable download formats
    Dataset updated
    Jun 11, 2025
    Dataset provided by
    PLOS ONE
    Authors
    Bang-Chuan Chen; Chung-Yi Shen; Jyh-Wen Chai; Ren-Hung Hwang; Wei-Chuan Chiang; Chi-Hsiang Chou; Wei-Min Liu
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    IoU and HD of UNets family when the optimizer is set as Adagrad. The best-performing results are highlighted using bold font.

  9. The Digital Forensics 2023 dataset - DF2023

    • zenodo.org
    • data.niaid.nih.gov
    Updated Apr 3, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    David Fischinger; David Fischinger; Martin Boyer; Martin Boyer (2025). The Digital Forensics 2023 dataset - DF2023 [Dataset]. http://doi.org/10.5281/zenodo.7326540
    Explore at:
    Dataset updated
    Apr 3, 2025
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    David Fischinger; David Fischinger; Martin Boyer; Martin Boyer
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    For a detailed description of the DF2023 dataset, please refer to:

     @inproceedings{Fischinger2023DFNet,
       title={DF2023: The Digital Forensics 2023 Dataset for Image Forgery Detection},
       author={David Fischinger and Martin Boyer},
       journal={The 25th Irish Machine Vision and Image Processing conference. (IMVIP)},
       year={2023}
     }
    

    DF2023 is a dataset for image forgery detection and localization. The training and validation datasets contain 1,000,000/5,000 manipulated images (and the ground truth masks).

    The DF2023 training dataset comprises:

    • 100K forged images produced by removal (inpainting) operations
    • 200K images produced by enhancement modifications
    • 300K copy-move manipulated images and
    • 400K spliced images

    === Naming convention ===

    The naming convention of DF2023 encodes information about the applied manipulations. Each image name has the following form:

    COCO_DF_0123456789_NNNNNNNN.{EXT} (e.g. COCO_DF_E000G40117_00200620.jpg)

    After the identifier of the image data source ("COCO") and the self-reference to the Digital Forensics ("DF") dataset, there are 10 digits as placeholders for the manipulation. Position 0 defines the manipulation types copy-move, splicing, removal, enhancement ([C,S,R,E]). The following digits 1-9 represent donor patch manipulations. For positions [1,2,7,8] (resample, flip, noise and brightness), a binary value indicates if this manipulation was applied to the donor image patch. Position 3 (rotate) indicates by the values 0-3 if the rotation was executed by 0, 90, 180 or 270 degrees. Position 4 defines if BoxBlur (B) or GaussianBlur (G) was used. Position 5 specifies the blurring radius. A value of 0 indicates that no blurring was executed. Position 6 indicates which of the Python-PIL contrast filters EDGE ENHANCE, EDGE ENHANCE MORE, SHARPEN, UnsharpMask or ImageEnhance (values 1-5) was applied. If none of them was applied, this value is set to 0. Finally, position 9 is set to the JPEG compression factor modulo 10, a value of 0 indicates that no JPEG compression was applied. The 8 characters NNNNNNNN in the image name template stand for a running number of the images.

    === Terms of Use / Licence ===

    The DF2023 dataset is based on the MS COCO dataset. Therefore, rules for using the images form MS COCO apply also for DF2023:

    Images

    The COCO Consortium does not own the copyright of the images. Use of the images must abide by the Flickr Terms of Use. The users of the images accept full responsibility for the use of the dataset, including but not limited to the use of any copies of copyrighted images that they may create from the dataset.

  10. Human V1 Dataset

    • universe.roboflow.com
    zip
    Updated Mar 24, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    aarongo.socialusername@gmail.com (2025). Human V1 Dataset [Dataset]. https://universe.roboflow.com/aarongo-socialusername-gmail-com/human-dataset-v1
    Explore at:
    zipAvailable download formats
    Dataset updated
    Mar 24, 2025
    Dataset provided by
    Gmailhttp://gmail.com/
    Authors
    aarongo.socialusername@gmail.com
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Variables measured
    Humans Bounding Boxes
    Description

    Here are a few use cases for this project:

    1. Human Presence Detection: This computer vision model can be incorporated into security systems and smart home devices to identify the presence of humans in an area, allowing for customized responses, room automation, and improved safety.

    2. Crowd Size Estimation: The "human dataset v1" can be used by event organizers or city planners to estimate the size of gatherings or crowds at public events, helping them better allocate resources and manage these events more efficiently.

    3. Surveillance and Security Enhancement: The model can be integrated into video surveillance systems to more accurately identify humans, helping to filter out false alarms caused by animals and other non-human entities.

    4. Collaborative Robotics: Robots equipped with this computer vision model can more easily identify and differentiate humans from their surroundings, allowing them to more effectively collaborate with people in shared spaces while ensuring human safety.

    5. Smart Advertising: The "human dataset v1" can be utilized by digital signage and advertising systems to detect and count the number of human viewers, enabling targeted advertising and measuring the effectiveness of marketing campaigns.

  11. R

    Udacity Self Driving Car Object Detection Dataset - fixed-large

    • public.roboflow.com
    zip
    Updated Mar 24, 2025
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Roboflow (2025). Udacity Self Driving Car Object Detection Dataset - fixed-large [Dataset]. https://public.roboflow.com/object-detection/self-driving-car/2
    Explore at:
    zipAvailable download formats
    Dataset updated
    Mar 24, 2025
    Dataset authored and provided by
    Roboflow
    License

    MIT Licensehttps://opensource.org/licenses/MIT
    License information was derived automatically

    Variables measured
    Bounding Boxes of obstacles
    Description

    Overview

    The original Udacity Self Driving Car Dataset is missing labels for thousands of pedestrians, bikers, cars, and traffic lights. This will result in poor model performance. When used in the context of self driving cars, this could even lead to human fatalities.

    We re-labeled the dataset to correct errors and omissions. We have provided convenient downloads in many formats including VOC XML, COCO JSON, Tensorflow Object Detection TFRecords, and more.

    Some examples of labels missing from the original dataset: https://i.imgur.com/A5J3qSt.jpg" alt="Examples of Missing Labels">

    Stats

    The dataset contains 97,942 labels across 11 classes and 15,000 images. There are 1,720 null examples (images with no labels).

    All images are 1920x1200 (download size ~3.1 GB). We have also provided a version downsampled to 512x512 (download size ~580 MB) that is suitable for most common machine learning models (including YOLO v3, Mask R-CNN, SSD, and mobilenet).

    Annotations have been hand-checked for accuracy by Roboflow.

    https://i.imgur.com/bOFkueI.pnghttps://" alt="Class Balance">

    Annotation Distribution: https://i.imgur.com/NwcrQKK.png" alt="Annotation Heatmap">

    Use Cases

    Udacity is building an open source self driving car! You might also try using this dataset to do person-detection and tracking.

    Using this Dataset

    Our updates to the dataset are released under the MIT License (the same license as the original annotations and images).

    Note: the dataset contains many duplicated bounding boxes for the same subject which we have not corrected. You will probably want to filter them by taking the IOU for classes that are 100% overlapping or it could affect your model performance (expecially in stoplight detection which seems to suffer from an especially severe case of duplicated bounding boxes).

    About Roboflow

    Roboflow makes managing, preprocessing, augmenting, and versioning datasets for computer vision seamless.

    Developers reduce 50% of their boilerplate code when using Roboflow's workflow, save training time, and increase model reproducibility. :fa-spacer:

    Roboflow Wordmark

  12. f

    Fusion results with UNet2+ (optimizer = Adagrad, learning rate = 0.0001)....

    • plos.figshare.com
    xls
    Updated Jun 11, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Bang-Chuan Chen; Chung-Yi Shen; Jyh-Wen Chai; Ren-Hung Hwang; Wei-Chuan Chiang; Chi-Hsiang Chou; Wei-Min Liu (2025). Fusion results with UNet2+ (optimizer = Adagrad, learning rate = 0.0001). The best-performing results are highlighted using bold font. [Dataset]. http://doi.org/10.1371/journal.pone.0323692.t004
    Explore at:
    xlsAvailable download formats
    Dataset updated
    Jun 11, 2025
    Dataset provided by
    PLOS ONE
    Authors
    Bang-Chuan Chen; Chung-Yi Shen; Jyh-Wen Chai; Ren-Hung Hwang; Wei-Chuan Chiang; Chi-Hsiang Chou; Wei-Min Liu
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Fusion results with UNet2+ (optimizer = Adagrad, learning rate = 0.0001). The best-performing results are highlighted using bold font.

  13. h

    Dogs-images-text-pair

    • huggingface.co
    Updated Jun 25, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Mikdad Mrhij (2025). Dogs-images-text-pair [Dataset]. https://huggingface.co/datasets/MikdadMrhij/Dogs-images-text-pair
    Explore at:
    Dataset updated
    Jun 25, 2025
    Authors
    Mikdad Mrhij
    License

    MIT Licensehttps://opensource.org/licenses/MIT
    License information was derived automatically

    Description

    This dataset is curated from subsets of public datasets such as MS COCO, LAION-ART, and SBU Captions. It specifically filters for samples featuring dog-related content. The goal of this dataset is to support text-to-image generation models focused on dog objects and related scenes. https://github.com/rom1504/img2dataset/blob/main/dataset_examples/SBUcaptions.md https://github.com/rom1504/img2dataset/blob/main/dataset_examples/mscoco.md… See the full description on the dataset page: https://huggingface.co/datasets/MikdadMrhij/Dogs-images-text-pair.

  14. SuperAnimal-Quadruped-80K

    • zenodo.org
    application/gzip
    Updated Nov 1, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Zenodo (2024). SuperAnimal-Quadruped-80K [Dataset]. http://doi.org/10.5281/zenodo.14016777
    Explore at:
    application/gzipAvailable download formats
    Dataset updated
    Nov 1, 2024
    Dataset provided by
    Zenodohttp://zenodo.org/
    Time period covered
    Jun 9, 2024
    Description

    Introduction

    This dataset supports Ye et al. 2024 Nature Communications. Please cite this dataset and paper if you use this resource. Please also see Ye et al. 2024 for the full DataSheet that accompanies this download, including the meta data for how to use this data is you want to compare model results on benchmark tasks. Below is just a summary. Also see the dataset licensing below.

    Training Data

    It consists of being trained together on the following datasets:

    • AwA-Pose Quadruped dataset, see full details at (1).
    • AnimalPose See full details at (2).
    • AcinoSet See full details at (3).
    • Horse-30 Horse-30 dataset, benchmark task is called Horse-10; See full details at (4).
    • StanfordDogs See full details at (5, 6).
    • AP-10K See full details at (7).
    • iRodent We utilized the iNaturalist API functions for scraping observations with the taxon ID of Suborder Myomorpha (8). The functions allowed us to filter the large amount of observations down to the ones with photos under the CC BY-NC creative license. The most common types of rodents from the collected observations are Muskrat (Ondatra zibethicus), Brown Rat (Rattus norvegicus), House Mouse (Mus musculus), Black Rat (Rattus rattus), Hispid Cotton Rat (Sigmodon hispidus), Meadow Vole (Microtus pennsylvanicus), Bank Vole (Clethrionomys glareolus), Deer Mouse (Peromyscus maniculatus), White-footed Mouse (Peromyscus leucopus), Striped Field Mouse (Apodemus agrarius). We then generated segmentation masks over target animals in the data by processing the media through an algorithm we designed that uses a Mask Region Based Convolutional Neural Networks(Mask R-CNN) (9) model with a ResNet-50-FPN backbone (10), pretrained on the COCO datasets (11). The processed 443 images were then manually labeled with both pose annotations and segmentation masks. iRodent data is banked at https://zenodo.org/record/8250392.
    • APT-36K See full details at (12).

    https://images.squarespace-cdn.com/content/v1/57f6d51c9f74566f55ecf271/1690988780004-AG00N6OU1R21MZ0AU9RE/modelcard-SAQ.png?format=1500w" target="_blank" rel="noopener">Here is an image with a keypoint guide.

    Ethical Considerations

    • No experimental data was collected for this model; all datasets used are cited above.

    Caveats and Recommendations

    • Please note that each dataest was labeled by separate labs & separate individuals, therefore while we map names to a unified pose vocabulary, there will be annotator bias in keypoint placement (See Ye et al. 2024 for our Supplementary Note on annotator bias). You will also note the dataset is highly diverse across species, but collectively has more representation of domesticated animals like dogs, cats, horses, and cattle. We recommend if performance of a model trained on this data is not as good as you need it to be, first try video adaptation (see Ye et al. 2024), or fine-tune the weights with your own labeling.

    License

    Modified MIT.

    Copyright 2023-present by Mackenzie Mathis, Shaokai Ye, and contributors.

    Permission is hereby granted to you (hereafter "LICENSEE") a fully-paid, non-exclusive,
    and non-transferable license for academic, non-commercial purposes only (hereafter “LICENSE”)
    to use the "DATASET" subject to the following conditions:

    The above copyright notice and this permission notice shall be included in all copies or substantial
    portions of the Software:

    This data or resulting software may not be used to harm any animal deliberately.

    LICENSEE acknowledges that the DATASET is a research tool.
    THE DATASET IS PROVIDED “AS IS”, WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING
    BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE DATASET OR THE USE OR OTHER DEALINGS IN THE DATASET.

    If this license is not appropriate for your application, please contact Prof. Mackenzie W. Mathis
    (mackenzie@post.harvard.edu) for a commercial use license.

    Please cite Ye et al if you use this DATASET in your work.

    References

    1. Prianka Banik, Lin Li, and Xishuang Dong. A novel dataset for keypoint detection of quadruped animals from images. ArXiv, abs/2108.13958, 2021
    2. Jinkun Cao, Hongyang Tang, Haoshu Fang, Xiaoyong Shen, Cewu Lu, and Yu-Wing Tai. Cross-domain adaptation for animal pose estimation. 2019 IEEE/CVF International Conference on Computer Vision (ICCV), pages 9497–9506, 2019.
    3. Daniel Joska, Liam Clark, Naoya Muramatsu, Ricardo Jericevich, Fred Nicolls, Alexander Mathis, Mackenzie W. Mathis, and Amir Patel. Acinoset: A 3d pose estimation dataset and baseline models for cheetahs in the wild. 2021 IEEE International Conference on Robotics and Automation (ICRA), pages 13901–13908, 2021.
    4. Alexander Mathis, Thomas Biasi, Steffen Schneider, Mert Yuksekgonul, Byron Rogers, Matthias Bethge, and Mackenzie W Mathis. Pretraining boosts out-of-domain robustness for pose estimation. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, pages 1859–1868, 2021.
    5. Aditya Khosla, Nityananda Jayadevaprakash, Bangpeng Yao, and Li Fei-Fei. Novel dataset for fine-grained image categorization. In First Workshop on Fine-Grained Visual Categorization, IEEE Conference on Computer Vision and Pattern Recognition, Colorado Springs, CO, June 2011.
    6. Benjamin Biggs, Thomas Roddick, Andrew Fitzgibbon, and Roberto Cipolla. Creatures great and smal: Recovering the shape and motion of animals from video. In Asian Conference on Computer Vision, pages 3–19. Springer, 2018.
    7. Hang Yu, Yufei Xu, Jing Zhang, Wei Zhao, Ziyu Guan, and Dacheng Tao. Ap-10k: A benchmark for animal pose estimation in the wild. In Thirty-fifth Conference on Neural Information Processing Systems Datasets and Benchmarks Track (Round 2), 2021.
    8. iNaturalist. OGBIF Occurrence Download. https://doi.org/10.15468/dl.p7nbxt. iNaturalist, July 2020
    9. Kaiming He, Georgia Gkioxari, Piotr Dollár, and Ross Girshick. Mask r-cnn. In Proceedings of the IEEE international conference on computer vision, pages 2961–2969, 2017.
    10. Tsung-Yi Lin, Piotr Dollár, Ross Girshick, Kaiming He, Bharath Hariharan, and Serge Belongie. Feature pyramid networks for object detection, 2016.
    11. Tsung-Yi Lin, Michael Maire, Serge J. Belongie, Lubomir D. Bourdev, Ross B. Girshick, James Hays, Pietro Perona, Deva Ramanan, Piotr Doll’ar, and C. Lawrence Zitnick. Microsoft COCO: common objects in context. CoRR, abs/1405.0312, 2014
    12. Yuxiang Yang, Junjie Yang, Yufei Xu, Jing Zhang, Long Lan, and Dacheng Tao. Apt-36k: A large-scale benchmark for animal pose estimation and tracking. Advances in Neural Information Processing Systems, 35:17301–17313, 2022

    Versioning Note:

    - V2 includes fixes to Stanford Dog data; it affected less than 1% of the data.

  15. u

    FMPD - Freshwater Microscopy Phytoplankton Dataset

    • portalinvestigacion.udc.gal
    • producciocientifica.uv.es
    Updated 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Rivas-Villar, David; Figueroa, Jorge; Carballeira, Rafael; Rouco, José; Novo, Jorge; Rivas-Villar, David; Figueroa, Jorge; Carballeira, Rafael; Rouco, José; Novo, Jorge (2024). FMPD - Freshwater Microscopy Phytoplankton Dataset [Dataset]. https://portalinvestigacion.udc.gal/documentos/668fc499b9e7c03b01be2414
    Explore at:
    Dataset updated
    2024
    Authors
    Rivas-Villar, David; Figueroa, Jorge; Carballeira, Rafael; Rouco, José; Novo, Jorge; Rivas-Villar, David; Figueroa, Jorge; Carballeira, Rafael; Rouco, José; Novo, Jorge
    Description

    This dataset, FMPD (Freshwater Microscopy Phytoplankton Dataset), is released for non-comercial academic or research purposes only, subject to attribution through citation of the following papers

    • Figueroa, J. Rouco, J. Novo, "Phytoplankton detection and recognition in freshwater digital microscopy images using deep learning object detectors", Heliyon, 2023.

    • D. Rivas-Villar, J. Rouco, M. G. Penedo, R. Carballeira, J. Novo, "Fully automatic detection and classification of phytoplankton specimens in digital microscopy images", Computer Methods and Programs in Biomedicine, 200, 105923, 2021

    Please also consider the citation of any of the other related papers from the dataset authors.

    Data:

    The FMPD dataset is a set of multi-specimen microscopy images of freshwater phytoplankton. These images have been captured with fixed settings, equal for each image, including illumination, focal point and magnification. The dataset contains 293 images from water sampled at lake of Doniños (Ferrol, Galicia, Spain) (UTM 555593 X, 4815672 Y; Datum ETRS89) on multiple visits throughout the year. This ensures seasonal representability.

    The phytoplankton sample was concentrated by filtering volume of 0.5 L through GF/F glass fiber filters and was then resuspended in 50 mL. Phytoplankton samples were preserved using 5% (v/v) glutaraldehyde, because it is efficient at preserving both cellular structures and pigment. The fixed sample was stored in the dark at constant temperature (10 oC) until analysis. The phytoplankton sample was homogenised for 2 min prior to microscopic examination. In addition, the sample was subjected to vacuum for one minute to break the vacuoles of some cyanobacterial taxa and prevent them from floating. Aliquots of the phytoplankton sample with a total volume of 1 mL were examined under light microscopy using a Nikon Eclipse E600 equipped with an E-Plan 10× objective (N.A. 0.25). Light microscopy images were taken with an AxioCam ICc5 Zeiss digital camera, maintaining the same illumination and focus throughout the image acquisition process and following regular transects until the entire surface of the sample was covered.

    The dataset contains 293 multi-specimen phytoplankton images. As mentioned, these images have fixed magnification, illumination and focal point. The produced images are saved in .tif format with a size of 2080x1540 pixels and are located in the dataset folder. The ground truth consists of bounding boxes that enclose the phytoplankton specimens, with an associated label identifying the species. Currently, this dataset has tags for:

    • Non-phytoplankton: particles, debris, zooplankton or any other object that could be mistaken as phytoplankton- Woronichinia naegeliana: Toxin-producing cyanobacteria- Anabaena Spiroides: Toxin-producing cyanobacteria- Dinobryon Sp.: Harmless but challenging as it can both appear solitary or in colonies- Other-phytoplankton: Other phytoplankton species.

    Annotations are provided in a .json file in the format typically used by the coco dataset, in the annotations.json file.

    Holdout train-test splits, as well as k-fold cross-validation splits, are provided in the splits folder, available in .json format. These splits correspond to those used in the previously mentioned papers to be cited, facilitating straightforward comparisons. Additionally, the annotations for each subset are included in separate files within the same folder for ease of use. It should be noted that the annotations.json contains all of these subsets of annotations.

  16. Kellogg's Coco Pops brand profile in the United Kingdom 2022

    • statista.com
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Statista, Kellogg's Coco Pops brand profile in the United Kingdom 2022 [Dataset]. https://www.statista.com/forecasts/1406322/kellogg-s-coco-pops-online-grocery-delivery-brand-profile-in-uk
    Explore at:
    Dataset authored and provided by
    Statistahttp://statista.com/
    Time period covered
    May 12, 2022 - Jan 17, 2023
    Area covered
    United Kingdom
    Description

    In 2022, brand awareness of Kellogg's Coco Pops was ** percent among survey respondents. When asked how many UK consumers ate Coco Pops in the past 12 months, just over a third of respondents stated that they had.

    Want more brand data?Explore more Statista Brand Profiles. If you want to compare brands, do deep-dives by survey items of your choice, filter by total online population or users of a certain brand, or drill down on your very own hand-tailored target groups, our Statista Brand Profiler has you covered.

  17. d

    Code, imagery, and annotations for training a deep learning model to detect...

    • catalog.data.gov
    • data.usgs.gov
    Updated Jul 6, 2024
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    U.S. Geological Survey (2024). Code, imagery, and annotations for training a deep learning model to detect wildlife in aerial imagery [Dataset]. https://catalog.data.gov/dataset/code-imagery-and-annotations-for-training-a-deep-learning-model-to-detect-wildlife-in-aeri
    Explore at:
    Dataset updated
    Jul 6, 2024
    Dataset provided by
    U.S. Geological Survey
    Description

    There are 3 child zip files included in this data release. 01_Codebase.zip contains a codebase for using deep learning to filter images based on the probability of any bird occurrence. It includes instructions and files necessary for training, validating, and testing a machine learning detection algorithm. 02_Imagery.zip contains imagery that were collected using a Partenavia P68 fixed-wing airplane using a PhaseOne iXU-R 180 forward motion compensating 80-megapixel digital frame camera with a 70 mm Rodenstock lens. The imagery were cropped into smaller patches of 720x720 pixels for training and 1440x1440 pixels for validation and test datasets. These data were collected for developing machine learning algorithms for the detection and classification of avian targets in aerial imagery. These data can be paired with annotation values to train and evaluate object detection and classification models. 03_Annotations.zip contains a collection of bounding boxes around avian targets in aerial imagery formatted as COCO JSON file. The data are nested under evaluation and test folders and contain both ground truth targets and predicted targets.These data were collected for two main functions. The ground truth avian targets were manually annotated and can be used to train avian detection algorithms using machine learning methods. The predicted targets can be used to evaluate model performance while referencing the larger work associated with these data.

  18. b

    Particulate Carbon Concentrations and Stable Carbon Isotopes in Marine...

    • datacart.bco-dmo.org
    • bco-dmo.org
    csv
    Updated May 10, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Sijia Dong; Frank J. Pavia; William M. Berelson; Jess F. Adkins (2024). Particulate Carbon Concentrations and Stable Carbon Isotopes in Marine Particles Captured by In Situ Mclane Pumps at Cocos Ridge Coco Ridge (Eastern Equatorial Pacific) during cruise SR2113 between November - December 2021 [Dataset]. http://doi.org/10.26008/1912/bco-dmo.925258.1
    Explore at:
    csv(10.18 KB)Available download formats
    Dataset updated
    May 10, 2024
    Dataset provided by
    Biological and Chemical Data Management Office
    Authors
    Sijia Dong; Frank J. Pavia; William M. Berelson; Jess F. Adkins
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Time period covered
    Dec 2, 2021 - Dec 14, 2021
    Area covered
    Variables measured
    Date, Depth, GFF_ID, LSF_TC, SSF_TC, LSF_PIC, SSF_PIC, Latitude, Longitude, LSF_TC_13C, and 5 more
    Measurement technique
    McLane Pump, Elemental Analyzer, Spectrometer
    Description

    This dataset includes particulate carbon concentrations and isotopes collected in December, 2021. Suspended particles are collected at 4 different stations near Cocos Ridge, at two different size fractions using Mclane pumps. The two size fractions are large size fraction (LSF) that is >51 um, and small size fraction (SSF) that is 0.5 -- 51 um. Concentrations and stable carbon isotopes of particulate inorganic carbon (PIC) and total carbon (TC) are measured and reported. PIC content are measured by acidifying a subsample of the Glass Fiber Filter (GFF) and measuring total CO2 released using a Picarro Cavity Ring-down Spectroscopy. TC content are analyzed by burning a subsample of the GFF on Elemental Analyzer (EA). Samples were collected during SR2113 onboard Sally Ride, under the project "new approaches to study calcium carbonate dissolution on the sea floor and its impact on paleo-proxy interpretations", as a water-column side determination of particle compositions and carbonate dissolution. This data reveals changes in concentrations and stable carbon isotopes with water depth, and has implications for multiple biogeochemical processes associated with both the inorganic and the organic carbon within marine particles in the water column.

  19. TreeAI Global Initiative - Advancing tree species identification from aerial...

    • zenodo.org
    Updated Mar 8, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Mirela Beloiu Schwenke; Mirela Beloiu Schwenke; Zhongyu Xia; Arthur Gessler; Arthur Gessler; Teja Kattenborn; Teja Kattenborn; Clemens Mosig; Clemens Mosig; Stefano Puliti; Stefano Puliti; Lars Waser; Lars Waser; Nataliia Rehush; Nataliia Rehush; Yan Cheng; Yan Cheng; Liang Xinliang; Verena C. Griess; Verena C. Griess; Martin Mokroš; Martin Mokroš; Zhongyu Xia; Liang Xinliang (2025). TreeAI Global Initiative - Advancing tree species identification from aerial images with deep learning [Dataset]. http://doi.org/10.5281/zenodo.14888706
    Explore at:
    Dataset updated
    Mar 8, 2025
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Mirela Beloiu Schwenke; Mirela Beloiu Schwenke; Zhongyu Xia; Arthur Gessler; Arthur Gessler; Teja Kattenborn; Teja Kattenborn; Clemens Mosig; Clemens Mosig; Stefano Puliti; Stefano Puliti; Lars Waser; Lars Waser; Nataliia Rehush; Nataliia Rehush; Yan Cheng; Yan Cheng; Liang Xinliang; Verena C. Griess; Verena C. Griess; Martin Mokroš; Martin Mokroš; Zhongyu Xia; Liang Xinliang
    License

    Attribution-NonCommercial-NoDerivs 4.0 (CC BY-NC-ND 4.0)https://creativecommons.org/licenses/by-nc-nd/4.0/
    License information was derived automatically

    Description

    TreeAI - Advancing Tree Species Identification from Aerial Images with Deep Learning

    Data Structure for the TreeAI Database Used in the TreeAI4Species Competition

    The data are in the COCO format, each folder contains training and validation subfolders with images and labels with the tree species ID.
    Training: Images (.png) and Labels (.txt)
    Validation: Images (.png) and Labels (.txt)
    Images: RGB bands, 8-bit, chip size 640 x 640 pixels = 32 x 32 m, 5 cm pixel spatial resolution.
    Labels: labels are prepared for object detection tasks, the number of classes varies per dataset, e.g. dataset 12_RGB_all_L has 53 classes, and the Latin name of the species is given for each class ID in the file named classDatasetName.xlsx.
    Species class: classDatasetName.xlsx contains 3 columns Species_ID, Labels (number of labels), and Species_Class (Latin name of the species).
    Masked images: The data set with partial labels was masked, i.e. a buffer of 30 pixels was created around a label, and the image was masked based on these buffers, e.g. 34_RGB_all_L_PascalVoc_640Mask.
    Additional filters to clean up the data:
    Labels at the edge: only images with labels at the edge were removed.
    Valid labels: images with labels that were completely within an image have been retained.
    Table 1. Description of the datasets included in the TreeAI database.

    a) Fully labeled images (i.e. the image has all the trees delineated and each polygon has species information)

    b) Partially labeled images (i.e. the image has only some trees delineated, and each polygon has species information)

    No.Dataset nameTraining imagesValidation imagesFully labeledPartially labeled
    112_RGB5cm_FullyLabeled1066304x
    2ObjectDetection_TreeSpecies42284x
    334_RGB_all_L_PascalVoc_640Mask951272 x
    434_RGB_PartiallyLabeled640917262 x
    Steps to access the dataset and participate in the TreeAI4Species competition:
    • Register: Access to the data will be granted upon registering for the competition, see the registration form: https://form.ethz.ch/research/tree-ai-global-database/treeai-competition.html
    • Request the dataset: Download the competition record after registration by requesting it. Enter your full name, purpose e.g. accept the TreeAI4Species data license, affiliation, and the country of affiliation in the request. This allows us to check whether you are already registered.
    • Test dataset: Only the participants registered for the competition will receive the test dataset.
    • Submit your DL models for evaluation by June 2025.
    • Award: The best models win a prize.
    • Publication: All participants in the competition who submit the required files for evaluation will be included in the subsequent publication.

    License

    == CC BY-NC-ND (Attribution-NonCommercial-NoDerivatives) ==
    Dear user,
    DATA ANALYSIS AND PUBLICATION
    The TreeAI database is released under a variant of the CC BY-NC-ND license. This database is confidential and can be used only for the TreeAI4Species data science competition. It is not permitted to pass on the data or the characteristics directly derived from it to third parties. Written consent from the data supplier is required for use for any other purpose.
    LIABILITY
    The data are based on the current state of existing scientific knowledge. However, there is no liability for the completeness. This is the first version of the database, and we plan to improve the tree annotations and include new tree species. Therefore, another version will be released in the future.
    The data can only be used for the purpose described by the user when requesting the data.
    ------------------------------------------------------
    ETH Zürich
    Dr. Mirela Beloiu Schwenke
    Institute of Terrestrial Ecosystems
    Department of Environmental Systems Science, CHN K75
    Universitätstrasse 16, 8092 Zürich, Schweiz
    mirela.beloiu@usys.ethz.ch

  20. f

    Data from: Wastewater treatment using adsorption process in column for...

    • scielo.figshare.com
    jpeg
    Updated Jun 1, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Flávia Rhuana Pereira Sales; Reynaldo Borges Galvão Serra; Gesivaldo Jesus Alves de Figueirêdo; Paulo Henrique Almeida da Hora; Antonio Cícero de Sousa (2023). Wastewater treatment using adsorption process in column for agricultural purposes [Dataset]. http://doi.org/10.6084/m9.figshare.7676681.v1
    Explore at:
    jpegAvailable download formats
    Dataset updated
    Jun 1, 2023
    Dataset provided by
    SciELO journals
    Authors
    Flávia Rhuana Pereira Sales; Reynaldo Borges Galvão Serra; Gesivaldo Jesus Alves de Figueirêdo; Paulo Henrique Almeida da Hora; Antonio Cícero de Sousa
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Abstract Water is essential for life, important for the ecosystem and it is in great demand due to its scarcity. This study explored the reuse of the wastewater of the Water Treatment Plant in Gramame for agricultural purposes. A qualitative and quantitative investigation of the effluent was carried out through the characterization of its physical and chemical parameters, comparing the results to what is allowed by the current legislation, CONAMA Resolution 357/2005. After this process, the activated charcoal of coco-da-baia mesocarp, adsorbent material, was prepared and tested in a filter system in a column with a continuous flow and ascendant entrance, in which the kinetic effect was evaluated. This technique was evaluated by correlating the reduced values in the adsorption material, respecting the initial effluent concentration, obtaining a reduction of 50% in the hardness, 87.5% in chloride and 66.6% in acidity. These results verified the adequacy of the technique in potential hydrogenation (pH) and abrupt reduction of color and turbidity. This treatment is suggested to qualify the effluent for use in agricultural, safe for humans and the environment. The adsorbent substrate efficiency was verified by correlating it with the Thomas isothermal model.

  21. Not seeing a result you expected?
    Learn how you can add new datasets to our index.

Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
Schowing, Thibault (2024). COCO dataset and neural network weights for micro-FTIR particle detection on filters. [Dataset]. https://data.niaid.nih.gov/resources?id=zenodo_10839526

COCO dataset and neural network weights for micro-FTIR particle detection on filters.

Explore at:
Dataset updated
Aug 13, 2024
Dataset authored and provided by
Schowing, Thibault
License

Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically

Description

The IMPTOX project has received funding from the EU's H2020 framework programme for research and innovation under grant agreement n. 965173. Imptox is part of the European MNP cluster on human health.

More information about the project here.

Description: This repository includes the trained weights and a custom COCO-formatted dataset used for developing and testing a Faster R-CNN R_50_FPN_3x object detector, specifically designed to identify particles in micro-FTIR filter images.

Contents:

Weights File (neuralNetWeights_V3.pth):

Format: .pth

Description: This file contains the trained weights for a Faster R-CNN model with a ResNet-50 backbone and a Feature Pyramid Network (FPN), trained for 3x schedule. These weights are specifically tuned for detecting particles in micro-FTIR filter images.

Custom COCO Dataset (uFTIR_curated_square.v5-uftir_curated_square_2024-03-14.coco-segmentation.zip):

Format: .zip

Description: This zip archive contains a custom COCO-formatted dataset, including JPEG images and their corresponding annotation file. The dataset consists of images of micro-FTIR filters with annotated particles.

Contents:

Images: JPEG format images of micro-FTIR filters.

Annotations: A JSON file in COCO format providing detailed annotations of the particles in the images.

Management: The dataset can be managed and manipulated using the Pycocotools library, facilitating easy integration with existing COCO tools and workflows.

Applications: The provided weights and dataset are intended for researchers and practitioners in the field of microscopy and particle detection. The dataset and model can be used for further training, validation, and fine-tuning of object detection models in similar domains.

Usage Notes:

The neuralNetWeights_V3.pth file should be loaded into a PyTorch model compatible with the Faster R-CNN architecture, such as Detectron2.

The contents of uFTIR_curated_square.v5-uftir_curated_square_2024-03-14.coco-segmentation.zip should be extracted and can be used with any COCO-compatible object detection framework for training and evaluation purposes.

Code can be found on the related Github repository.

Search
Clear search
Close search
Google apps
Main menu