5 datasets found
  1. d

    Data from: X-ray CT data with semantic annotations for the paper "A workflow...

    • catalog.data.gov
    • agdatacommons.nal.usda.gov
    Updated Jun 5, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Agricultural Research Service (2025). X-ray CT data with semantic annotations for the paper "A workflow for segmenting soil and plant X-ray CT images with deep learning in Google’s Colaboratory" [Dataset]. https://catalog.data.gov/dataset/x-ray-ct-data-with-semantic-annotations-for-the-paper-a-workflow-for-segmenting-soil-and-p-d195a
    Explore at:
    Dataset updated
    Jun 5, 2025
    Dataset provided by
    Agricultural Research Service
    Description

    Leaves from genetically unique Juglans regia plants were scanned using X-ray micro-computed tomography (microCT) on the X-ray μCT beamline (8.3.2) at the Advanced Light Source (ALS) in Lawrence Berkeley National Laboratory (LBNL), Berkeley, CA USA). Soil samples were collected in Fall of 2017 from the riparian oak forest located at the Russell Ranch Sustainable Agricultural Institute at the University of California Davis. The soil was sieved through a 2 mm mesh and was air dried before imaging. A single soil aggregate was scanned at 23 keV using the 10x objective lens with a pixel resolution of 650 nanometers on beamline 8.3.2 at the ALS. Additionally, a drought stressed almond flower bud (Prunus dulcis) from a plant housed at the University of California, Davis, was scanned using a 4x lens with a pixel resolution of 1.72 µm on beamline 8.3.2 at the ALS Raw tomographic image data was reconstructed using TomoPy. Reconstructions were converted to 8-bit tif or png format using ImageJ or the PIL package in Python before further processing. Images were annotated using Intel’s Computer Vision Annotation Tool (CVAT) and ImageJ. Both CVAT and ImageJ are free to use and open source. Leaf images were annotated in following Théroux-Rancourt et al. (2020). Specifically, Hand labeling was done directly in ImageJ by drawing around each tissue; with 5 images annotated per leaf. Care was taken to cover a range of anatomical variation to help improve the generalizability of the models to other leaves. All slices were labeled by Dr. Mina Momayyezi and Fiona Duong.To annotate the flower bud and soil aggregate, images were imported into CVAT. The exterior border of the bud (i.e. bud scales) and flower were annotated in CVAT and exported as masks. Similarly, the exterior of the soil aggregate and particulate organic matter identified by eye were annotated in CVAT and exported as masks. To annotate air spaces in both the bud and soil aggregate, images were imported into ImageJ. A gaussian blur was applied to the image to decrease noise and then the air space was segmented using thresholding. After applying the threshold, the selected air space region was converted to a binary image with white representing the air space and black representing everything else. This binary image was overlaid upon the original image and the air space within the flower bud and aggregate was selected using the “free hand” tool. Air space outside of the region of interest for both image sets was eliminated. The quality of the air space annotation was then visually inspected for accuracy against the underlying original image; incomplete annotations were corrected using the brush or pencil tool to paint missing air space white and incorrectly identified air space black. Once the annotation was satisfactorily corrected, the binary image of the air space was saved. Finally, the annotations of the bud and flower or aggregate and organic matter were opened in ImageJ and the associated air space mask was overlaid on top of them forming a three-layer mask suitable for training the fully convolutional network. All labeling of the soil aggregate and soil aggregate images was done by Dr. Devin Rippner. These images and annotations are for training deep learning models to identify different constituents in leaves, almond buds, and soil aggregates Limitations: For the walnut leaves, some tissues (stomata, etc.) are not labeled and only represent a small portion of a full leaf. Similarly, both the almond bud and the aggregate represent just one single sample of each. The bud tissues are only divided up into buds scales, flower, and air space. Many other tissues remain unlabeled. For the soil aggregate annotated labels are done by eye with no actual chemical information. Therefore particulate organic matter identification may be incorrect. Resources in this dataset:Resource Title: Annotated X-ray CT images and masks of a Forest Soil Aggregate. File Name: forest_soil_images_masks_for_testing_training.zipResource Description: This aggregate was collected from the riparian oak forest at the Russell Ranch Sustainable Agricultural Facility. The aggreagate was scanned using X-ray micro-computed tomography (microCT) on the X-ray μCT beamline (8.3.2) at the Advanced Light Source (ALS) in Lawrence Berkeley National Laboratory (LBNL), Berkeley, CA USA) using the 10x objective lens with a pixel resolution of 650 nanometers. For masks, the background has a value of 0,0,0; pores spaces have a value of 250,250, 250; mineral solids have a value= 128,0,0; and particulate organic matter has a value of = 000,128,000. These files were used for training a model to segment the forest soil aggregate and for testing the accuracy, precision, recall, and f1 score of the model.Resource Title: Annotated X-ray CT images and masks of an Almond bud (P. Dulcis). File Name: Almond_bud_tube_D_P6_training_testing_images_and_masks.zipResource Description: Drought stressed almond flower bud (Prunis dulcis) from a plant housed at the University of California, Davis, was scanned by X-ray micro-computed tomography (microCT) on the X-ray μCT beamline (8.3.2) at the Advanced Light Source (ALS) in Lawrence Berkeley National Laboratory (LBNL), Berkeley, CA USA) using the 4x lens with a pixel resolution of 1.72 µm using. For masks, the background has a value of 0,0,0; air spaces have a value of 255,255, 255; bud scales have a value= 128,0,0; and flower tissues have a value of = 000,128,000. These files were used for training a model to segment the almond bud and for testing the accuracy, precision, recall, and f1 score of the model.Resource Software Recommended: Fiji (ImageJ),url: https://imagej.net/software/fiji/downloads Resource Title: Annotated X-ray CT images and masks of Walnut leaves (J. Regia) . File Name: 6_leaf_training_testing_images_and_masks_for_paper.zipResource Description: Stems were collected from genetically unique J. regia accessions at the 117 USDA-ARS-NCGR in Wolfskill Experimental Orchard, Winters, California USA to use as scion, and were grafted by Sierra Gold Nursery onto a commonly used commercial rootstock, RX1 (J. microcarpa × J. regia). We used a common rootstock to eliminate any own-root effects and to simulate conditions for a commercial walnut orchard setting, where rootstocks are commonly used. The grafted saplings were repotted and transferred to the Armstrong lathe house facility at the University of California, Davis in June 2019, and kept under natural light and temperature. Leaves from each accession and treatment were scanned using X-ray micro-computed tomography (microCT) on the X-ray μCT beamline (8.3.2) at the Advanced Light Source (ALS) in Lawrence Berkeley National Laboratory (LBNL), Berkeley, CA USA) using the 10x objective lens with a pixel resolution of 650 nanometers. For masks, the background has a value of 170,170,170; Epidermis value= 85,85,85; Mesophyll value= 0,0,0; Bundle Sheath Extension value= 152,152,152; Vein value= 220,220,220; Air value = 255,255,255.Resource Software Recommended: Fiji (ImageJ),url: https://imagej.net/software/fiji/downloads

  2. Data Annotation Tools Market Research Report 2033

    • growthmarketreports.com
    csv, pdf, pptx
    Updated Jun 30, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Growth Market Reports (2025). Data Annotation Tools Market Research Report 2033 [Dataset]. https://growthmarketreports.com/report/data-annotation-tools-market-global-geographical-industry-analysis
    Explore at:
    csv, pdf, pptxAvailable download formats
    Dataset updated
    Jun 30, 2025
    Dataset authored and provided by
    Growth Market Reports
    Time period covered
    2024 - 2032
    Area covered
    Global
    Description

    Data Annotation Tools Market Outlook



    According to our latest research, the global Data Annotation Tools market size reached USD 2.1 billion in 2024. The market is set to expand at a robust CAGR of 26.7% from 2025 to 2033, projecting a remarkable value of USD 18.1 billion by 2033. The primary growth driver for this market is the escalating adoption of artificial intelligence (AI) and machine learning (ML) across various industries, which necessitates high-quality labeled data for model training and validation.




    One of the most significant growth factors propelling the data annotation tools market is the exponential rise in AI-powered applications across sectors such as healthcare, automotive, retail, and BFSI. As organizations increasingly integrate AI and ML into their core operations, the demand for accurately annotated data has surged. Data annotation tools play a crucial role in transforming raw, unstructured data into structured, labeled datasets that can be efficiently used to train sophisticated algorithms. The proliferation of deep learning and natural language processing technologies further amplifies the need for comprehensive data labeling solutions. This trend is particularly evident in industries like healthcare, where annotated medical images are vital for diagnostic algorithms, and in automotive, where labeled sensor data supports the evolution of autonomous vehicles.




    Another prominent driver is the shift toward automation and digital transformation, which has accelerated the deployment of data annotation tools. Enterprises are increasingly adopting automated and semi-automated annotation platforms to enhance productivity, reduce manual errors, and streamline the data preparation process. The emergence of cloud-based annotation solutions has also contributed to market growth by enabling remote collaboration, scalability, and integration with advanced AI development pipelines. Furthermore, the growing complexity and variety of data types, including text, audio, image, and video, necessitate versatile annotation tools capable of handling multimodal datasets, thus broadening the market's scope and applications.




    The market is also benefiting from a surge in government and private investments aimed at fostering AI innovation and digital infrastructure. Several governments across North America, Europe, and Asia Pacific have launched initiatives and funding programs to support AI research and development, including the creation of high-quality, annotated datasets. These efforts are complemented by strategic partnerships between technology vendors, research institutions, and enterprises, which are collectively advancing the capabilities of data annotation tools. As regulatory standards for data privacy and security become more stringent, there is an increasing emphasis on secure, compliant annotation solutions, further driving innovation and market demand.




    From a regional perspective, North America currently dominates the data annotation tools market, driven by the presence of major technology companies, well-established AI research ecosystems, and significant investments in digital transformation. However, Asia Pacific is emerging as the fastest-growing region, fueled by rapid industrialization, expanding IT infrastructure, and a burgeoning startup ecosystem focused on AI and data science. Europe also holds a substantial market share, supported by robust regulatory frameworks and active participation in AI research. Latin America and the Middle East & Africa are gradually catching up, with increasing adoption in sectors such as retail, automotive, and government. The global landscape is characterized by dynamic regional trends, with each market contributing uniquely to the overall growth trajectory.





    Component Analysis



    The data annotation tools market is segmented by component into software and services, each playing a pivotal role in the market's overall ecosystem. Software solutions form the backbone of the market, providing the technical infrastructure for auto

  3. AI-Powered Medical Imaging Annotation Market Research Report 2033

    • growthmarketreports.com
    csv, pdf, pptx
    Updated Jun 28, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Growth Market Reports (2025). AI-Powered Medical Imaging Annotation Market Research Report 2033 [Dataset]. https://growthmarketreports.com/report/ai-powered-medical-imaging-annotation-market
    Explore at:
    csv, pptx, pdfAvailable download formats
    Dataset updated
    Jun 28, 2025
    Dataset authored and provided by
    Growth Market Reports
    Time period covered
    2024 - 2032
    Area covered
    Global
    Description

    AI-Powered Medical Imaging Annotation Market Outlook




    According to our latest research, the global AI-powered medical imaging annotation market size reached USD 1.24 billion in 2024, demonstrating robust traction across healthcare and life sciences sectors. The market is projected to expand at a compound annual growth rate (CAGR) of 23.7% from 2025 to 2033, reaching an estimated USD 9.31 billion by 2033. This significant growth is primarily driven by the increasing adoption of artificial intelligence (AI) in medical diagnostics, the rising prevalence of chronic diseases necessitating advanced imaging techniques, and the urgent need for high-quality annotated datasets to train sophisticated AI algorithms for clinical applications.




    A pivotal growth factor for the AI-powered medical imaging annotation market is the escalating demand for precision medicine and personalized healthcare. As healthcare providers and researchers strive for tailored treatment plans, the need for accurate and detailed medical image annotation becomes paramount. AI-driven annotation platforms enable rapid, consistent, and scalable labeling of complex imaging data such as CT, MRI, and X-ray scans, facilitating the development of advanced diagnostic tools. Furthermore, the integration of AI in annotation workflows reduces human error, improves annotation speed, and enhances the quality of datasets, all of which are essential for training reliable machine learning models used in disease detection, prognosis, and treatment planning.




    Another significant driver is the exponential growth in medical imaging data generated globally. With the proliferation of advanced imaging modalities and the increasing use of digital health records, healthcare systems are inundated with vast quantities of imaging data. Manual annotation of such data is time-consuming, labor-intensive, and prone to inconsistencies. AI-powered annotation solutions address these challenges by automating the labeling process, ensuring uniformity, and enabling real-time collaboration among radiologists, data scientists, and clinicians. This not only accelerates the deployment of AI-powered diagnostic tools but also supports large-scale clinical research initiatives aimed at uncovering novel biomarkers and improving patient outcomes.




    The growing emphasis on regulatory compliance and data standardization also fuels market expansion. Regulatory bodies such as the FDA and EMA increasingly mandate the use of annotated datasets for the validation and approval of AI-driven diagnostic devices. As a result, healthcare organizations and medical device manufacturers are investing heavily in AI-powered annotation platforms that comply with stringent data privacy and security standards. Moreover, the emergence of cloud-based annotation solutions enhances accessibility and scalability, allowing stakeholders from diverse geographies to collaborate seamlessly on large annotation projects, thereby accelerating innovation and commercialization in the medical imaging domain.




    Regionally, North America dominates the AI-powered medical imaging annotation market due to its advanced healthcare infrastructure, high adoption of AI technologies, and substantial investments in medical research. Europe follows closely, benefiting from strong regulatory support and a well-established healthcare ecosystem. The Asia Pacific region is poised for the fastest growth, driven by increasing healthcare expenditure, rapid digitalization, and government initiatives promoting AI adoption in healthcare. Latin America and the Middle East & Africa are emerging markets, gradually embracing AI-powered solutions to address gaps in diagnostic capabilities and improve healthcare access. This regional diversification underscores the global relevance and transformative potential of AI-powered medical imaging annotation.





    Component Analysis




    The component segment of the AI-powered medical imaging annotation market is bifurcated into software and services, each pla

  4. Data from: Automatic patient-level recognition of four Plasmodium species on...

    • zenodo.org
    bin, zip
    Updated Sep 26, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Emilie Guemas; Baptiste Routier; Théo Ghelfenstein-Ferreira; Camille Cordier; Sophie Hartuis; Bénédicte Marion; Sébastien Bertout; Damien Costa; Grégoire Pasquier; Emilie Guemas; Baptiste Routier; Théo Ghelfenstein-Ferreira; Camille Cordier; Sophie Hartuis; Bénédicte Marion; Sébastien Bertout; Damien Costa; Grégoire Pasquier (2023). Data from: Automatic patient-level recognition of four Plasmodium species on thin blood smear by a Real Time Detector Transformer (RT-DETR) object detection algorithm: a proof-of-concept and evaluation [Dataset]. http://doi.org/10.5281/zenodo.8358829
    Explore at:
    bin, zipAvailable download formats
    Dataset updated
    Sep 26, 2023
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Emilie Guemas; Baptiste Routier; Théo Ghelfenstein-Ferreira; Camille Cordier; Sophie Hartuis; Bénédicte Marion; Sébastien Bertout; Damien Costa; Grégoire Pasquier; Emilie Guemas; Baptiste Routier; Théo Ghelfenstein-Ferreira; Camille Cordier; Sophie Hartuis; Bénédicte Marion; Sébastien Bertout; Damien Costa; Grégoire Pasquier
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Automatic patient-level recognition of four Plasmodium species on thin blood smear by a Real Time Dectector Transformer (RT-DETR) object detection algorithm: a proof-of-concept and evaluation

    Emilie Guemas, Baptiste Routier, Théo Ghelfenstein-Ferreira, Camille Cordier, Sophie Hartuis, Bénédicte Marion, Sébastien Bertout, Emmanuelle Varlet-Marie, Damien Costa, Grégoire Pasquier

    Abstract:

    Malaria remains a global health problem with 247 million cases and 619,000 deaths in 2021. Diagnostic of Plasmodium species is important for administering the appropriate treatment. The gold-standard diagnosis from accurate species identification remains the thin blood smear. Nevertheless, this method is time-consuming and requires highly skilled and trained microscopists. To overcome these issues, new diagnostic tools based on deep learning are emerging. This study aimed to evaluate the performances of a RT-DETR (Real-Time Detection Transformer)object detection algorithm to discriminate Plasmodium species on thin blood smears images. The algorithm was trained and validated on a dataset consisting in 24,720 images from 475 thin blood smears corresponding to 2,002,597 labels. Performances were calculated with a test dataset of 4,508 images from 170 smears corresponding to 358,825labels coming from six French university hospital. At the patient level, the RT-DETR algorithm exhibited an overall accuracy of 79.4% (135/170) with a recall of 74% (40/54) and 81.9% (95/116) for negative and positive smears, respectively. Among Plasmodium positive smears, the global sensitivity was 82.7% (91/110) with a sensitivity of 90% (38/42), 81.8% (18/22) and 76.1% (35/46) for P. falciparum, P. malariae and P. ovale/vivax, respectively. The YOLOv5 model achieved a World Health Organization (WHO) competence level 2 for species identification. Besides, the RT-DETR algorithm may be run in real-time on low-cost devices such as a smartphone and could be suitable for deployment in low-resource setting areas where microscopy experts are lacking.

    Data collection:

    The training and validation dataset included 24,720 pictures taken from 475 manually May Grunwald-Giemsa (MGG)-stained thin blood smears from the Montpellier University Hospital collection and for a smaller part from the Toulouse University Hospital collection. In Montpellier, the pictures were taken with a Flexcam C1 microscope camera (Leica) attached to a Leica DM 2000 microscope and Leica DF450C microscope camera adapted with a Leica DM2500 microscope at X1000 magnification. Labelling of pictures was performed manually, and then automatically with manual correction with a Computer Visual Annotation Tools (CVAT) free software. Nine categories of labels were used: white blood cells (n=3,338), red blood cells (n=1,887,781), platelets (n=48,520), Trypanosoma brucei (n=2,773), and red blood cells infected by P. falciparum (n=43,545), P. ovale (n=4,651), P. vivax (n=4,115), P. malariae (n=2,849) and Babesia divergens (n=5,142).

    The test dataset included 4,508 pictures taken from 170 thin blood smears from the same number of patients from the Parasitology laboratories of University Hospitals of Montpellier, Toulouse, Rouen, Lille, Nantes and Saint-Louis in Paris (Table 1). Among these 170 patients, 54 were not infected, including two patients with Howell-Jolly bodies, and 116 were infected with hematozoa. For each patient, between 20 and 30 photos were taken from one thin blood smear with at least one hematozoan parasite per picture for infected patients.

    Accurate species diagnostic was made by a senior parasitologist, and for recent smears, it was confirmed by specific PCR, either performed locally (Toulouse) or at the Malaria French National Reference Center (Montpellier, Saint Louis, Rouen, Lille, Nantes).

  5. Data from: CVB: A Video Dataset of Cattle Visual Behaviors

    • researchdata.edu.au
    • data.csiro.au
    datadownload
    Updated Jun 13, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Flavio Pereira Alvarenga; Aaron Ingham; Lars Petersson; Brano Kusy; Vivien Rolland; Brendan Do; Neil Bagnall; Jody McNally; Greg Bishop-Hurley; Reza Arablouei; Ali Zia; Renuka Sharma (2023). CVB: A Video Dataset of Cattle Visual Behaviors [Dataset]. http://doi.org/10.25919/3G3T-P068
    Explore at:
    datadownloadAvailable download formats
    Dataset updated
    Jun 13, 2023
    Dataset provided by
    CSIROhttp://www.csiro.au/
    Authors
    Flavio Pereira Alvarenga; Aaron Ingham; Lars Petersson; Brano Kusy; Vivien Rolland; Brendan Do; Neil Bagnall; Jody McNally; Greg Bishop-Hurley; Reza Arablouei; Ali Zia; Renuka Sharma
    License

    Attribution-NonCommercial-ShareAlike 4.0 (CC BY-NC-SA 4.0)https://creativecommons.org/licenses/by-nc-sa/4.0/
    License information was derived automatically

    Time period covered
    Aug 1, 2022 - Apr 28, 2023
    Area covered
    Description

    Existing image/video datasets for cattle behavior recognition are mostly small, lack well-defined labels, or are collected in unrealistic controlled environments. This limits the utility of machine learning (ML) models learned from them. Therefore, we introduce a new dataset, called Cattle Visual Behaviors (CVB), that consists of 502 video clips, each fifteen seconds long, captured in natural lighting conditions, and annotated with eleven visually perceptible behaviors of grazing cattle. By creating and sharing CVB, our aim is to develop improved models capable of recognizing all important behaviors accurately and to assist other researchers and practitioners in developing and evaluating new ML models for cattle behavior classification using video data. The dataset is presented in the form of following three sub-directories. 1. raw_frames: contains 450 frames in each sub folder, representing 15 sec video, taking at a frames rate of 30 FPS, 2. annotations: contains the json files corresponding to the raw_frames folder. We have one json file for one video, containing the bounding box annotations for each cattle and their associated behaviors, and 3. CVB_in_AVA_format: contains the CVB data in the standard AVA dataset format which we have used to apply SlowFast model. Lineage: We use the Computer Vision Annotation Tool (CVAT) to collect our annotations. To make the procedure more efficient, we perform an initial detection and tracking of cattle in the videos using appropriate pre-trained models. The results are corrected by domain experts along with cattle behavior labeling in CVAT. The pre-hoc detection and tracking step significantly reduces the manual annotation time and effort.

  6. Not seeing a result you expected?
    Learn how you can add new datasets to our index.

Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
Agricultural Research Service (2025). X-ray CT data with semantic annotations for the paper "A workflow for segmenting soil and plant X-ray CT images with deep learning in Google’s Colaboratory" [Dataset]. https://catalog.data.gov/dataset/x-ray-ct-data-with-semantic-annotations-for-the-paper-a-workflow-for-segmenting-soil-and-p-d195a

Data from: X-ray CT data with semantic annotations for the paper "A workflow for segmenting soil and plant X-ray CT images with deep learning in Google’s Colaboratory"

Related Article
Explore at:
Dataset updated
Jun 5, 2025
Dataset provided by
Agricultural Research Service
Description

Leaves from genetically unique Juglans regia plants were scanned using X-ray micro-computed tomography (microCT) on the X-ray μCT beamline (8.3.2) at the Advanced Light Source (ALS) in Lawrence Berkeley National Laboratory (LBNL), Berkeley, CA USA). Soil samples were collected in Fall of 2017 from the riparian oak forest located at the Russell Ranch Sustainable Agricultural Institute at the University of California Davis. The soil was sieved through a 2 mm mesh and was air dried before imaging. A single soil aggregate was scanned at 23 keV using the 10x objective lens with a pixel resolution of 650 nanometers on beamline 8.3.2 at the ALS. Additionally, a drought stressed almond flower bud (Prunus dulcis) from a plant housed at the University of California, Davis, was scanned using a 4x lens with a pixel resolution of 1.72 µm on beamline 8.3.2 at the ALS Raw tomographic image data was reconstructed using TomoPy. Reconstructions were converted to 8-bit tif or png format using ImageJ or the PIL package in Python before further processing. Images were annotated using Intel’s Computer Vision Annotation Tool (CVAT) and ImageJ. Both CVAT and ImageJ are free to use and open source. Leaf images were annotated in following Théroux-Rancourt et al. (2020). Specifically, Hand labeling was done directly in ImageJ by drawing around each tissue; with 5 images annotated per leaf. Care was taken to cover a range of anatomical variation to help improve the generalizability of the models to other leaves. All slices were labeled by Dr. Mina Momayyezi and Fiona Duong.To annotate the flower bud and soil aggregate, images were imported into CVAT. The exterior border of the bud (i.e. bud scales) and flower were annotated in CVAT and exported as masks. Similarly, the exterior of the soil aggregate and particulate organic matter identified by eye were annotated in CVAT and exported as masks. To annotate air spaces in both the bud and soil aggregate, images were imported into ImageJ. A gaussian blur was applied to the image to decrease noise and then the air space was segmented using thresholding. After applying the threshold, the selected air space region was converted to a binary image with white representing the air space and black representing everything else. This binary image was overlaid upon the original image and the air space within the flower bud and aggregate was selected using the “free hand” tool. Air space outside of the region of interest for both image sets was eliminated. The quality of the air space annotation was then visually inspected for accuracy against the underlying original image; incomplete annotations were corrected using the brush or pencil tool to paint missing air space white and incorrectly identified air space black. Once the annotation was satisfactorily corrected, the binary image of the air space was saved. Finally, the annotations of the bud and flower or aggregate and organic matter were opened in ImageJ and the associated air space mask was overlaid on top of them forming a three-layer mask suitable for training the fully convolutional network. All labeling of the soil aggregate and soil aggregate images was done by Dr. Devin Rippner. These images and annotations are for training deep learning models to identify different constituents in leaves, almond buds, and soil aggregates Limitations: For the walnut leaves, some tissues (stomata, etc.) are not labeled and only represent a small portion of a full leaf. Similarly, both the almond bud and the aggregate represent just one single sample of each. The bud tissues are only divided up into buds scales, flower, and air space. Many other tissues remain unlabeled. For the soil aggregate annotated labels are done by eye with no actual chemical information. Therefore particulate organic matter identification may be incorrect. Resources in this dataset:Resource Title: Annotated X-ray CT images and masks of a Forest Soil Aggregate. File Name: forest_soil_images_masks_for_testing_training.zipResource Description: This aggregate was collected from the riparian oak forest at the Russell Ranch Sustainable Agricultural Facility. The aggreagate was scanned using X-ray micro-computed tomography (microCT) on the X-ray μCT beamline (8.3.2) at the Advanced Light Source (ALS) in Lawrence Berkeley National Laboratory (LBNL), Berkeley, CA USA) using the 10x objective lens with a pixel resolution of 650 nanometers. For masks, the background has a value of 0,0,0; pores spaces have a value of 250,250, 250; mineral solids have a value= 128,0,0; and particulate organic matter has a value of = 000,128,000. These files were used for training a model to segment the forest soil aggregate and for testing the accuracy, precision, recall, and f1 score of the model.Resource Title: Annotated X-ray CT images and masks of an Almond bud (P. Dulcis). File Name: Almond_bud_tube_D_P6_training_testing_images_and_masks.zipResource Description: Drought stressed almond flower bud (Prunis dulcis) from a plant housed at the University of California, Davis, was scanned by X-ray micro-computed tomography (microCT) on the X-ray μCT beamline (8.3.2) at the Advanced Light Source (ALS) in Lawrence Berkeley National Laboratory (LBNL), Berkeley, CA USA) using the 4x lens with a pixel resolution of 1.72 µm using. For masks, the background has a value of 0,0,0; air spaces have a value of 255,255, 255; bud scales have a value= 128,0,0; and flower tissues have a value of = 000,128,000. These files were used for training a model to segment the almond bud and for testing the accuracy, precision, recall, and f1 score of the model.Resource Software Recommended: Fiji (ImageJ),url: https://imagej.net/software/fiji/downloads Resource Title: Annotated X-ray CT images and masks of Walnut leaves (J. Regia) . File Name: 6_leaf_training_testing_images_and_masks_for_paper.zipResource Description: Stems were collected from genetically unique J. regia accessions at the 117 USDA-ARS-NCGR in Wolfskill Experimental Orchard, Winters, California USA to use as scion, and were grafted by Sierra Gold Nursery onto a commonly used commercial rootstock, RX1 (J. microcarpa × J. regia). We used a common rootstock to eliminate any own-root effects and to simulate conditions for a commercial walnut orchard setting, where rootstocks are commonly used. The grafted saplings were repotted and transferred to the Armstrong lathe house facility at the University of California, Davis in June 2019, and kept under natural light and temperature. Leaves from each accession and treatment were scanned using X-ray micro-computed tomography (microCT) on the X-ray μCT beamline (8.3.2) at the Advanced Light Source (ALS) in Lawrence Berkeley National Laboratory (LBNL), Berkeley, CA USA) using the 10x objective lens with a pixel resolution of 650 nanometers. For masks, the background has a value of 170,170,170; Epidermis value= 85,85,85; Mesophyll value= 0,0,0; Bundle Sheath Extension value= 152,152,152; Vein value= 220,220,220; Air value = 255,255,255.Resource Software Recommended: Fiji (ImageJ),url: https://imagej.net/software/fiji/downloads

Search
Clear search
Close search
Google apps
Main menu