89 datasets found
  1. Variable Message Signal annotated images for object detection

    • zenodo.org
    • portalcientifico.universidadeuropea.com
    zip
    Updated Oct 2, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Gonzalo de las Heras de Matías; Gonzalo de las Heras de Matías; Javier Sánchez-Soriano; Javier Sánchez-Soriano; Enrique Puertas; Enrique Puertas (2022). Variable Message Signal annotated images for object detection [Dataset]. http://doi.org/10.5281/zenodo.5904211
    Explore at:
    zipAvailable download formats
    Dataset updated
    Oct 2, 2022
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Gonzalo de las Heras de Matías; Gonzalo de las Heras de Matías; Javier Sánchez-Soriano; Javier Sánchez-Soriano; Enrique Puertas; Enrique Puertas
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    If you use this dataset, please cite this paper: Puertas, E.; De-Las-Heras, G.; Sánchez-Soriano, J.; Fernández-Andrés, J. Dataset: Variable Message Signal Annotated Images for Object Detection. Data 2022, 7, 41. https://doi.org/10.3390/data7040041

    This dataset consists of Spanish road images taken from inside a vehicle, as well as annotations in XML files in PASCAL VOC format that indicate the location of Variable Message Signals within them. Also, a CSV file is attached with information regarding the geographic position, the folder where the image is located, and the text in Spanish. This can be used to train supervised learning computer vision algorithms, such as convolutional neural networks. Throughout this work, the process followed to obtain the dataset, image acquisition, and labeling, and its specifications are detailed. The dataset is constituted of 1216 instances, 888 positives, and 328 negatives, in 1152 jpg images with a resolution of 1280x720 pixels. These are divided into 576 real images and 576 images created from the data-augmentation technique. The purpose of this dataset is to help in road computer vision research since there is not one specifically for VMSs.

    The folder structure of the dataset is as follows:

    • vms_dataset/
      • data.csv
      • real_images/
        • imgs/
        • annotations/
      • data-augmentation/
        • imgs/
        • annotations/

    In which:

    • data.csv: Each row contains the following information separated by commas (,): image_name, x_min, y_min, x_max, y_max, class_name, lat, long, folder, text.
    • real_images: Images extracted directly from the videos.
    • data-augmentation: Images created using data-augmentation
    • imgs: Image files in .jpg format.
    • annotations: Annotation files in .xml format.
  2. d

    Data from: Exploring deep learning techniques for wild animal behaviour...

    • datadryad.org
    • data.niaid.nih.gov
    • +1more
    zip
    Updated Jan 23, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Exploring deep learning techniques for wild animal behaviour classification using animal-borne accelerometers [Dataset]. https://datadryad.org/stash/dataset/doi:10.5061/dryad.2ngf1vhwk
    Explore at:
    zipAvailable download formats
    Dataset updated
    Jan 23, 2024
    Dataset provided by
    Dryad
    Authors
    Ryoma Otsuka; Naoya Yoshimura; Kei Tanigaki; Shiho Koyama; Yuichi Mizutani; Ken Yoda; Takuya Maekawa
    Time period covered
    2023
    Description

    Machine learning‐based behaviour classification using acceleration data is a powerful tool in bio‐logging research. Deep learning architectures such as convolutional neural networks (CNN), long short‐term memory (LSTM) and self‐attention mechanisms as well as related training techniques have been extensively studied in human activity recognition. However, they have rarely been used in wild animal studies. The main challenges of acceleration‐based wild animal behaviour classification include data shortages, class imbalance problems, various types of noise in data due to differences in individual behaviour and where the loggers were attached and complexity in data due to complex animal‐specific behaviours, which may have limited the application of deep learning techniques in this area. To overcome these challenges, we explored the effectiveness of techniques for efficient model training: data augmentation, manifold mixup and pre‐training of deep learning models with unlabelled data, usin...

  3. f

    Hyperparameters for each classification model.

    • plos.figshare.com
    xls
    Updated Sep 26, 2024
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Rodrigo Gutiérrez Benítez; Alejandra Segura Navarrete; Christian Vidal-Castro; Claudia Martínez-Araneda (2024). Hyperparameters for each classification model. [Dataset]. http://doi.org/10.1371/journal.pone.0310707.t007
    Explore at:
    xlsAvailable download formats
    Dataset updated
    Sep 26, 2024
    Dataset provided by
    PLOS ONE
    Authors
    Rodrigo Gutiérrez Benítez; Alejandra Segura Navarrete; Christian Vidal-Castro; Claudia Martínez-Araneda
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Over the last ten years, social media has become a crucial data source for businesses and researchers, providing a space where people can express their opinions and emotions. To analyze this data and classify emotions and their polarity in texts, natural language processing (NLP) techniques such as emotion analysis (EA) and sentiment analysis (SA) are employed. However, the effectiveness of these tasks using machine learning (ML) and deep learning (DL) methods depends on large labeled datasets, which are scarce in languages like Spanish. To address this challenge, researchers use data augmentation (DA) techniques to artificially expand small datasets. This study aims to investigate whether DA techniques can improve classification results using ML and DL algorithms for sentiment and emotion analysis of Spanish texts. Various text manipulation techniques were applied, including transformations, paraphrasing (back-translation), and text generation using generative adversarial networks, to small datasets such as song lyrics, social media comments, headlines from national newspapers in Chile, and survey responses from higher education students. The findings show that the Convolutional Neural Network (CNN) classifier achieved the most significant improvement, with an 18% increase using the Generative Adversarial Networks for Sentiment Text (SentiGan) on the Aggressiveness (Seriousness) dataset. Additionally, the same classifier model showed an 11% improvement using the Easy Data Augmentation (EDA) on the Gender-Based Violence dataset. The performance of the Bidirectional Encoder Representations from Transformers (BETO) also improved by 10% on the back-translation augmented version of the October 18 dataset, and by 4% on the EDA augmented version of the Teaching survey dataset. These results suggest that data augmentation techniques enhance performance by transforming text and adapting it to the specific characteristics of the dataset. Through experimentation with various augmentation techniques, this research provides valuable insights into the analysis of subjectivity in Spanish texts and offers guidance for selecting algorithms and techniques based on dataset features.

  4. Z

    Data from: Exploring deep learning techniques for wild animal behaviour...

    • data.niaid.nih.gov
    Updated Jan 23, 2024
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Maekawa, Takuya (2024). Data from: Exploring deep learning techniques for wild animal behaviour classification using animal-borne accelerometers [Dataset]. https://data.niaid.nih.gov/resources?id=zenodo_10557258
    Explore at:
    Dataset updated
    Jan 23, 2024
    Dataset provided by
    Mizutani, Yuichi
    Yoshimura, Naoya
    Yoda, Ken
    Otsuka, Ryoma
    Maekawa, Takuya
    Tanigaki, Kei
    Koyama, Shiho
    License

    CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
    License information was derived automatically

    Description

    1: Machine learning-based behaviour classification using acceleration data is a powerful tool in bio-logging research. Deep learning architectures such as convolutional neural networks (CNN), long short-term memory (LSTM), and self-attention mechanism as well as related training techniques have been extensively studied in human activity recognition. However, they have rarely been used in wild animal studies. The main challenges of acceleration-based wild animal behaviour classification include data shortages, class imbalance problems, various types of noise in data due to differences in individual behaviour and where the loggers were attached, and complexity in data due to complex animal-specific behaviours, which may have limited the application of deep learning techniques in this area.

    2: To overcome these challenges, we explored the effectiveness of techniques for efficient model training: data augmentation, manifold mixup, and pre-training of deep learning models with unlabelled data, using datasets from two species of wild seabirds and state-of-the-art deep learning model architectures.

    3: Data augmentation improved the overall model performance when one of various techniques (none, scaling, jittering, permutation, time-warping, and rotation) was randomly applied to each data during mini-batch training. Manifold mixup also improved model performance, but not as much as random data augmentation. Pre-training with unlabelled data did not improve model performance. The state-of-the-art deep learning models, including a model consisting of four CNN layers, an LSTM layer, and a multi-head attention layer, as well as its modified version with shortcut connection, showed better performance among other comparative models. Using only raw acceleration data as inputs, these models outperformed classic machine learning approaches that used 119 handcrafted features.

    4: Our experiments showed that deep learning techniques are promising for acceleration-based behaviour classification of wild animals and highlighted some challenges (e.g. effective use of unlabelled data). There is scope for greater exploration of deep learning techniques in wild animal studies (e.g. advanced data augmentation, multimodal sensor data use, transfer learning, and self-supervised learning). We hope that this study will stimulate the development of deep learning techniques for wild animal behaviour classification using time-series sensor data.

    This abstract is cited from the original article "Exploring deep learning techniques for wild animal behaviour classification using animal-borne accelerometers" in Methods in Ecology and Evolution (Otsuka et al., 2024).Please see README for the details of the datasets.

  5. n

    Data from: New Deep Learning Methods for Medical Image Analysis and...

    • curate.nd.edu
    pdf
    Updated Nov 11, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Pengfei Gu (2024). New Deep Learning Methods for Medical Image Analysis and Scientific Data Generation and Compression [Dataset]. http://doi.org/10.7274/26156719.v1
    Explore at:
    pdfAvailable download formats
    Dataset updated
    Nov 11, 2024
    Dataset provided by
    University of Notre Dame
    Authors
    Pengfei Gu
    License

    https://www.law.cornell.edu/uscode/text/17/106https://www.law.cornell.edu/uscode/text/17/106

    Description

    Medical image analysis is critical to biological studies, health research, computer- aided diagnoses, and clinical applications. Recently, deep learning (DL) techniques have achieved remarkable successes in medical image analysis applications. However, these techniques typically require large amounts of annotations to achieve satisfactory performance. Therefore, in this dissertation, we seek to address this critical problem: How can we develop efficient and effective DL algorithms for medical image analysis while reducing annotation efforts? To address this problem, we have outlined two specific aims: (A1) Utilize existing annotations effectively from advanced models; (A2) extract generic knowledge directly from unannotated images.

    To achieve the aim (A1): First, we introduce a new data representation called TopoImages, which encodes the local topology of all the image pixels. TopoImages can be complemented with the original images to improve medical image analysis tasks. Second, we propose a new augmentation method, SAMAug-C, that lever- ages the Segment Anything Model (SAM) to augment raw image input and enhance medical image classification. Third, we propose two advanced DL architectures, kCBAC-Net and ConvFormer, to enhance the performance of 2D and 3D medical image segmentation. We also present a gate-regularized network training (GrNT) approach to improve multi-scale fusion in medical image segmentation. To achieve the aim (A2), we propose a novel extension of known Masked Autoencoders (MAEs) for self pre-training, i.e., models pre-trained on the same target dataset, specifically for 3D medical image segmentation.

    Scientific visualization is a powerful approach for understanding and analyzing various physical or natural phenomena, such as climate change or chemical reactions. However, the cost of scientific simulations is high when factors like time, ensemble, and multivariate analyses are involved. Additionally, scientists can only afford to sparsely store the simulation outputs (e.g., scalar field data) or visual representations (e.g., streamlines) or visualization images due to limited I/O bandwidths and storage space. Therefore, in this dissertation, we seek to address this critical problem: How can we develop efficient and effective DL algorithms for scientific data generation and compression while reducing simulation and storage costs?

    To tackle this problem: First, we propose a DL framework that generates un- steady vector fields data from a set of streamlines. Based on this method, domain scientists only need to store representative streamlines at simulation time and recon- struct vector fields during post-processing. Second, we design a novel DL method that translates scalar fields to vector fields. Using this approach, domain scientists only need to store scalar field data at simulation time and generate vector fields from their scalar field counterparts afterward. Third, we present a new DL approach that compresses a large collection of visualization images generated from time-varying data for communicating volume visualization results.

  6. cars_wagonr_swift

    • kaggle.com
    zip
    Updated Sep 11, 2019
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Ajay (2019). cars_wagonr_swift [Dataset]. https://www.kaggle.com/ajaykgp12/cars-wagonr-swift
    Explore at:
    zip(44486490 bytes)Available download formats
    Dataset updated
    Sep 11, 2019
    Authors
    Ajay
    License

    https://creativecommons.org/publicdomain/zero/1.0/https://creativecommons.org/publicdomain/zero/1.0/

    Description

    Context

    Data science beginners start with curated set of data, but it's a well known fact that in a real Data Science Project, major time is spent on collecting, cleaning and organizing data . Also domain expertise is considered as important aspect of creating good ML models. Being an automobile enthusiast, I tool up this challenge to collect images of two of the popular car models from a used car website, where users upload the images of the car they want to sell and then train a Deep Neural Network to identify model of a car from car images. In my search for images I found that approximately 10 percent of the cars pictures did not represent the intended car correctly and those pictures have to be deleted from final data.

    Content

    There are 4000 images of two of the popular cars (Swift and Wagonr) in India of make Maruti Suzuki with 2000 pictures belonging to each model. The data is divided into training set with 2400 images , validation set with 800 images and test set with 800 images. The data was randomized before splitting into training, test and validation set.

    The starter kernal is provided for keras with CNN. I have also created github project documenting advanced techniques in pytorch and keras for image classification like data augmentation, dropout, batch normalization and transfer learning

    Inspiration

    1. With small dataset like this, how much accuracy can we achieve and whether more data is always better. The baseline model trained in Keras achieves 88% accuracy on validation set, can we achieve even better performance and by how much.

    2. Is the data collected for the two car models representative of all possible car from all over country or there is sample bias .

    3. I would also like someone to extend the concept to build a use case so that if user uploads an incorrect car picture of car , the ML model could automatically flag it. For example user uploading incorrect model or an image which is not a car

  7. S

    GalliformeSpectra: A Hen Breed Dataset

    • scidb.cn
    • data.mendeley.com
    Updated Oct 30, 2023
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Galib Muhammad Shahriar Himel; Md Masudul Islam (2023). GalliformeSpectra: A Hen Breed Dataset [Dataset]. http://doi.org/10.57760/sciencedb.12798
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Oct 30, 2023
    Dataset provided by
    Science Data Bank
    Authors
    Galib Muhammad Shahriar Himel; Md Masudul Islam
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    This dataset includes a variety of low-resolution images showcasing ten globally recognized hen breeds. These breeds were carefully chosen from different parts of the world to ensure a diverse and comprehensive representation. The dataset serves as a visual resource, offering a detailed depiction of the unique characteristics of these hen breeds, which aids in their accurate classification. It consists of ten distinct categories: Bielefeld, Blackorpington, Brahma, Buckeye, Fayoumi, Leghorn, Newhampshire, Plymouthrock, Sussex, and Turken, comprising a total of 1010 original JPG images that were later resized and converted to PNG format. After applying augmentation techniques, the total number of images increased to 5050. The dataset is organized into three variations: one with original images, another with resized images, and a third with augmented images. Each variation is further divided into ten separate folders, each dedicated to a specific hen breed. The images vary in size and have been subjected to data augmentation, which is essential for training machine vision deep learning models. Augmentation includes transformations like left and right flips, random adjustments in brightness (0.7-1.1), and random zoom (90% area). Consequently, an additional 505 augmented images were created from the original images in each category, resulting in a dataset comprising a total of 5050 augmented images (505 per category).

  8. ProxyFAUG: Proximity-based Fingerprint Augmentation (data)

    • zenodo.org
    • data.niaid.nih.gov
    csv
    Updated Jun 13, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Grigorios Anagnostopoulos; Grigorios Anagnostopoulos; Alexandros Kalousis; Alexandros Kalousis (2022). ProxyFAUG: Proximity-based Fingerprint Augmentation (data) [Dataset]. http://doi.org/10.5281/zenodo.4457391
    Explore at:
    csvAvailable download formats
    Dataset updated
    Jun 13, 2022
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Grigorios Anagnostopoulos; Grigorios Anagnostopoulos; Alexandros Kalousis; Alexandros Kalousis
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    The supplementary data of the paper "ProxyFAUG: Proximity-based Fingerprint Augmentation".

    Open access Author’s accepted manuscript version: https://arxiv.org/abs/2102.02706v2

    Published paper: https://ieeexplore.ieee.org/document/9662590

    The train/validation/test sets used in the paper "ProxyFAUG: Proximity-based Fingerprint Augmentation", after having passed the preprocessing process described in the paper, are made available here. Moreover, the augmentations produced by the proposed ProxyFAUG method are also made available with the files (x_aug_train.csv, y_aug_train.csv). More specifically:

    x_train_pre.csv : The features side (x) information of the preprocessed training set.

    x_val_pre.csv : The features side (x) information of the preprocessed validation set.

    x_test_pre.csv : The features side (x) information of the preprocessed test set.

    x_aug_train.csv : The features side (x) information of the fingerprints generated by ProxyFAUG.

    y_train.csv : The location ground truth information (y) of the training set.

    y_val.csv : The location ground truth information (y) of the validation set.

    y_test.csv : The location ground truth information (y) of the test set.

    y_aug_train.csv : The location ground truth information (y) of the fingerprints generated by ProxyFAUG.

    Note that in the paper, the original training set (x_train_pre.csv) is used as a baseline, and is compared against the scenario where the concatenation of the original and the generated training sets (concatenation of x_train_pre.csv and x_aug_train.csv) is used.

    The full code implementation related to the paper is available here:

    Code: https://zenodo.org/record/4457353

    -----------------------------------------------------------------------------------------------------------------------------------------------------------------

    The original full dataset used in this study, is the public dataset sigfox_dataset_antwerp.csv which can be access here:

    https://zenodo.org/record/3904158#.X4_h7y8RpQI

    The above link is related to the publication "Sigfox and LoRaWAN Datasets for Fingerprint Localization in Large Urban and Rural Areas", in which the original full dataset was published. The publication is available here:

    http://www.mdpi.com/2306-5729/3/2/13

    The credit for the creation of the original full dataset goes to Aernouts, Michiel; Berkvens, Rafael; Van Vlaenderen, Koen; and Weyn, Maarten.

    -----------------------------------------------------------------------------------------------------------------------------------------------------------------

    The train/validation/test split of the original dataset that is used in this paper, is taken from our previous work "A Reproducible Analysis of RSSI Fingerprinting for Outdoors Localization Using Sigfox: Preprocessing and Hyperparameter Tuning". Using the same train/validation/test split in different works strengthens the consistency of the comparison of results. All relevant material of that work is listed below:

    Preprint: https://arxiv.org/abs/1908.06851

    Paper: https://ieeexplore.ieee.org/document/8911792

    Code: https://zenodo.org/record/3228752

    Data: https://zenodo.org/record/3228744

  9. m

    An Image Dataset of Rice Varieties

    • data.mendeley.com
    Updated Dec 5, 2023
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Md Masudul Islam (2023). An Image Dataset of Rice Varieties [Dataset]. http://doi.org/10.17632/3mn9843tz2.3
    Explore at:
    Dataset updated
    Dec 5, 2023
    Authors
    Md Masudul Islam
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    This extensive dataset presents a meticulously curated collection of low-resolution images showcasing 20 well-established rice varieties native to diverse regions of Bangladesh. The rice samples were carefully gathered from both rural areas and local marketplaces, ensuring a comprehensive and varied representation. Serving as a visual compendium, the dataset provides a thorough exploration of the distinct characteristics of these rice varieties, facilitating precise classification.

    Dataset Composition

    The dataset encompasses 18 distinct classes, encompassing Subol Lota, Bashmoti (Deshi), Ganjiya, Shampakatari, Sugandhi Katarivog, BR-28, BR-29, Paijam, Bashful, Lal Aush, BR-Jirashail, Gutisharna, Birui, Najirshail, Pahari Birui, Polao (Katari), Polao (Chinigura), Amon, Shorna-5, and Lal Binni. In total, the dataset comprises 4,730 original JPG images and 23,650 augmented images.

    Image Capture and Dataset Organization

    These images were captured using an iPhone 11 camera with a 5x zoom feature. Each image capturing these rice varieties was diligently taken between October 18 and November 29, 2023. To facilitate efficient data management and organization, the dataset is structured into two variants: Original images and Augmented images. Each variant is systematically categorized into 20 distinct sub-directories, each corresponding to a specific rice variety.

    Original Image Dataset

    The primary image set comprises 4,730 JPG images, uniformly sized at 853 × 853 pixels. Due to the initial low resolution, the file size was notably 268 MB. Employing compression through a zip program significantly optimized the dataset, resulting in a final size of 254 MB.

    Augmented Image Dataset

    To address the substantial image volume requirements of deep learning models for machine vision, data augmentation techniques were implemented. Transformations such as rotation (90° left, 90° right, 180° left) and flip were applied, generating an additional set of augmented images in every class, totaling 23,650 augmented images. These augmented images, also in JPG format and uniformly sized at 512 × 512 pixels, initially amounted to 781 MB. However, post-compression, the dataset was further streamlined to 699 MB.

    Dataset Storage and Access

    The raw and augmented datasets are stored in two distinct zip files, namely 'Original.zip' and 'Augmented.zip'. Both zip files contain 20 sub-folders representing a unique rice variety, namely 1_Subol_Lota, 2_Bashmoti, 3_Ganjiya, 4_Shampakatari, 5_Katarivog, 6_BR28, 7_BR29, 8_Paijam, 9_Bashful, 10_Lal_Aush, 11_Jirashail, 12_Gutisharna, 13_Red_Cargo,14_Najirshail, 15_Katari_Polao, 16_Lal_Biroi, 17_Chinigura_Polao, 18_Amon, 19_Shorna5, 20_Lal_Binni.

  10. Data from: Image-based automated species identification: Can virtual data...

    • zenodo.org
    • data.niaid.nih.gov
    • +1more
    zip
    Updated Jun 4, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Morris Klasen; Morris Klasen; Jonas Eberle; Dirk Ahrens; Volker Steinhage; Jonas Eberle; Dirk Ahrens; Volker Steinhage (2022). Image-based automated species identification: Can virtual data augmentation overcome problems of insufficient sampling? [Dataset]. http://doi.org/10.5061/dryad.f1vhhmgx9
    Explore at:
    zipAvailable download formats
    Dataset updated
    Jun 4, 2022
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Morris Klasen; Morris Klasen; Jonas Eberle; Dirk Ahrens; Volker Steinhage; Jonas Eberle; Dirk Ahrens; Volker Steinhage
    License

    CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
    License information was derived automatically

    Description

    Automated species identification and delimitation is challenging, particularly in rare and thus often scarcely sampled species, which do not allow sufficient discrimination of infraspecific versus interspecific variation. Typical problems arising from either low or exaggerated interspecific morphological differentiation are best met by automated methods of machine learning that learn efficient and effective species identification from training samples. However, limited infraspecific sampling remains a key challenge also in machine learning.

    In this study, we assessed whether a data augmentation approach may help to overcome the problem of scarce training data in automated visual species identification. The stepwise augmentation of data comprised image rotation as well as visual and virtual augmentation. The visual data augmentation applies classic approaches of data augmentation and generation of artificial images using a Generative Adversarial Networks (GAN) approach. Descriptive feature vectors are derived from bottleneck features of a VGG-16 convolutional neural network (CNN) that are then stepwise reduced in dimensionality using Global Average Pooling and PCA to prevent overfitting. Finally, data augmentation employs synthetic additional sampling in feature space by an oversampling algorithm in vector space (SMOTE). Applied on four different image datasets, which include scarab beetle genitalia (Pleophylla, Schizonycha) as well as wing patterns of bees (Osmia) and cattleheart butterflies (Parides), our augmentation approach outperformed a deep learning baseline approach by means of resulting identification accuracy with non-augmented data as well as a traditional 2D morphometric approach (Procrustes analysis of scarab beetle genitalia).

  11. Fresh and Stale Images of Fruits and Vegetables

    • kaggle.com
    Updated May 17, 2021
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    RAGHAV R POTDAR (2021). Fresh and Stale Images of Fruits and Vegetables [Dataset]. https://www.kaggle.com/raghavrpotdar/fresh-and-stale-images-of-fruits-and-vegetables/code
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    May 17, 2021
    Dataset provided by
    Kaggle
    Authors
    RAGHAV R POTDAR
    License

    https://creativecommons.org/publicdomain/zero/1.0/https://creativecommons.org/publicdomain/zero/1.0/

    Description

    Context

    This dataset contains images of 6 fruits and vegetables: apple, banana, bitter gourd, capsicum, orange, and tomato. The images of each fruit or vegetable are grouped into two categories: fresh and stale. The purpose behind the creation of this dataset is the development of a machine learning model to classify fruits and vegetables as fresh or stale. This feature is a part of our final year project titled ‘Food Aayush’. (Github Link)

    Data Collection and Preprocessing

    For collecting the images to create the dataset, images of the fruits and vegetables were captured daily using a mobile phone camera. Depending on the visual properties of the fruit or vegetable in each image and on the day when the image was captured, each image was labelled as fresh or stale. Additionally, videos of the fruits and vegetables were taken, and the frames of these videos were extracted to collect a large number of images conveniently. The machine learning model requires a 224 x 224-pixel image. So, the images were cropped to extract the center square of the image and resized in 512 x 512 pixels using a data pre-processing library in Keras. Frame Extraction

    Data Augmentation: We used ImageDataGenerator library from Keras for augmentation. We on average created 20 augmentations per image which indeed improve our models accuracy. Data Augmentation

    Acknowledgements

    We would like to give credit to this dataset as we have obtained the images in some of the classes from here. Dataset

    Inspiration

    Our BE final year project, titled ‘Food Aayush’, is an application that can be used for the classification of fruits and vegetables as fresh or stale, the classification of cooking oils into different rancidity levels, and the analysis of various parameters related to the nutritional value of food and people’s dietary intake. We have trained a machine learning model for the classification of fruits and vegetables. This dataset was created for training the machine learning model. Your data will be in front of the world's largest data science community. What questions do you want to see answered?

  12. Data from: How many specimens make a sufficient training set for automated...

    • zenodo.org
    • search.dataone.org
    • +2more
    bin, zip
    Updated May 31, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    James M. Mulqueeney; James M. Mulqueeney; Alex Searle-Barnes; Anieke Brombacher; Marisa Sweeney; Anjali Goswami; Thomas H. G. Ezard; Alex Searle-Barnes; Anieke Brombacher; Marisa Sweeney; Anjali Goswami; Thomas H. G. Ezard (2024). Data from: How many specimens make a sufficient training set for automated three dimensional feature extraction? [Dataset]. http://doi.org/10.5061/dryad.1rn8pk12f
    Explore at:
    bin, zipAvailable download formats
    Dataset updated
    May 31, 2024
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    James M. Mulqueeney; James M. Mulqueeney; Alex Searle-Barnes; Anieke Brombacher; Marisa Sweeney; Anjali Goswami; Thomas H. G. Ezard; Alex Searle-Barnes; Anieke Brombacher; Marisa Sweeney; Anjali Goswami; Thomas H. G. Ezard
    License

    CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
    License information was derived automatically

    Measurement technique
    <p><strong>Data collection</strong></p> <p>50 planktonic foraminifera, comprising 4 <em>Menardella menardii</em>, 17 <em>Menardella limbata</em>, 18 <em>Menardella exilis</em>, and 11 <em>Menardella pertenuis</em> specimens, were used in our analyses (electronic supplementary material, figures S1 and S2). The taxonomic classification of these species was established based on the analysis of morphological characteristics observed in their shells. In this context, all species are characterised by lenticular, low trochosprial tests with a prominent keel [13]. Discrimination among these species is achievable, as <em>M. limbata</em> can be distinguished from its ancestor, <em>M. menardii</em>, by having a greater number of chambers and a smaller umbilicus. Moreover, <em>M. exilis</em> and <em>M. pertenuis</em> can be discerned from <em>M. limbata</em> by their thinner, more polished tests and reduced trochospirality. Furthermore, <em>M. pertenuis</em> is identifiable by a thin plate extending over the umbilicus and possessing a greater number of chambers in the final whorl compared to <em>M. exilis </em>[13].</p> <p>The samples containing these individuals and species spanned 5.65 million years ago (Ma) to 2.85 Ma [14] and were collected from the Ceara Rise in the Equatorial Atlantic region at Ocean Drilling Program (ODP) Site 925, which comprised Hole 925B (4°12.248'N, 43°29.349'W), Hole 925C 20 (4°12.256'N, 43°29.349'W), and Hole 925D (4°12.260'N, 43°29.363'W). See Curry et al., [15] for more details. This group was chosen to provide inter- and intraspecific species variation, and to provide contemporary data to test how morphological distinctiveness maps to taxonomic hypotheses [16].</p> <p>The non-destructive imaging of both internal and external structures of the foraminifera was conducted at the µ-VIS X-ray Imaging Centre, University of Southampton, UK, using a Zeiss Xradia 510 Versa X-ray tomography scanner. Employing a rotational target system, the scanner operated at a voltage of 110 kV and a power of 10 W. Projections were reconstructed using Zeiss Xradia software, resulting in 16-bit greyscale .tiff stacks characterised by a voxel size of 1.75 μm and an average dimension of 992 x 1015 pixels for each 2D slice.</p> <p><strong>Generation of training sets</strong></p> <p>We extracted the external calcite and internal cavity spaces from the micro-CT scans of the 50 individuals using manual segmentation within Dragonfly v. 2021.3 (Object Research Systems, Canada). This step took approximately 480 minutes per specimen (24,000 minutes total) and involved the manual labelling of 11,947 2D images. Segmentation data for each specimen were exported as multi-label (3 labels: external, internal, and background) 8-bit multipage .tiff stacks and paired with the original CT image data to allow for training (see figure 2).</p> <p>The 50 specimens were categorised into three distinct groups (electronic supplementary material, table S1): 20 training image stacks, 10 validation image stacks, and 20 test image stacks. From the training image category, we generated six distinct training sets, varying in size from 1 to 20 specimens (see table 1). These were used to assess the impact of training set size on segmentation accuracy, as determined through a comparative analysis against the validation set (see Section 2.3).</p> <p>From the initial six training sets, we created six additional training sets through data augmentation using the NumPy library [17] in Python. This augmentation method was chosen for its simplicity and accessibility to researchers with limited computational expertise, as it can be easily implemented using a straightforward batch code. This augmentation process entailed rotating the original images five times (the maximum amount permitted using this method), effectively producing six distinct 3D orientations per specimen for each of the original training sets (see figure 3). The augmented training sets comprised between 6 and 120 .tiff stacks (see table 1).</p> <p><strong>Training the neural networks</strong></p> <p>CNNs were trained using the offline version of Biomedisa, which utilises a 3D U-Net architecture [18] – the primary model employed for image segmentation [19], and is optimised using Keras with a TensorFlow backend. We used patches of size 64 x 64 x 64 voxels, which were then scaled to a size of 256 x 256 x 256 voxels. This scaling was performed to improve the network's ability to capture spatial features and mitigate potential information loss during training. We trained 3 networks for each of the training sets to check the extent of stochastic variation on the results [20].</p> <p>To train our models in Biomedisa, we used a stochastic gradient descent with a learning rate of 0.01, a decay of 1 × 10<sup>-6</sup>, momentum of 0.9, and Nesterov momentum enabled. A stride size of 32 pixels and a batch size of 24 samples per epoch were used alongside an automated cropping feature, which has been demonstrated to enhance accuracy [21]. The training of each network was performed on a Tesla V100S-PCIE-32GB graphics card with 30989 MB of available memory. All the analyses and training procedures were conducted on the High-Performance Computing (HPC) system at the Natural History Museum, London.</p> <p>To measure network accuracy, we used the Dice similarity coefficient (Dice score), a metric commonly used in used in biomedical image segmentation studies [22, 23]. The Dice score quantifies the level of overlap between two segmentations, providing a value between 0 (no overlap) and 1 (perfect match). </p> <p>We conducted experiments to evaluate the potential efficiency gains of using an early stopping mechanism within Biomedisa. After testing a variety of epoch limits, we opted for an early stopping criterion set at 25 epochs, which was found to be the lowest value as to which all models trained correctly for every training set. By "trained correctly" we mean if there is no increase in Dice score within a 25-epoch window, the optimal network is selected, and training is terminated. To gauge its impact of early stopping on network accuracy, we compared the results obtained from the original six training sets under early stopping to those obtained on a full run of 200 epochs.</p> <p><strong>Evaluation of feature extraction </strong></p> <p>We used the median accuracy network from each of the 12 training sets to produce segmentation data for the external and internal structures of the 20 test specimens. The median accuracy was selected as it provides a more robust estimate of performance by ensuring that outliers had less impact on the overall result. We then compared the volumetric and shape measurements from the manual data to those from each training set. The volumetric measurements were total volume (comprising both external and internal volumes) and percentage calcite (calculated as the ratio of external volume to internal volume, multiplied by 100).</p> <p>To compare shape, mesh data for the external and internal structures was generated from the segmentation data of the 12 training sets and the manual data. Meshes were decimated to 50,000 faces and smoothed before being scaled and aligned using Python and Generalized Procrustes Surface Analysis (GPSA) [24], respectively. Shape was then analysed using the landmark-free morphometry pipeline, as outlined by Toussaint et al., [25]. We used a kernel width of 0.1mm and noise parameter of 1.0 for both the analysis of shape for both the external and internal data, using a Keops kernel (PyKeops; <a href="https://pypi.org/project/pykeops/">https://pypi.org/project/pykeops/</a>) as it performs better with large data [25]. The analyses were run for 150 iterations, using an initial step size of 0.01. The manually generated mesh for the individual st049_bl1_fo2 was used as the atlas for both the external and internal shape comparisons. </p>
    Description

    Deep learning has emerged as a robust tool for automating feature extraction from 3D images, offering an efficient alternative to labour-intensive and potentially biased manual image segmentation methods. However, there has been limited exploration into the optimal training set sizes, including assessing whether artificial expansion by data augmentation can achieve consistent results in less time and how consistent these benefits are across different types of traits. In this study, we manually segmented 50 planktonic foraminifera specimens from the genus Menardella to determine the minimum number of training images required to produce accurate volumetric and shape data from internal and external structures. The results reveal unsurprisingly that deep learning models improve with a larger number of training images with eight specimens being required to achieve 95% accuracy. Furthermore, data augmentation can enhance network accuracy by up to 8.0%. Notably, predicting both volumetric and shape measurements for the internal structure poses a greater challenge compared to the external structure, due to low contrast differences between different materials and increased geometric complexity. These results provide novel insight into optimal training set sizes for precise image segmentation of diverse traits and highlight the potential of data augmentation for enhancing multivariate feature extraction from 3D images.

  13. Augmented Intelligence Market Analysis North America, Europe, APAC, South...

    • technavio.com
    Updated Oct 15, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Technavio (2024). Augmented Intelligence Market Analysis North America, Europe, APAC, South America, Middle East and Africa - US, UK, China, Japan, India - Size and Forecast 2024-2028 [Dataset]. https://www.technavio.com/report/augmented-intelligence-market-industry-analysis
    Explore at:
    Dataset updated
    Oct 15, 2024
    Dataset provided by
    TechNavio
    Authors
    Technavio
    Time period covered
    2021 - 2025
    Area covered
    United Kingdom, United States, Global
    Description

    Snapshot img

    Augmented Intelligence Market Size 2024-2028

    The augmented intelligence market size is forecast to increase by USD 61.3 billion at a CAGR of 33.1% between 2023 and 2028.

    Augmented Intelligence (IA) is revolutionizing business operations by amplifying human intelligence with advanced technologies such as Machine Learning (ML), Deep Learning, Natural Language Processing (NLP), and Virtual Assistants. IA is increasingly being adopted by enterprises to enhance decision-making capabilities and improve business outcomes. The implementation of IA in Business Intelligence (BI) tools is a significant trend, enabling organizations to derive insights from Big Data and perform predictive analytics.
    However, the shortage of IA experts poses a challenge to the widespread adoption of these technologies. ML and DL algorithms are integral to IA, enabling systems to learn and make decisions autonomously. NLP is used to understand human language and interact with virtual assistants, while Big Data and Data Analytics provide the foundation for IA applications. Predictive analytics is a key benefit of IA, enabling organizations to anticipate future trends and make informed decisions. IA is transforming various industries, including healthcare, finance, and retail, by augmenting human intelligence and automating routine tasks.
    

    What will be the Size of the Market During the Forecast Period?

    Request Free Sample

    Augmented Intelligence (IA), also known as Intelligence Amplification, refers to the use of advanced technologies such as machine learning (ML), deep learning (DL), and natural language processing (NLP) to support and enhance human intelligence. IA systems are designed to process vast amounts of data and provide insights that would be difficult or impossible for humans to identify on their own. Machine Learning and Deep Learning are at the core of IA systems. ML algorithms learn from data and improve their performance over time, while DL algorithms can identify complex patterns and relationships within data.
    Additionally, NLP enables computers to understand human language, enabling more effective communication between humans and machines. IA is being adopted across various industries, including streaming video services, factory automation, political think tanks, medical analysis, and more. In factory automation, IA systems are used to optimize production processes and improve quality control. In medical analysis, IA is used to analyze patient data and provide doctors with accurate diagnoses and treatment recommendations. In political think tanks, IA is used to analyze large datasets and identify trends and patterns. IA systems rely on big data and data analytics to function effectively.
    However, predictive analytics is a key application of IA, allowing organizations to make informed decisions based on data trends and patterns. Data scientists are essential in developing and implementing IA systems, ensuring that they are accurate, unbiased, and free from fatigue or distraction. Decision-making: IA systems are designed to augment human decision-making by providing accurate and relevant information in real-time. Autonomous systems and reactive machines are examples of IA applications that can make decisions based on data and environmental inputs. However, it is important to note that IA systems are not infallible and have an error rate that must be considered in decision-making.
    In conclusion, cybernetics, the study of communication and control in machines and living beings, plays a crucial role in IA development. Algorithms are used to process data and provide insights, and IA systems are designed to learn and adapt over time, improving their performance and accuracy. Limitations: IA systems are not without limitations. Bias in data can lead to inaccurate or unfair outcomes, and user viewing habits can influence the performance of recommendation systems. It is essential to address these limitations and ensure that IA systems are designed to augment human intelligence in a symbiotic relationship, rather than replacing it.
    

    How is this market segmented and which is the largest segment?

    The market research report provides comprehensive data (region-wise segment analysis), with forecasts and estimates in 'USD billion' for the period 2024-2028, as well as historical data from 2018-2022 for the following segments.

    Technology
    
      Machine learning
      NLP
      Computer vision
      Others
    
    
    Geography
    
      North America
    
        US
    
    
      Europe
    
        UK
    
    
      APAC
    
        China
        India
        Japan
    
    
      South America
    
    
    
      Middle East and Africa
    

    By Technology Insights

    The machine learning segment is estimated to witness significant growth during the forecast period.
    

    Augmented Intelligence, also known as Intelligence Amplification, is a technology that enhances human intelligence by integrating Machine Learning (ML), Deep Learning, Natural Language

  14. Data from: SEMFIRE forest dataset for semantic segmentation and data...

    • zenodo.org
    application/gzip, bin +2
    Updated Jan 20, 2022
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Dominik Bittner; Dominik Bittner; Maria Eduarda Andrada; Maria Eduarda Andrada; David Portugal; David Portugal; João Filipe Ferreira; João Filipe Ferreira (2022). SEMFIRE forest dataset for semantic segmentation and data augmentation [Dataset]. http://doi.org/10.5281/zenodo.5819064
    Explore at:
    zip, application/gzip, bin, txtAvailable download formats
    Dataset updated
    Jan 20, 2022
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Dominik Bittner; Dominik Bittner; Maria Eduarda Andrada; Maria Eduarda Andrada; David Portugal; David Portugal; João Filipe Ferreira; João Filipe Ferreira
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    SEMFIRE Datasets (Forest environment dataset)

    These datasets are used for semantic segmentation and data augmentation and contain various forestry scenes. They were collected as part of the research work conducted by the Institute of Systems and Robotics, University of Coimbra team within the scope of the Safety, Exploration and Maintenance of Forests with Ecological Robotics (SEMFIRE, ref. CENTRO-01-0247-FEDER-032691) research project coordinated by Ingeniarius Ltd.

    The semantic segmentation algorithms attempt to identify various semantic classes (e.g. background, live flammable materials, trunks, canopies etc.) in the images of the datasets.

    The datasets include diverse image types, e.g. original camera images and their labeled images. In total the SEMFIRE datasets include about 1700 image pairs. Each dataset includes corresponding .bag files.

    To launch those .bag files on your ROS environment, use the instructions on the following Github repository

    Description of each dataset:

    1. 2019_2020_quinta_do_bolao_coimbra: Robot moving on a path through a forest environment
    2. 2020_ctcv_parking_lot_coimbra: Robot moving in a circle in a parking lot for testings
    3. 2020_sete_fontes_forest: A set of forest images acquired by hand-held apparatus

    Each dataset consists of following directories:

    1. images directory: diverse image types, e.g. original camera images and their labeled images
    2. rosbags directory: .bag files, which correspond to the image directory

    Each images directory consists of following directories:

    • img: original camera images
    • lbl: single channel images (ground truth) with corresponding labels for each image in img
    • lbl_colored: camera images in lbl colorized according to different semantic classes (for more details see the datasets descriptions)
    • lbl_overlaid: camera images in img overlaid with corresponding labels (colored)

    Each rosbags directory contains .bag files with the following topics:

    • 2019_2020_quinta_do_bolao_coimbra_rosbags:
      • /back_lslidar_packet
      • /dalsa_camera_720p/compressed
      • /flir_ax8/compressed
      • /front_lslidar_packet
      • /gps_fix
      • /gps_time
      • /gps_vel
      • /imu/data
      • /realsense/aligned_depth_to_color/image_raw
      • /realsense/color/camera_info
      • /realsense/color/image_raw/compressed
      • /realsense/depth/camera_info
      • /realsense/depth/image_rect_raw/compressed
      • /realsense/extrinsics/depth_to_color
    • 2020_ctcv_parking_lot_coimbra_rosbags:
      • /dalsa_camera_720p/compressed
      • /gps_fix
      • /gps_ime
      • /fused_point_cloud
      • /imu/data
      • /imu/mag
      • /imu/rpy
    • 2020_sete_fontes_forest_rosbags:
      • /realsense/camera_info
      • /realsense/depth_compressed/compressedDepth
      • /realsense/nir/left/compressed
      • /realsense/nir/right/compressed
      • /realsense/rgb/compressed

    All datasets include a detailed description as a text file. In addition, they include a rosbag_info.txt file with a description for each ROS inside the .bag files as well as a description for each ROS topic.

    The following table shows the statistical description of typical portuguese woodland configurations with structured plantations of Pinus pinaster (Pp, pine trees) and Eucalyptus globulus (Eg, eucalyptus).

    "Low density" structured plantation"High density" structured plantation
    Tree density (assuming plantation in rows spaced 3m apart in all cases)

    Eg: 900 trees/ha

    Pp: 450 trees/ha

    Eg: 1400 trees/ha

    Pp: 1250 trees/ha

    Average heights and corresponding ages of plantation trees

    Eg: 12m (6 years old)

    Pp: 10m (15 years old)

    Eg: 12m (6 years old)

    Pp: 10m (15 years old)

    Maximum heights and corresponding fully-matured ages of plantation trees

    Eg: 20m (11 years old)

    Pp: 30m (40 years old)

    Eg: 20m (11 years old)

    Pp: 30m (40 years old)

    Diameter at chest level (DCL – 1,3m) of plantation trees (average/maximum)

    Eg: 15cm/25cm

    Pp: 20cm/50cm

    Eg: 15cm/25cm

    Pp: 20cm/50cm

    Natural density of herbaceous plants

    30% of woodland area

    30% of woodland area

    Natural density of bush and shrubbery

    30% of woodland area

    30% of woodland area

    Natural density of arboreal plants (not part of plantation)

    5% of woodland area

    5% of woodland area

  15. f

    Data from: BEGAN: Boltzmann-Reweighted Data Augmentation for Enhanced...

    • figshare.com
    • acs.figshare.com
    xlsx
    Updated Nov 21, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Jialei Dai; Yutong Zhang; Chen Shi; Yang Liu; Peng Xiu; Yong Wang (2024). BEGAN: Boltzmann-Reweighted Data Augmentation for Enhanced GAN-Based Molecule Design in Insect Pheromone Receptors [Dataset]. http://doi.org/10.1021/acs.jpcb.4c06729.s003
    Explore at:
    xlsxAvailable download formats
    Dataset updated
    Nov 21, 2024
    Dataset provided by
    ACS Publications
    Authors
    Jialei Dai; Yutong Zhang; Chen Shi; Yang Liu; Peng Xiu; Yong Wang
    License

    Attribution-NonCommercial 4.0 (CC BY-NC 4.0)https://creativecommons.org/licenses/by-nc/4.0/
    License information was derived automatically

    Description

    Identifying small molecules that bind strongly to target proteins in rational molecular design is crucial. Machine learning techniques, such as generative adversarial networks (GAN), are now essential tools for generating such molecules. In this study, we present an enhanced method for molecule generation using objective-reinforced GANs. Specifically, we introduce BEGAN (Boltzmann-enhanced GAN), a novel approach that adjusts molecule occurrence frequencies during training based on the Boltzmann distribution exp(−ΔU/τ), where ΔU represents the estimated binding free energy derived from docking algorithms and τ is a temperature-related scaling hyperparameter. This Boltzmann reweighting process shifts the generation process toward molecules with higher binding affinities, allowing the GAN to explore molecular spaces with superior binding properties. The reweighting process can also be refined through multiple iterations without altering the overall distribution shape. To validate our approach, we apply it to the design of sex pheromone analogs targeting Spodoptera frugiperda pheromone receptor SfruOR16, illustrating that the Boltzmann reweighting significantly increases the likelihood of generating promising sex pheromone analogs with improved binding affinities to SfruOR16, further supported by atomistic molecular dynamics simulations. Furthermore, we conduct a comprehensive investigation into parameter dependencies and propose a reasonable range for the hyperparameter τ. Our method offers a promising approach for optimizing molecular generation for enhanced protein binding, potentially increasing the efficiency of molecular discovery pipelines.

  16. P

    Grapevine Leaves Image Dataset Dataset

    • paperswithcode.com
    Updated Nov 30, 2022
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    (2022). Grapevine Leaves Image Dataset Dataset [Dataset]. https://paperswithcode.com/dataset/grapevine-leaves-image-dataset
    Explore at:
    Dataset updated
    Nov 30, 2022
    Description

    KOKLU Murat (a), UNLERSEN M. Fahri (b), OZKAN Ilker Ali (a), ASLAN M. Fatih(c), SABANCI Kadir (c)

    (a) Department of Computer Engineering, Selcuk University, Turkey, Konya, Turkey (b) Department of Electrical and Electronics Engineering, Necmettin Erbakan University, Konya, Turkey (c) Department of Electrical-Electronic Engineering, Karamanoglu Mehmetbey University, Karaman, Turkey

    Citation Request : Koklu, M., Unlersen, M. F., Ozkan, I. A., Aslan, M. F., & Sabanci, K. (2022). A CNN-SVM study based on selected deep features for grapevine leaves classification. Measurement, 188, 110425. Doi:https://doi.org/10.1016/j.measurement.2021.110425

    Link: https://doi.org/10.1016/j.measurement.2021.110425

    https://www.kaggle.com/mkoklu42 DATASET: https://www.muratkoklu.com/datasets/

    Highlights • Classification of five classes of grapevine leaves by MobileNetv2 CNN Model. • Classification of features using SVMs with different kernel functions. • Implementing a feature selection algorithm for high classification percentage. • Classification with highest accuracy using CNN-SVM Cubic model.

    Abstract: The main product of grapevines is grapes that are consumed fresh or processed. In addition, grapevine leaves are harvested once a year as a by-product. The species of grapevine leaves are important in terms of price and taste. In this study, deep learning-based classification is conducted by using images of grapevine leaves. For this purpose, images of 500 vine leaves belonging to 5 species were taken with a special self-illuminating system. Later, this number was increased to 2500 with data augmentation methods. The classification was conducted with a state-of-art CNN model fine-tuned MobileNetv2. As the second approach, features were extracted from pre-trained MobileNetv2′s Logits layer and classification was made using various SVM kernels. As the third approach, 1000 features extracted from MobileNetv2′s Logits layer were selected by the Chi-Squares method and reduced to 250. Then, classification was made with various SVM kernels using the selected features. The most successful method was obtained by extracting features from the Logits layer and reducing the feature with the Chi-Squares method. The most successful SVM kernel was Cubic. The classification success of the system has been determined as 97.60%. It was observed that feature selection increased the classification success although the number of features used in classification decreased.

    Keywords: Deep learning, Transfer learning, SVM, Grapevine leaves, Leaf identification

  17. d

    UiT_TILs - Replication Data for \"A Pragmatic Machine Learning Approach to...

    • search.dataone.org
    • dataverse.no
    Updated Jan 5, 2024
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Kilvaer, Thomas K (2024). UiT_TILs - Replication Data for \"A Pragmatic Machine Learning Approach to Quantify Tumor Infiltrating Lymphocytes in Whole Slide Images\" [Dataset]. https://search.dataone.org/view/sha256%3A4eef9d178e9c9237ad4ac29c41c448973f1169c26df38a0339d92f9efb127c3e
    Explore at:
    Dataset updated
    Jan 5, 2024
    Dataset provided by
    DataverseNO
    Authors
    Kilvaer, Thomas K
    Time period covered
    Jan 1, 1993 - Jan 1, 2003
    Description

    This dataset can be used to replicate the findings in "A Pragmatic Machine Learning Approach to Quantify Tumor Infiltrating Lymphocytes in Whole Slide Images". The motivation for this paper is that increased levels of tumor infiltrating lymphocytes (TILs) indicate favorable outcomes in many types of cancer. Our aim is to leverage computational pathology to automatically quantify TILs in standard diagnostic whole-tissue hematoxylin and eosin stained section slides (H&E slides). Our approach is to transfer an open source machine learning method for segmentation and classification of nuclei in H&E slides trained on public data to TIL quantification without manual labeling of our data. Our results show that improved data augmentation improves immune cell detection in H&E WSIs. Moreover, the resulting TIL quantification correlates to patient prognosis and compares favorably to the current state-of-the-art method for immune cell detection in non-small lung cancer (current standard CD8 cells in DAB stained TMAs HR 0.34 95% CI 0.17-0.68 vs TILs in HE WSIs: HoVer-Net PanNuke Model HR 0.30 95% CI 0.15-0.60). Moreover, we implemented a cloud based system to train, deploy, and visually inspect machine learning based annotation for H&E slides. Our pragmatic approach bridges the gap between machine learning research, translational clinical research and clinical implementation. However, validation in prospective studies is needed to assert that the method works in a clinical setting. The dataset is comprised of three parts: 1) Twenty image patches with and without overlays used by pathologists to manually evaluate the output of the deep learning models, 2) The models trained and subsequently used for inference in the paper, 3) the patient dataset with corresponding image patches used to clinically validate the output of the deep learning models. The tissue samples were collected from patients diagnosed between 1993 and 2003. Supplementing information was collected retrospectively in the time period 2006-2017. The images were produced in 2017.

  18. Concrete Crack Images for Classification

    • kaggle.com
    Updated Apr 29, 2022
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    ArnavR (2022). Concrete Crack Images for Classification [Dataset]. https://www.kaggle.com/datasets/arnavr10880/concrete-crack-images-for-classification/data
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Apr 29, 2022
    Dataset provided by
    Kagglehttp://kaggle.com/
    Authors
    ArnavR
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Description

    The dataset contains concrete images having cracks. The data is collected from various METU Campus Buildings. The dataset is divided into two as negative and positive crack images for image classification. Each class has 20000images with a total of 40000 images with 227 x 227 pixels with RGB channels. The dataset is generated from 458 high-resolution images (4032x3024 pixel) with the method proposed by Zhang et al (2016). High-resolution images have variance in terms of surface finish and illumination conditions. No data augmentation in terms of random rotation or flipping is applied.

    Acknowledgements

    Özgenel, Çağlar Fırat (2019), “Concrete Crack Images for Classification”, Mendeley Data, V2, doi: 10.17632/5y9wdsg2zt.2

  19. Z

    COMPOSED FAULT DATASET (COMFAULDA)

    • data.niaid.nih.gov
    • ieee-dataport.org
    • +1more
    Updated Jul 11, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    DenysPestana-Viana (2024). COMPOSED FAULT DATASET (COMFAULDA) [Dataset]. https://data.niaid.nih.gov/resources?id=zenodo_8208454
    Explore at:
    Dataset updated
    Jul 11, 2024
    Dataset provided by
    Dionísio Martins
    Amaro Lima
    Luiz Vaz
    Diego Hadadd
    DenysPestana-Viana
    Ricardo Homero
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    The measurement and diagnosis of the severity of failures in rotating machines allow the execution of predictive maintenance actions on equipment. These actions make it possible to monitor the operating parameters of the machine and to perform the prediction of failures, thus avoiding production losses, severe damage to the equipment, and safeguarding the integrity of the equipment operators. This paper describes the construction of a dataset composed of vibration signals of a rotating machine. The acquisition has taken into consideration seven distinct operating scenarios, with different speed values. Unlike the few datasets that currently exist, the resulting dataset contains simple and combined faults with several severity levels. The considered operating setups are normal condition, unbalance, horizontal misalignment, vertical misalignment, unbalance combined with horizontal misalignment, unbalance combined with vertical misalignment, and vertical misalignment combined with horizontal misalignment. The dataset described in this paper can be utilized by machine learning researchers that intend to detect faults in rotating machines in an automatic manner. In this context, several related topics might be investigated, such as feature extraction and/or selection, reduction of feature space, data augmentation methods, and prognosis of rotating machines through the analysis of failure severity parameters.

  20. F

    Database and code to the paper: A multiscale CNN-based intrinsic...

    • data.uni-hannover.de
    zip
    Updated Oct 1, 2024
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Institut für Baumechanik und Numerische Mechanik (2024). Database and code to the paper: A multiscale CNN-based intrinsic permeability prediction in deformable porous media [Dataset]. https://data.uni-hannover.de/dataset/data_and_ml_code
    Explore at:
    zip(938061358), zip(188562)Available download formats
    Dataset updated
    Oct 1, 2024
    Dataset authored and provided by
    Institut für Baumechanik und Numerische Mechanik
    License

    Attribution-NonCommercial 3.0 (CC BY-NC 3.0)https://creativecommons.org/licenses/by-nc/3.0/
    License information was derived automatically

    Description

    This work in the related paper introduces a novel application for predicting the macroscopic intrinsic permeability tensor in deformable porous media, using a limited set of micro-CT images of real microgeometries. The primary goal is to develop an efficient, machine-learning (ML)-based method that overcomes the limitations of traditional permeability estimation techniques, which often rely on time-consuming experiments or computationally expensive fluid dynamics simulations. The novelty of this work lies in leveraging Convolutional Neural Networks (CNN) to predict pore-fluid flow behavior under deformation and anisotropic flow conditions. Particularly, the described approach employs binarized CT images of porous micro-structure as inputs to predict the symmetric second-order permeability tensor, a critical parameter in continuum porous media flow modeling. The methodology comprises four key steps: (1) constructing a dataset of CT images from Bentheim sandstone at different volumetric strain levels; (2) performing pore-scale simulations of single-phase flow using the lattice Boltzmann method (LBM) to generate permeability data; (3) training the CNN model with the processed CT images as inputs and permeability tensors as outputs; and (4) exploring techniques to improve model generalization, including data augmentation and alternative CNN architectures. Examples are provided to demonstrate the CNN’s capability to accurately predict the permeability tensor, a crucial parameter in various disciplines such as geotechnical engineering, hydrology, and material science.

Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
Gonzalo de las Heras de Matías; Gonzalo de las Heras de Matías; Javier Sánchez-Soriano; Javier Sánchez-Soriano; Enrique Puertas; Enrique Puertas (2022). Variable Message Signal annotated images for object detection [Dataset]. http://doi.org/10.5281/zenodo.5904211
Organization logo

Variable Message Signal annotated images for object detection

Explore at:
5 scholarly articles cite this dataset (View in Google Scholar)
zipAvailable download formats
Dataset updated
Oct 2, 2022
Dataset provided by
Zenodohttp://zenodo.org/
Authors
Gonzalo de las Heras de Matías; Gonzalo de las Heras de Matías; Javier Sánchez-Soriano; Javier Sánchez-Soriano; Enrique Puertas; Enrique Puertas
License

Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically

Description

If you use this dataset, please cite this paper: Puertas, E.; De-Las-Heras, G.; Sánchez-Soriano, J.; Fernández-Andrés, J. Dataset: Variable Message Signal Annotated Images for Object Detection. Data 2022, 7, 41. https://doi.org/10.3390/data7040041

This dataset consists of Spanish road images taken from inside a vehicle, as well as annotations in XML files in PASCAL VOC format that indicate the location of Variable Message Signals within them. Also, a CSV file is attached with information regarding the geographic position, the folder where the image is located, and the text in Spanish. This can be used to train supervised learning computer vision algorithms, such as convolutional neural networks. Throughout this work, the process followed to obtain the dataset, image acquisition, and labeling, and its specifications are detailed. The dataset is constituted of 1216 instances, 888 positives, and 328 negatives, in 1152 jpg images with a resolution of 1280x720 pixels. These are divided into 576 real images and 576 images created from the data-augmentation technique. The purpose of this dataset is to help in road computer vision research since there is not one specifically for VMSs.

The folder structure of the dataset is as follows:

  • vms_dataset/
    • data.csv
    • real_images/
      • imgs/
      • annotations/
    • data-augmentation/
      • imgs/
      • annotations/

In which:

  • data.csv: Each row contains the following information separated by commas (,): image_name, x_min, y_min, x_max, y_max, class_name, lat, long, folder, text.
  • real_images: Images extracted directly from the videos.
  • data-augmentation: Images created using data-augmentation
  • imgs: Image files in .jpg format.
  • annotations: Annotation files in .xml format.
Search
Clear search
Close search
Google apps
Main menu