Facebook
TwitterThis dataset contains images in png format for RSNA Screening Mammography Breast Cancer Detection competition. All the images have a 768x768 size and are organized in two folders (one for each label). This allows inferring the labels from the folders (which might be helpful when using keras image_dataset_from_directory for example). - The processing of the DICOM files was done in this notebook by (Radek Osmulski ): https://www.kaggle.com/code/radek1/how-to-process-dicom-images-to-pngs?scriptVersionId=113529850 - The contribution made in this dataset is to organize the files in a way that enables label inferring from the directory name.
🎉If this dataset helps you in your work don't hesitate to upvote 🎉
Facebook
TwitterAttribution-ShareAlike 4.0 (CC BY-SA 4.0)https://creativecommons.org/licenses/by-sa/4.0/
License information was derived automatically
Overview
This dataset is a cleaned, standardized, and comprehensive collection of plant leaf images designed for training high-accuracy classification models. It addresses a common challenge in agricultural computer vision by merging four popular but distinctly formatted datasets (PlantVillage, PlantDoc, PlantWild, and PlantSeg).
The primary goal is to provide a clean and robust dataset. All images have been organized by crop_name/disease_name/image.jpg, and all directory names have been standardized to a snake_case format. Furthermore, ambiguous or duplicate class names (e.g., Apple scab and Scab) have been merged into single, unified categories.
This makes the dataset directly compatible with modern deep learning frameworks like PyTorch and TensorFlow.
The data is organized in a hierarchical format perfect for use with ImageFolder-style data loaders. All directories have been standardized to lowercase snake_case.
/ ├── apple/ │ ├── scab/ │ │ ├── image1.jpg │ │ └── ... │ ├── black_rot/ │ └── ... ├── tomato/ │ ├── bacterial_spot/ │ └── ... ├── corn/ │ └── ... └── ...
crop_name (e.g., apple, tomato, corn)disease_name (e.g., scab, healthy, leaf_mold)This structure allows for easy training of both large, multi-crop models and specialized, crop-specific submodels.
Source Datasets & Loading
This dataset was created by merging the following public sources:
A comprehensive cleaning process was applied to merge duplicate/synonymous disease folders and standardize all folder names.
The associated "starter notebook" provides the essential MultiCropDiseaseDataset class required to easily load this complex, multi-crop structure. This class correctly parses the folders and returns three items for each image (image, crop_label, and disease_label) for direct use in a PyTorch DataLoader.
other_crops FolderYou will find a folder named other_crops. This folder contains a wide variety of other plants (e.g., bean_halo_blight, rice_blast) from the source datasets. The MultiCropDiseaseDataset class in the starter notebook will load this folder along with all other crops, treating other_crops as its own distinct category.
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Dataset Introduction The advent of neural networks capable of learning salient features from variance in the radar data has expanded the breadth of radar applications, often as an alternative sensor or a complementary modality to camera vision. Gesture recognition for command control is arguably the most commonly explored application. Nevertheless, more suitable benchmarking datasets than currently available are needed to assess and compare the merits of the different proposed solutions and explore a broader range of scenarios than simple hand-gesturing a few centimeters away from a radar transmitter/receiver. Most current publicly available radar datasets used in gesture recognition provide limited diversity, do not provide access to raw ADC data, and are not significantly challenging. To address these shortcomings, we created and make available a new dataset that combines FMCW radar and dynamic vision camera of 10 aircraft marshalling signals (whole body) at several distances and angles from the sensors, recorded from 13 people. The two modalities are hardware synchronized using the radar's PRI signal. Moreover, in the supporting publication we propose a sparse encoding of the time domain (ADC) signals that achieve a dramatic data rate reduction (>76%) while retaining the efficacy of the downstream FFT processing (<2% accuracy loss on recognition tasks), and can be used to create an sparse event-based representation of the radar data. In this way the dataset can be used as a two-modality neuromorphic dataset. Synchronization of the two modalities The PRI pulses from the radar have been hard-wired to the event stream of the DVS sensor, and timestamped using the DVS clock. Based on this signal the DVS event stream has been segmented such that groups of events (time-bins) of the DVS are mapped with individual radar pulses (chirps). Data storage DVS events (x,y coords and timestamps) are stored in structured arrays, and one such structured array object is associated with the data of a radar transmission (pulse/chirp). A radar transmission is a vector of 512 ADC levels that correspond to sampling points of chirping signal (FMCW radar) that lasts about ~1.3ms. Every 192 radar transmissions are stacked in a matrix called a radar frame (each transmission is a row in that matrix). A data capture (recording) consisting of some thousands of continuous radar transmissions is therefore segmented in a number of radar frames. Finally radar frames and the corresponding DVS structured arrays are stored in separate containers in a custom-made multi-container file format (extension .rad). We provide a (rad file) parser for extracting the data out of these files. There is one file per capture of continuous gesture recording of about 10s. Note the number of 192 transmissions per radar frame is an ad-hoc segmentation that suits the purpose of obtaining sufficient signal resolution in a 2D FFT typical in radar signal processing, for the range resolution of the specific radar. It also served the purpose of fast streaming storing of the data during capture. For extracting individual data points for the dataset however, one can pool together (concat) all the radar frames from a single capture file and re-segment them according to liking. The data loader that we provide offers this, with a default of re-segmenting every 769 transmissions (about 1s of gesturing). Data captures directory organization (radar8Ghz-DVS-marshaling_signals_20220901_publication_anonymized.7z) The dataset captures (recordings) are organized in a common directory structure which encompasses additional metadata information about the captures. dataset_dir///--/ofxRadar8Ghz_yyyy-mm-dd_HH-MM-SS.rad Identifiers
stage [train, test].
room: [conference_room, foyer, open_space].
subject: [0-9]. Note that 0 stands for no person, and 1 for an unlabeled, random person (only present in test).
gesture: ['none', 'emergency_stop', 'move_ahead', 'move_back_v1', 'move_back_v2', 'slow_down' 'start_engines', 'stop_engines', 'straight_ahead', 'turn_left', 'turn_right'].
distance: 'xxx', '100', '150', '200', '250', '300', '350', '400', '450'. Note that xxx is used for none gestures when there is no person present in front of the radar (i.e. background samples), or when a person is walking in front of the radar with varying distances but performing no gesture.
The test data captures contain both subjects that appear in the train data as well as previously unseen subjects. Similarly the test data contain captures from the spaces that train data were recorded at, as well as from a new unseen open space.
Files List
radar8Ghz-DVS-marshaling_signals_20220901_publication_anonymized.7z
This is the actual archive bundle with the data captures (recordings).
rad_file_parser_2.py
Parser for individual .rad files, which contain capture data.
loader.py
A convenience PyTorch Dataset loader (partly Tonic compatible). You practically only need this to quick-start if you don't want to delve too much into code reading. When you init a DvsRadarAircraftMarshallingSignals class object it automatically downloads the dataset archive and the .rad file parser, unpacks the archive, and imports the .rad parser to load the data. One can then request from it a training set, a validation set and a test set as torch.Datasets to work with.
aircraft_marshalling_signals_howto.ipynb
Jupyter notebook for exemplary basic use of loader.py
Contact
For further information or questions try contacting first M. Sifalakis or F. Corradi.
Facebook
TwitterPytorch Image Models (timm)
timm is a deep-learning library created by Ross Wightman and is a collection of SOTA computer vision models, layers, utilities, optimizers, schedulers, data-loaders, augmentations and also training/validating scripts with ability to reproduce ImageNet training results.
https://github.com/rwightman/pytorch-image-models
# Installation
!python -m pip install /kaggle/input/timm067py3/timm-0.6.7-py3-none-any.whl
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
This dataset contains a complete deep learning pipeline for digitizing ECG images into time series data for the PhysioNet ECG Image Digitization Challenge. The solution extracts 12-lead ECG signals from degraded images including scans, photos, and physically damaged printouts.
Encoder: ResNet-inspired CNN with residual connections
Decoder: Bidirectional LSTM with attention
The solution includes comprehensive image enhancement:
Training robustness achieved through realistic augmentations: - Gaussian noise injection (simulates scanning noise) - Motion and Gaussian blur (simulates camera shake) - Grid distortion (simulates paper warping) - Random brightness/contrast adjustments - Elastic transformations (simulates physical deformation)
model.py: Neural network architecture (ResNet encoder + LSTM decoder)preprocessing.py: Image preprocessing and augmentation pipelinedataset.py: PyTorch data loaders for training and inferencemetrics.py: Modified SNR evaluation metrictrain.py: Complete training loop with checkpointinginference.py: Batch prediction and submission generationapp.py: Streamlit interactive visualization interfaceensemble.py: Multi-model ensemble methodspostprocessing.py: Signal refinement and constraint enforcementtransfer_learning.py: Pre-trained backbone integrationcross_validation.py: K-fold validation frameworkhyperparameter_tuning.py: Automated grid searchconfig.py: Centralized configuration managementutils.py: Helper functions for data handlingkaggle_submission_standalone.py: Self-contained Kaggle notebook
Facebook
TwitterAttribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Immobilized fluorescently stained zebrafish through the eXtended Field of view Light Field Microscope 2D-3D dataset
This dataset comprises three immobilized fluorescently stained zebrafish imaged through the eXtended Field of view Light Field Microscope (XLFM, also known as Fourier Light Field Microscope). The images were preprocessed with the SLNet, which extracts the sparse signals from the images (a.k.a. the neural activity).
If you intend to use this with Pytorch, you can find a data loader and working source code to load and train networks here.
This dataset is part of the publication: Fast light-field 3D microscopy with out-of-distribution detection and adaptation through Conditional Normalizing Flows.
The fish present are:
1x NLS GCaMP6s
1x Pan-neuronal nuclear localized GCaMP6s Tg(HuC:H2B:GCaMP6s)
1x Soma localized GCaMP7f Tg(HuC:somaGCaMP7f)
The dataset is structured as follows::
XLFM_dataset
Dataset/
GCaMP6s_NLS_1/
SLNet_preprocessed/
XLFM_image/
XLFM_image_stack.tif: tif stack of 600 preprocessed XLFM images.
XLFM_stack/
XLFM_stack_nnn.tif: 3D stack corresponding to frame nnn.
Neural_activity_coordinates.csv: 3D coordinates of neurons found through the suite2p framework.
Raw/
XLFM_image/
XLFM_image_stack.tif: tif stack of 600 raw XLFM images.
(other samples)
lenslet_centers_python.txt: 2D coordinates of the lenset in the XLFM images.
PSF_241depths_16bit.tif: 3D PSF of the microscope can be used for 3D deconvolution. Spanning 734 × 734 × 550𝜇𝑚3 used to deconvolve this volumes.
In this dataset, we provide a subset of the images and volumes.
Due to space constraints, we provide the 3D volumes only for:
SLNet_preprocessed/XLFM_stack/
10 interleaved frames between frames 0-499 (can be used for training a network).
20 consecutive frames, 500-520 (can be used for testing).
raw/
No volumes are provided for raw data, but they can be reconstructed through 3D deconvolution.
Enjoy, and feel free to contact us for any information request, like the full PSF, 3 more samples or longer image sequences.
Not seeing a result you expected?
Learn how you can add new datasets to our index.
Facebook
TwitterThis dataset contains images in png format for RSNA Screening Mammography Breast Cancer Detection competition. All the images have a 768x768 size and are organized in two folders (one for each label). This allows inferring the labels from the folders (which might be helpful when using keras image_dataset_from_directory for example). - The processing of the DICOM files was done in this notebook by (Radek Osmulski ): https://www.kaggle.com/code/radek1/how-to-process-dicom-images-to-pngs?scriptVersionId=113529850 - The contribution made in this dataset is to organize the files in a way that enables label inferring from the directory name.
🎉If this dataset helps you in your work don't hesitate to upvote 🎉