12 datasets found
  1. Doodleverse/Segmentation Zoo Res-UNet models for 4-class (water, whitewater,...

    • zenodo.org
    bin, json, png, txt
    Updated Jul 16, 2024
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Daniel Buscombe; Daniel Buscombe (2024). Doodleverse/Segmentation Zoo Res-UNet models for 4-class (water, whitewater, sediment and other) segmentation of Sentinel-2 and Landsat-7/8 3-band (RGB) images of coasts. [Dataset]. http://doi.org/10.5281/zenodo.6950472
    Explore at:
    json, txt, png, binAvailable download formats
    Dataset updated
    Jul 16, 2024
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Daniel Buscombe; Daniel Buscombe
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Doodleverse/Segmentation Zoo Res-UNet models for 4-class (water, whitewater, sediment and other) segmentation of Sentinel-2 and Landsat-7/8 3-band (RGB) images of coasts.

    These Residual-UNet model data are based on RGB (red, green, and blue) images of coasts and associated labels.

    Models have been created using Segmentation Gym* using the following dataset**: https://doi.org/10.5281/zenodo.7335647

    Classes: {0=water, 1=whitewater, 2=sediment, 3=other}

    File descriptions

    For each model, there are 5 files with the same root name:

    1. '.json' config file: this is the file that was used by Segmentation Gym* to create the weights file. It contains instructions for how to make the model and the data it used, as well as instructions for how to use the model for prediction. It is a handy wee thing and mastering it means mastering the entire Doodleverse.

    2. '.h5' weights file: this is the file that was created by the Segmentation Gym* function `train_model.py`. It contains the trained model's parameter weights. It can called by the Segmentation Gym* function `seg_images_in_folder.py`. Models may be ensembled.

    3. '_modelcard.json' model card file: this is a json file containing fields that collectively describe the model origins, training choices, and dataset that the model is based upon. There is some redundancy between this file and the `config` file (described above) that contains the instructions for the model training and implementation. The model card file is not used by the program but is important metadata so it is important to keep with the other files that collectively make the model and is such is considered part of the model

    4. '_model_history.npz' model training history file: this numpy archive file contains numpy arrays describing the training and validation losses and metrics. It is created by the Segmentation Gym function `train_model.py`

    5. '.png' model training loss and mean IoU plot: this png file contains plots of training and validation losses and mean IoU scores during model training. A subset of data inside the .npz file. It is created by the Segmentation Gym function `train_model.py`

    Additionally, BEST_MODEL.txt contains the name of the model with the best validation loss and mean IoU

    References

    *Segmentation Gym: Buscombe, D., & Goldstein, E. B. (2022). A reproducible and reusable pipeline for segmentation of geoscientific imagery. Earth and Space Science, 9, e2022EA002332. https://doi.org/10.1029/2022EA002332 See: https://github.com/Doodleverse/segmentation_gym

    ** Buscombe, Daniel, Goldstein, Evan, Bernier, Julie, Bosse, Stephen, Colacicco, Rosa, Corak, Nick, Fitzpatrick, Sharon, del Jesús González Guillén, Anais, Ku, Venus, Paprocki, Julie, Platt, Lindsay, Steele, Bethel, Wright, Kyle, & Yasin, Brandon. (2022). Images and 4-class labels for semantic segmentation of Sentinel-2 and Landsat RGB satellite images of coasts (water, whitewater, sediment, other) (v1.0) [Data set]. Zenodo. https://doi.org/10.5281/zenodo.7335647

  2. f

    Summary of the false segmentation rates obtained with the proposed method.

    • plos.figshare.com
    xls
    Updated Jun 1, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Yuliang Wang; Zaicheng Zhang; Huimin Wang; Shusheng Bi (2023). Summary of the false segmentation rates obtained with the proposed method. [Dataset]. http://doi.org/10.1371/journal.pone.0130178.t001
    Explore at:
    xlsAvailable download formats
    Dataset updated
    Jun 1, 2023
    Dataset provided by
    PLOS ONE
    Authors
    Yuliang Wang; Zaicheng Zhang; Huimin Wang; Shusheng Bi
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    The values in parenthesis are corresponding percentages.Among the three cell lines, MCF 10A cells have the lowest overall false tracking rate.£ Oversegmentation: the number of detected cells is more than their actual number in a given area;† Undersegmentation: the number of detected cells is less than their actual number in a given area;‡ Over detection: the debris or artifacts present in the field of view are falsely detected as cells.Summary of the false segmentation rates obtained with the proposed method.

  3. f

    Summary for Test dataset.

    • plos.figshare.com
    xls
    Updated Jun 1, 2023
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Dongjin Han; Hackjoon Shim; Byunghwan Jeon; Yeonggul Jang; Youngtaek Hong; Sunghee Jung; Seongmin Ha; Hyuk-Jae Chang (2023). Summary for Test dataset. [Dataset]. http://doi.org/10.1371/journal.pone.0156837.t006
    Explore at:
    xlsAvailable download formats
    Dataset updated
    Jun 1, 2023
    Dataset provided by
    PLOS ONE
    Authors
    Dongjin Han; Hackjoon Shim; Byunghwan Jeon; Yeonggul Jang; Youngtaek Hong; Sunghee Jung; Seongmin Ha; Hyuk-Jae Chang
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Summary for Test dataset.

  4. Auditory Attention Detection Dataset KULeuven

    • zenodo.org
    bin, txt, zip
    Updated Jun 11, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Neetha Das; Neetha Das; Tom Francart; Tom Francart; Alexander Bertrand; Alexander Bertrand (2025). Auditory Attention Detection Dataset KULeuven [Dataset]. http://doi.org/10.5281/zenodo.4004271
    Explore at:
    bin, zip, txtAvailable download formats
    Dataset updated
    Jun 11, 2025
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Neetha Das; Neetha Das; Tom Francart; Tom Francart; Alexander Bertrand; Alexander Bertrand
    License

    Attribution-NonCommercial-ShareAlike 4.0 (CC BY-NC-SA 4.0)https://creativecommons.org/licenses/by-nc-sa/4.0/
    License information was derived automatically

    Time period covered
    Jan 2024
    Description
    ***************************************

    Please cite the original paper where this data set was presented:

    Biesmans, W., Das, N., Francart, T., & Bertrand, A. (2016). Auditory-inspired speech envelope extraction methods for improved EEG-based auditory attention detection in a cocktail party scenario. IEEE Transactions on Neural Systems and Rehabilitation Engineering, 25(5), 402-412.

    ***************************************

    IMPORTANT MESSAGE FROM THE AUTHORS (January 2024):

    We have observed that this dataset is widely used in research, establishing it as a standard benchmark for evaluating novel decoding strategies in auditory attention decoding (AAD). We emphasize the critical importance of rigorous cross-validation in such studies. In particular, researchers should be aware of two common and significant validation pitfalls:

    1. Trial fingerprints: Avoid splitting data from the same experimental trial into training and testing segments. Classifiers can detect whether a test segment belongs to a specific trial if other, non-overlapping segments from that trial are in the training set.

    2. Gaze bias: AAD algorithms that directly classify EEG signals (often referred to as spatial AAD or SpAAD) without explicitly correlating the decoder output signal to the speech stimulus, should not be evaluated on this dataset, as it is affected by gaze-related bias. Instead, use the gaze-controlled dataset of Rotaru et al. available at https://zenodo.org/records/11058711

    Further details on both issues are provided below.

    In the original study by Biesmans et al., which produced this dataset, linear correlation-based methods were employed, and a straightforward random cross-validation sufficed. However, with the recent surge in the application of machine learning techniques, particularly deep neural networks, in tackling the AAD challenge, a more stringent cross-validation approach becomes imperative. Deep networks are susceptible to overfitting to trial-specific patterns in EEG data, even from very brief segments (less than 1 second), leading to the ability to identify the trial source. Since a subject typically maintains attention to the same speaker throughout a trial, having knowledge of the trial effectively results in a perfect attention decoding.

    We observed that many research papers utilizing our dataset still adhere to the basic random cross-validation method, neglecting the separation of trials into training and testing sets. Consequently, these studies frequently report remarkably high AAD accuracies when using extremely short EEG segments (one or a few seconds). Nevertheless, research has demonstrated that such an approach yields inaccurate and excessively optimistic outcomes. Accuracies often plummet significantly, sometimes even falling below chance levels, when employing a proper cross-validation where this trial bias is removed (e.g., leave-one-trial-out, leave-one-story-out, or leave-one-subject-out cross-validation).

    This overfitting effect is described in: Corentin Puffay et al., "Relating EEG to continuous speech using deep neural networks: a review", Journal of Neural Engineering 20, 041003, 2023 DOI:10.1088/1741-2552/ace73f

    Moreover, it's important to note that AAD strategies which directly classify an EEG snippet, rather than explicitly computing a correlation between the decoder output and the corresponding speech envelopes, may be susceptible to an eye-gaze bias. This bias refers to the tendency of the subject to subtly and often unknowingly direct their gaze towards the attended speaker. Given that EEG equipment can inadvertently capture these gaze patterns, it becomes possible to leverage this gaze information, whether intentionally or unintentionally, to enhance AAD performance. It's crucial to highlight that there is a relatively strong eye gaze bias in this dataset (such gaze bias is present in the majority of public AAD datasets).

    This eye-gaze overfitting effects is discussed in: Rotaru et al. "What are we really decoding? Unveiling biases in EEG-based decoding of the spatial focus of auditory attention", Journal of Neural Engineering, vol. 21, 016017, DOI: https://doi.org/10.1088/1741-2552/ad2214. Also available on bioRxiv: https://doi.org/10.1101/2023.07.13.548824

    To test whether your method is not using gaze as a shortcut, use the Rotaru et al. data set available at https://zenodo.org/records/11058711

    ***************************************

    Explanation about the data set:

    This work was done at ExpORL, Dept. Neurosciences, KULeuven and Dept. Electrical Engineering (ESAT), KULeuven.
    This dataset contains EEG data collected from 16 normal-hearing subjects. EEG recordings were made in a soundproof, electromagnetically shielded room at ExpORL, KULeuven. The BioSemi ActiveTwo system was used to record 64-channel EEG signals at 8196 Hz sample rate. The audio signals, low pass filtered at 4 kHz, were administered to each subject at 60 dBA through a pair of insert phones (Etymotic ER3A). The experiments were conducted using the APEX 3 program developed at ExpORL [1].

    Four Dutch short stories [2], narrated by different male speakers, were used as stimuli. All silences longer than 500 ms in the audio files were truncated to 500 ms. Each story was divided into two parts of approximately 6 minutes each. During a presentation, the subjects were presented with the six-minutes part of two (out of four) stories played simultaneously. There were two stimulus conditions, i.e., `HRTF' or `dry' (dichotic). An experiment here is defined as a sequence of 4 presentations, 2 for each stimulus condition and ear of stimulation, with questions asked to the subject after each presentation. All subjects sat through three experiments within a single recording session. An example for the design of an experiment is shown in Table 1 in [3]. The first two experiments included four presentations each. During a presentation, the subjects were instructed to listen to the story in one ear, while ignoring the story in the other ear. After each presentation, the subjects were presented with a set of multiple-choice questions about the story they were listening to in order to help them stay motivated to focus on the task. In the next presentation, the subjects were presented with the next part of the two stories. This time they were instructed to attend to their other ear. In this manner, one experiment involved four presentations in which the subjects listened to a total of two stories, switching attended ear between presentations. The second experiment had the same design but with two other stories. Note that the Table was different for each subject or recording session, i.e., each of the elements in the table were permuted between different recording sessions to ensure that the different conditions (stimulus condition and the attended ear) were equally distributed over the four presentations. Finally, the third experiment included a set of presentations where the first two minutes of the story parts from the first experiment, i.e., a total of four shorter presentations, were repeated three times, to build a set of recordings of repetitions. Thus, a total of approximately 72 minutes of EEG was recorded per subject.

    We refer to EEG recorded from each presentation as a trial. For each subject, we recorded 20 trials - 4 from the first experiment, 4 from the second experiment, and 12 from the third experiment (first 2 minutes of the 4 presentations from experiment 1 X 3 repetitions). The EEG data is stored in subject specific mat files of the format 'Sx', 'x' referring to the subject number. The audio data is stored as wav files in the folder 'stimuli'. Please note that the stories were not of equal lengths, and the subjects were allowed to finish listening to a story, even in cases where the competing story was over. Therefore, for each trial, we suggest referring to the length of the EEG recordings to truncate the ends of the corresponding audio data. This will ensure that the processed data (EEG and audio) contains only competing talker scenarios. Each trial was high-pass filtered (0.5 Hz cut off) and downsampled from the recorded sampling rate of 8192 Hz to 128 Hz. Artifacts were removed using the MWF-filtering method in [4]. Please get in touch with the team (of Prof. Alexander Bertrand or Prof. Tom Francart) if you wish to obtain the raw EEG data (without the mentioned high-pass filtering and artifact removal).

    Each trial (trial*.mat) contains the following information:

    RawData.Channels : channel numbers (1 to 64)
    RawData.EegData : EEG data (samples X channels)
    FileHeader.SampleRate : Sampling frequency of the saved data
    TrialID : a number between 1 to 20, showing the trial number
    attended_ear : the direction of attention of the subject. 'L' for left, 'R' for right
    stimuli : cell array with stimuli{1} and stimuli{2} indicating the name of audio files presented in the left ear and the right ear of the subject respectively
    condition : stimulus presentation condition. 'HRTF' - stimuli were filtered with HRTF functions to simulate audio from 90 degrees to the left and 90 degrees to the right of the speaker, 'dry' - a dichotic presentation in which there was one story track each presented separately via the left and the right earphones.
    experiment : the number of the experiment (1,2 or 3)
    part : part of the story track being presented (can be 1 to 4 for experiments 1 and 2, and 1 to 12 for experiment

  5. f

    The overall performance of the innervation zone detection algorithm.

    • plos.figshare.com
    xls
    Updated Jun 2, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Hamid Reza Marateb; Morteza Farahi; Monica Rojas; Miguel Angel Mañanas; Dario Farina (2023). The overall performance of the innervation zone detection algorithm. [Dataset]. http://doi.org/10.1371/journal.pone.0167954.t003
    Explore at:
    xlsAvailable download formats
    Dataset updated
    Jun 2, 2023
    Dataset provided by
    PLOS ONE
    Authors
    Hamid Reza Marateb; Morteza Farahi; Monica Rojas; Miguel Angel Mañanas; Dario Farina
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    The overall performance of the innervation zone detection algorithm.

  6. The running time of the innervation zone detection algorithm on each 60-ms...

    • plos.figshare.com
    xls
    Updated Jun 4, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Hamid Reza Marateb; Morteza Farahi; Monica Rojas; Miguel Angel Mañanas; Dario Farina (2023). The running time of the innervation zone detection algorithm on each 60-ms sEMG frame in MEAN±SD. [Dataset]. http://doi.org/10.1371/journal.pone.0167954.t006
    Explore at:
    xlsAvailable download formats
    Dataset updated
    Jun 4, 2023
    Dataset provided by
    PLOShttp://plos.org/
    Authors
    Hamid Reza Marateb; Morteza Farahi; Monica Rojas; Miguel Angel Mañanas; Dario Farina
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    The running time of the innervation zone detection algorithm on each 60-ms sEMG frame in MEAN±SD.

  7. f

    The performance of the proposed IZ detection algorithm on the simulated...

    • figshare.com
    xls
    Updated Jun 4, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Hamid Reza Marateb; Morteza Farahi; Monica Rojas; Miguel Angel Mañanas; Dario Farina (2023). The performance of the proposed IZ detection algorithm on the simulated dataset (MEAN±SD). [Dataset]. http://doi.org/10.1371/journal.pone.0167954.t002
    Explore at:
    xlsAvailable download formats
    Dataset updated
    Jun 4, 2023
    Dataset provided by
    PLOS ONE
    Authors
    Hamid Reza Marateb; Morteza Farahi; Monica Rojas; Miguel Angel Mañanas; Dario Farina
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    The performance of the proposed IZ detection algorithm on the simulated dataset (MEAN±SD).

  8. f

    Additional analysis of the proposed IZ detection algorithms.

    • plos.figshare.com
    xls
    Updated Jun 2, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Hamid Reza Marateb; Morteza Farahi; Monica Rojas; Miguel Angel Mañanas; Dario Farina (2023). Additional analysis of the proposed IZ detection algorithms. [Dataset]. http://doi.org/10.1371/journal.pone.0167954.t007
    Explore at:
    xlsAvailable download formats
    Dataset updated
    Jun 2, 2023
    Dataset provided by
    PLOS ONE
    Authors
    Hamid Reza Marateb; Morteza Farahi; Monica Rojas; Miguel Angel Mañanas; Dario Farina
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Additional analysis of the proposed IZ detection algorithms.

  9. f

    Confusion matrix for the best detection model; each element is shown in...

    • plos.figshare.com
    xls
    Updated Jun 9, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Arindam Dutta; Elena Steiner; Jeffrey Proulx; Visar Berisha; Daniel W. Bliss; Scott Poole; Steven Corman (2023). Confusion matrix for the best detection model; each element is shown in terms of number of 15 seconds segments. [Dataset]. http://doi.org/10.1371/journal.pone.0250301.t001
    Explore at:
    xlsAvailable download formats
    Dataset updated
    Jun 9, 2023
    Dataset provided by
    PLOS ONE
    Authors
    Arindam Dutta; Elena Steiner; Jeffrey Proulx; Visar Berisha; Daniel W. Bliss; Scott Poole; Steven Corman
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Confusion matrix for the best detection model; each element is shown in terms of number of 15 seconds segments.

  10. f

    Segment variables and their imagery sources.

    • plos.figshare.com
    • figshare.com
    xls
    Updated Jun 13, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Celia J. Hampton-Miller; Peter N. Neitlich; David K. Swanson (2023). Segment variables and their imagery sources. [Dataset]. http://doi.org/10.1371/journal.pone.0273893.t003
    Explore at:
    xlsAvailable download formats
    Dataset updated
    Jun 13, 2023
    Dataset provided by
    PLOS ONE
    Authors
    Celia J. Hampton-Miller; Peter N. Neitlich; David K. Swanson
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Segment variables and their imagery sources.

  11. The absolute and relative muscle fiber conduction velocity error in m/s, and...

    • plos.figshare.com
    xls
    Updated Jun 5, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Hamid Reza Marateb; Morteza Farahi; Monica Rojas; Miguel Angel Mañanas; Dario Farina (2023). The absolute and relative muscle fiber conduction velocity error in m/s, and percentage, respectively of the proposed algorithm (MEAN±SD). [Dataset]. http://doi.org/10.1371/journal.pone.0167954.t005
    Explore at:
    xlsAvailable download formats
    Dataset updated
    Jun 5, 2023
    Dataset provided by
    PLOShttp://plos.org/
    Authors
    Hamid Reza Marateb; Morteza Farahi; Monica Rojas; Miguel Angel Mañanas; Dario Farina
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    The absolute and relative muscle fiber conduction velocity error in m/s, and percentage, respectively of the proposed algorithm (MEAN±SD).

  12. The spatial distribution parameters of the simulated EMG frames (MEAN±SD).

    • plos.figshare.com
    xls
    Updated Jun 2, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Hamid Reza Marateb; Morteza Farahi; Monica Rojas; Miguel Angel Mañanas; Dario Farina (2023). The spatial distribution parameters of the simulated EMG frames (MEAN±SD). [Dataset]. http://doi.org/10.1371/journal.pone.0167954.t001
    Explore at:
    xlsAvailable download formats
    Dataset updated
    Jun 2, 2023
    Dataset provided by
    PLOShttp://plos.org/
    Authors
    Hamid Reza Marateb; Morteza Farahi; Monica Rojas; Miguel Angel Mañanas; Dario Farina
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    The spatial distribution parameters of the simulated EMG frames (MEAN±SD).

  13. Not seeing a result you expected?
    Learn how you can add new datasets to our index.

Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
Daniel Buscombe; Daniel Buscombe (2024). Doodleverse/Segmentation Zoo Res-UNet models for 4-class (water, whitewater, sediment and other) segmentation of Sentinel-2 and Landsat-7/8 3-band (RGB) images of coasts. [Dataset]. http://doi.org/10.5281/zenodo.6950472
Organization logo

Doodleverse/Segmentation Zoo Res-UNet models for 4-class (water, whitewater, sediment and other) segmentation of Sentinel-2 and Landsat-7/8 3-band (RGB) images of coasts.

Explore at:
json, txt, png, binAvailable download formats
Dataset updated
Jul 16, 2024
Dataset provided by
Zenodohttp://zenodo.org/
Authors
Daniel Buscombe; Daniel Buscombe
License

Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically

Description

Doodleverse/Segmentation Zoo Res-UNet models for 4-class (water, whitewater, sediment and other) segmentation of Sentinel-2 and Landsat-7/8 3-band (RGB) images of coasts.

These Residual-UNet model data are based on RGB (red, green, and blue) images of coasts and associated labels.

Models have been created using Segmentation Gym* using the following dataset**: https://doi.org/10.5281/zenodo.7335647

Classes: {0=water, 1=whitewater, 2=sediment, 3=other}

File descriptions

For each model, there are 5 files with the same root name:

1. '.json' config file: this is the file that was used by Segmentation Gym* to create the weights file. It contains instructions for how to make the model and the data it used, as well as instructions for how to use the model for prediction. It is a handy wee thing and mastering it means mastering the entire Doodleverse.

2. '.h5' weights file: this is the file that was created by the Segmentation Gym* function `train_model.py`. It contains the trained model's parameter weights. It can called by the Segmentation Gym* function `seg_images_in_folder.py`. Models may be ensembled.

3. '_modelcard.json' model card file: this is a json file containing fields that collectively describe the model origins, training choices, and dataset that the model is based upon. There is some redundancy between this file and the `config` file (described above) that contains the instructions for the model training and implementation. The model card file is not used by the program but is important metadata so it is important to keep with the other files that collectively make the model and is such is considered part of the model

4. '_model_history.npz' model training history file: this numpy archive file contains numpy arrays describing the training and validation losses and metrics. It is created by the Segmentation Gym function `train_model.py`

5. '.png' model training loss and mean IoU plot: this png file contains plots of training and validation losses and mean IoU scores during model training. A subset of data inside the .npz file. It is created by the Segmentation Gym function `train_model.py`

Additionally, BEST_MODEL.txt contains the name of the model with the best validation loss and mean IoU

References

*Segmentation Gym: Buscombe, D., & Goldstein, E. B. (2022). A reproducible and reusable pipeline for segmentation of geoscientific imagery. Earth and Space Science, 9, e2022EA002332. https://doi.org/10.1029/2022EA002332 See: https://github.com/Doodleverse/segmentation_gym

** Buscombe, Daniel, Goldstein, Evan, Bernier, Julie, Bosse, Stephen, Colacicco, Rosa, Corak, Nick, Fitzpatrick, Sharon, del Jesús González Guillén, Anais, Ku, Venus, Paprocki, Julie, Platt, Lindsay, Steele, Bethel, Wright, Kyle, & Yasin, Brandon. (2022). Images and 4-class labels for semantic segmentation of Sentinel-2 and Landsat RGB satellite images of coasts (water, whitewater, sediment, other) (v1.0) [Data set]. Zenodo. https://doi.org/10.5281/zenodo.7335647

Search
Clear search
Close search
Google apps
Main menu