13 datasets found
  1. f

    Comparison of random prediction simulation, logistic regression, random...

    • figshare.com
    xls
    Updated Jul 7, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Samantha Kanny; Grisha Post; Patricia Carbajales-Dale; William Cummings; Janet Evatt; Windsor Westbrook Sherrill (2025). Comparison of random prediction simulation, logistic regression, random forest, xgboost, balanced random forest, and balanced XGBoost models. [Dataset]. http://doi.org/10.1371/journal.pdig.0000930.t005
    Explore at:
    xlsAvailable download formats
    Dataset updated
    Jul 7, 2025
    Dataset provided by
    PLOS Digital Health
    Authors
    Samantha Kanny; Grisha Post; Patricia Carbajales-Dale; William Cummings; Janet Evatt; Windsor Westbrook Sherrill
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Comparison of random prediction simulation, logistic regression, random forest, xgboost, balanced random forest, and balanced XGBoost models.

  2. m

    Physiological Signals Based Mental Fatigue Analysis & Recognition: MEFAR

    • data.mendeley.com
    Updated May 23, 2023
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    SEYMA DERDIYOK (2023). Physiological Signals Based Mental Fatigue Analysis & Recognition: MEFAR [Dataset]. http://doi.org/10.17632/z3g26tphnv.3
    Explore at:
    Dataset updated
    May 23, 2023
    Authors
    SEYMA DERDIYOK
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    The dataset was created specifically for analyzing mental fatigue based on physiological signals. It includes the following signals:

    EEG (Electroencephalography): Measures the electrical activity of the brain. BVP (Blood Volume Pulse): Captures changes in blood volume in peripheral blood vessels. EDA (Electrodermal Activity): Tracks changes in the electrical conductance of the skin. HR (Heart Rate): Records the number of heartbeats per minute. TEMP (Temperature): Monitors the body temperature of the participants. ACC (Acceleration): Measures the acceleration experienced by the participants. The measurements were collected from 23 participants in both the morning and evening sessions. To evaluate the participants' mental fatigue levels, the Chalder Fatigue Scale was utilized. This scale assigns a score to each participant based on their responses, with a score of 12 or higher indicating a positive mental fatigue condition.

    The dataset includes both raw and processed data. Raw data refers to the original recorded signals, while processed data typically involves signal preprocessing techniques such as filtering, artifact removal, and feature extraction.The processed data includes processed data based on different sampling frequencies (1 Hz, 32 Hz, and 64 Hz). These operations include downsampling, midsampling, and upsampling. The processed datasets are named MEFAR_DOWN, MEFAR_MID, and MEFAR_UP, corresponding to the processed data with downsampling, midsampling, and upsampling, respectively.This way, it was possible to analyze and compare data with different sampling frequencies.

    The dataset has been used for training deep learning and transfer learning models, which suggests that it may be suitable for developing machine learning algorithms for mental fatigue detection based on physiological signals.

    Additionally, demographic information about the participants is available in an Excel file called "general_info." It is important to ensure that the participants' anonymity and privacy are maintained in accordance with ethical guidelines.

    By sharing this dataset, researchers interested in mental fatigue analysis can utilize it for further investigations, algorithm development, and validation.

  3. f

    Pre-program participant results using HED evaluation measures.

    • plos.figshare.com
    xls
    Updated Jul 7, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Samantha Kanny; Grisha Post; Patricia Carbajales-Dale; William Cummings; Janet Evatt; Windsor Westbrook Sherrill (2025). Pre-program participant results using HED evaluation measures. [Dataset]. http://doi.org/10.1371/journal.pdig.0000930.t002
    Explore at:
    xlsAvailable download formats
    Dataset updated
    Jul 7, 2025
    Dataset provided by
    PLOS Digital Health
    Authors
    Samantha Kanny; Grisha Post; Patricia Carbajales-Dale; William Cummings; Janet Evatt; Windsor Westbrook Sherrill
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Pre-program participant results using HED evaluation measures.

  4. f

    Odds ratio for evaluation measures (Odds Ratio Estimate, P-Value & 95%...

    • figshare.com
    • plos.figshare.com
    xls
    Updated Jul 7, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Samantha Kanny; Grisha Post; Patricia Carbajales-Dale; William Cummings; Janet Evatt; Windsor Westbrook Sherrill (2025). Odds ratio for evaluation measures (Odds Ratio Estimate, P-Value & 95% Confidence Interval). [Dataset]. http://doi.org/10.1371/journal.pdig.0000930.t004
    Explore at:
    xlsAvailable download formats
    Dataset updated
    Jul 7, 2025
    Dataset provided by
    PLOS Digital Health
    Authors
    Samantha Kanny; Grisha Post; Patricia Carbajales-Dale; William Cummings; Janet Evatt; Windsor Westbrook Sherrill
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Odds ratio for evaluation measures (Odds Ratio Estimate, P-Value & 95% Confidence Interval).

  5. m

    Data from: Mental issues, internet addiction and quality of life predict...

    • data.mendeley.com
    Updated Jul 31, 2024
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Andras Matuz (2024). Mental issues, internet addiction and quality of life predict burnout among Hungarian teachers: a machine learning analysis [Dataset]. http://doi.org/10.17632/2yy4j7rgvg.2
    Explore at:
    Dataset updated
    Jul 31, 2024
    Authors
    Andras Matuz
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Background: Burnout is usually defined as a state of emotional, physical, and mental exhaustion that affects people in various professions (e.g. physicians, nurses, teachers). The consequences of burnout involve decreased motivation, productivity, and overall diminished well-being. The machine learning-based prediction of burnout has therefore become the focus of recent research. In this study, the aim was to detect burnout using machine learning and to identify its most important predictors in a sample of Hungarian high-school teachers. Methods: The final sample consisted of 1,576 high-school teachers (522 male), who completed a survey including various sociodemographic and health-related questions and psychological questionnaires. Specifically, depression, insomnia, internet habits (e.g. when and why one uses the internet) and problematic internet usage were among the most important predictors tested in this study. Supervised classification algorithms were trained to detect burnout assessed by two well-known burnout questionnaires. Feature selection was conducted using recursive feature elimination. Hyperparameters were tuned via grid search with 5-fold cross-validation. Due to class imbalance, class weights (i.e. cost-sensitive learning), downsampling and a hybrid method (SMOTE-ENN) were applied in separate analyses. The final model evaluation was carried out on a previously unseen holdout test sample. Results: Burnout was detected in 19.7% of the teachers included in the final dataset. The best predictive performance on the holdout test sample was achieved by random forest with class weigths (AUC = .811; balanced accuracy = .745, sensitivity = .765; specificity = .726). The best predictors of burnout were Beck’s Depression Inventory scores, Athen’s Insomnia Scale scores, subscales of the Problematic Internet Use Questionnaire and self-reported current health status. Conclusions: The performances of the algorithms were comparable with previous studies; however, it is important to note that we tested our models on previously unseen holdout samples suggesting higher levels of generalizability. Another remarkable finding is that besides depression and insomnia, other variables such as problematic internet use and time spent online also turned out to be important predictors of burnout.

  6. f

    Comparison of model performance across downsampling and cross-validation...

    • figshare.com
    xls
    Updated Feb 13, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Sasja Maria Pedersen; Nicolai Damslund; Trine Kjær; Kim Rose Olsen (2025). Comparison of model performance across downsampling and cross-validation settings. [Dataset]. http://doi.org/10.1371/journal.pone.0317722.t003
    Explore at:
    xlsAvailable download formats
    Dataset updated
    Feb 13, 2025
    Dataset provided by
    PLOS ONE
    Authors
    Sasja Maria Pedersen; Nicolai Damslund; Trine Kjær; Kim Rose Olsen
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Comparison of model performance across downsampling and cross-validation settings.

  7. f

    Odds ratio for participant demographics characteristics (Odds Ratio...

    • plos.figshare.com
    xls
    Updated Jul 7, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Samantha Kanny; Grisha Post; Patricia Carbajales-Dale; William Cummings; Janet Evatt; Windsor Westbrook Sherrill (2025). Odds ratio for participant demographics characteristics (Odds Ratio Estimate, P-Value & 95% Confidence Interval). [Dataset]. http://doi.org/10.1371/journal.pdig.0000930.t003
    Explore at:
    xlsAvailable download formats
    Dataset updated
    Jul 7, 2025
    Dataset provided by
    PLOS Digital Health
    Authors
    Samantha Kanny; Grisha Post; Patricia Carbajales-Dale; William Cummings; Janet Evatt; Windsor Westbrook Sherrill
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Odds ratio for participant demographics characteristics (Odds Ratio Estimate, P-Value & 95% Confidence Interval).

  8. f

    Descriptive statistics for participant demographic characteristics.

    • plos.figshare.com
    xls
    Updated Jul 7, 2025
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Samantha Kanny; Grisha Post; Patricia Carbajales-Dale; William Cummings; Janet Evatt; Windsor Westbrook Sherrill (2025). Descriptive statistics for participant demographic characteristics. [Dataset]. http://doi.org/10.1371/journal.pdig.0000930.t001
    Explore at:
    xlsAvailable download formats
    Dataset updated
    Jul 7, 2025
    Dataset provided by
    PLOS Digital Health
    Authors
    Samantha Kanny; Grisha Post; Patricia Carbajales-Dale; William Cummings; Janet Evatt; Windsor Westbrook Sherrill
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Descriptive statistics for participant demographic characteristics.

  9. f

    Performance of ML models of DM progression after upsampling minority class...

    • plos.figshare.com
    xls
    Updated Apr 15, 2025
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Claudia C. Colmenares-Mejia; Andrés F. García-Suaza; Paul Rodríguez-Lesmes; Christian Lochmuller; Sara C. Atehortúa; J.E. Camacho-Cogollo; Juan P. Martínez; Juliana Rincón; Yohan R. Céspedes; Esteban Morales-Mendoza; Mario A. Isaza-Ruget (2025). Performance of ML models of DM progression after upsampling minority class and downsampling majority. [Dataset]. http://doi.org/10.1371/journal.pone.0321258.t008
    Explore at:
    xlsAvailable download formats
    Dataset updated
    Apr 15, 2025
    Dataset provided by
    PLOS ONE
    Authors
    Claudia C. Colmenares-Mejia; Andrés F. García-Suaza; Paul Rodríguez-Lesmes; Christian Lochmuller; Sara C. Atehortúa; J.E. Camacho-Cogollo; Juan P. Martínez; Juliana Rincón; Yohan R. Céspedes; Esteban Morales-Mendoza; Mario A. Isaza-Ruget
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Performance of ML models of DM progression after upsampling minority class and downsampling majority.

  10. f

    Data_Sheet_1_Ensemble graph neural network model for classification of major...

    • frontiersin.figshare.com
    pdf
    Updated Jun 2, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Sujitha Venkatapathy; Mikhail Votinov; Lisa Wagels; Sangyun Kim; Munseob Lee; Ute Habel; In-Ho Ra; Han-Gue Jo (2023). Data_Sheet_1_Ensemble graph neural network model for classification of major depressive disorder using whole-brain functional connectivity.PDF [Dataset]. http://doi.org/10.3389/fpsyt.2023.1125339.s001
    Explore at:
    pdfAvailable download formats
    Dataset updated
    Jun 2, 2023
    Dataset provided by
    Frontiers
    Authors
    Sujitha Venkatapathy; Mikhail Votinov; Lisa Wagels; Sangyun Kim; Munseob Lee; Ute Habel; In-Ho Ra; Han-Gue Jo
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Major depressive disorder (MDD) is characterized by impairments in mood and cognitive functioning, and it is a prominent source of global disability and stress. A functional magnetic resonance imaging (fMRI) can aid clinicians in their assessments of individuals for the identification of MDD. Herein, we employ a deep learning approach to the issue of MDD classification. Resting-state fMRI data from 821 individuals with MDD and 765 healthy controls (HCs) is employed for investigation. An ensemble model based on graph neural network (GNN) has been created with the goal of identifying patients with MDD among HCs as well as differentiation between first-episode and recurrent MDDs. The graph convolutional network (GCN), graph attention network (GAT), and GraphSAGE models serve as a base models for the ensemble model that was developed with individual whole-brain functional networks. The ensemble's performance is evaluated using upsampling and downsampling, along with 10-fold cross-validation. The ensemble model achieved an upsampling accuracy of 71.18% and a downsampling accuracy of 70.24% for MDD and HC classification. While comparing first-episode patients with recurrent patients, the upsampling accuracy is 77.78% and the downsampling accuracy is 71.96%. According to the findings of this study, the proposed GNN-based ensemble model achieves a higher level of accuracy and suggests that our model produces can assist healthcare professionals in identifying MDD.

  11. f

    Counting results on RFRB dataset.

    • plos.figshare.com
    xls
    Updated Oct 29, 2024
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Dunlu Lu; Yangxu Wang (2024). Counting results on RFRB dataset. [Dataset]. http://doi.org/10.1371/journal.pone.0307643.t007
    Explore at:
    xlsAvailable download formats
    Dataset updated
    Oct 29, 2024
    Dataset provided by
    PLOS ONE
    Authors
    Dunlu Lu; Yangxu Wang
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    With the development of deep learning technology, object detection has been widely applied in various fields. However, in cross-dataset object detection, conventional deep learning models often face performance degradation issues. This is particularly true in the agricultural field, where there is a multitude of crop types and a complex and variable environment. Existing technologies still face performance bottlenecks when dealing with diverse scenarios. To address these issues, this study proposes a lightweight, cross-dataset enhanced object detection method for the agricultural domain based on YOLOv9, named Multi-Adapt Recognition-YOLOv9 (MAR-YOLOv9). The traditional 32x downsampling Backbone network has been optimized, and a 16x downsampling Backbone network has been innovatively designed. A more streamlined and lightweight Main Neck structure has been introduced, along with innovative methods for feature extraction, up-sampling, and Concat connection. The hybrid connection strategy allows the model to flexibly utilize features from different levels. This solves the issues of increased training time and redundant weights caused by the detection neck and auxiliary branch structures in traditional YOLOv9, enabling MAR-YOLOv9 to maintain high performance while reducing the model’s computational complexity and improving detection speed, making it more suitable for real-time detection tasks. In comparative experiments on four plant datasets, MAR-YOLOv9 improved the mAP@0.5 accuracy by 39.18% compared to seven mainstream object detection algorithms, and by 1.28% compared to the YOLOv9 model. At the same time, the model size was reduced by 9.3%, and the number of model layers was decreased, reducing computational costs and storage requirements. Additionally, MAR-YOLOv9 demonstrated significant advantages in detecting complex agricultural images, providing an efficient, lightweight, and adaptable solution for object detection tasks in the agricultural field. The curated data and code can be accessed at the following link: https://github.com/YangxuWangamI/MAR-YOLOv9.

  12. f

    Quantitative results on WEDU dataset.

    • plos.figshare.com
    xls
    Updated Oct 29, 2024
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Dunlu Lu; Yangxu Wang (2024). Quantitative results on WEDU dataset. [Dataset]. http://doi.org/10.1371/journal.pone.0307643.t002
    Explore at:
    xlsAvailable download formats
    Dataset updated
    Oct 29, 2024
    Dataset provided by
    PLOS ONE
    Authors
    Dunlu Lu; Yangxu Wang
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    With the development of deep learning technology, object detection has been widely applied in various fields. However, in cross-dataset object detection, conventional deep learning models often face performance degradation issues. This is particularly true in the agricultural field, where there is a multitude of crop types and a complex and variable environment. Existing technologies still face performance bottlenecks when dealing with diverse scenarios. To address these issues, this study proposes a lightweight, cross-dataset enhanced object detection method for the agricultural domain based on YOLOv9, named Multi-Adapt Recognition-YOLOv9 (MAR-YOLOv9). The traditional 32x downsampling Backbone network has been optimized, and a 16x downsampling Backbone network has been innovatively designed. A more streamlined and lightweight Main Neck structure has been introduced, along with innovative methods for feature extraction, up-sampling, and Concat connection. The hybrid connection strategy allows the model to flexibly utilize features from different levels. This solves the issues of increased training time and redundant weights caused by the detection neck and auxiliary branch structures in traditional YOLOv9, enabling MAR-YOLOv9 to maintain high performance while reducing the model’s computational complexity and improving detection speed, making it more suitable for real-time detection tasks. In comparative experiments on four plant datasets, MAR-YOLOv9 improved the mAP@0.5 accuracy by 39.18% compared to seven mainstream object detection algorithms, and by 1.28% compared to the YOLOv9 model. At the same time, the model size was reduced by 9.3%, and the number of model layers was decreased, reducing computational costs and storage requirements. Additionally, MAR-YOLOv9 demonstrated significant advantages in detecting complex agricultural images, providing an efficient, lightweight, and adaptable solution for object detection tasks in the agricultural field. The curated data and code can be accessed at the following link: https://github.com/YangxuWangamI/MAR-YOLOv9.

  13. f

    Quantitative results on MTDC-UAV dataset.

    • plos.figshare.com
    xls
    Updated Oct 29, 2024
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Dunlu Lu; Yangxu Wang (2024). Quantitative results on MTDC-UAV dataset. [Dataset]. http://doi.org/10.1371/journal.pone.0307643.t001
    Explore at:
    xlsAvailable download formats
    Dataset updated
    Oct 29, 2024
    Dataset provided by
    PLOS ONE
    Authors
    Dunlu Lu; Yangxu Wang
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    With the development of deep learning technology, object detection has been widely applied in various fields. However, in cross-dataset object detection, conventional deep learning models often face performance degradation issues. This is particularly true in the agricultural field, where there is a multitude of crop types and a complex and variable environment. Existing technologies still face performance bottlenecks when dealing with diverse scenarios. To address these issues, this study proposes a lightweight, cross-dataset enhanced object detection method for the agricultural domain based on YOLOv9, named Multi-Adapt Recognition-YOLOv9 (MAR-YOLOv9). The traditional 32x downsampling Backbone network has been optimized, and a 16x downsampling Backbone network has been innovatively designed. A more streamlined and lightweight Main Neck structure has been introduced, along with innovative methods for feature extraction, up-sampling, and Concat connection. The hybrid connection strategy allows the model to flexibly utilize features from different levels. This solves the issues of increased training time and redundant weights caused by the detection neck and auxiliary branch structures in traditional YOLOv9, enabling MAR-YOLOv9 to maintain high performance while reducing the model’s computational complexity and improving detection speed, making it more suitable for real-time detection tasks. In comparative experiments on four plant datasets, MAR-YOLOv9 improved the mAP@0.5 accuracy by 39.18% compared to seven mainstream object detection algorithms, and by 1.28% compared to the YOLOv9 model. At the same time, the model size was reduced by 9.3%, and the number of model layers was decreased, reducing computational costs and storage requirements. Additionally, MAR-YOLOv9 demonstrated significant advantages in detecting complex agricultural images, providing an efficient, lightweight, and adaptable solution for object detection tasks in the agricultural field. The curated data and code can be accessed at the following link: https://github.com/YangxuWangamI/MAR-YOLOv9.

  14. Not seeing a result you expected?
    Learn how you can add new datasets to our index.

Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
Samantha Kanny; Grisha Post; Patricia Carbajales-Dale; William Cummings; Janet Evatt; Windsor Westbrook Sherrill (2025). Comparison of random prediction simulation, logistic regression, random forest, xgboost, balanced random forest, and balanced XGBoost models. [Dataset]. http://doi.org/10.1371/journal.pdig.0000930.t005

Comparison of random prediction simulation, logistic regression, random forest, xgboost, balanced random forest, and balanced XGBoost models.

Related Article
Explore at:
xlsAvailable download formats
Dataset updated
Jul 7, 2025
Dataset provided by
PLOS Digital Health
Authors
Samantha Kanny; Grisha Post; Patricia Carbajales-Dale; William Cummings; Janet Evatt; Windsor Westbrook Sherrill
License

Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically

Description

Comparison of random prediction simulation, logistic regression, random forest, xgboost, balanced random forest, and balanced XGBoost models.

Search
Clear search
Close search
Google apps
Main menu