4 datasets found
  1. ViTPose with retraining on infant data

    • zenodo.org
    bin
    Updated Apr 23, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Lennart Jahn; Lennart Jahn (2025). ViTPose with retraining on infant data [Dataset]. http://doi.org/10.5281/zenodo.14833182
    Explore at:
    binAvailable download formats
    Dataset updated
    Apr 23, 2025
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Lennart Jahn; Lennart Jahn
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Retrained ViTPose Models

    This dataset contains five retrained ViTPose-large models.
    They were generated for
    Jahn, L., Flügge, S., Zhang, D. et al. Comparison of marker-less 2D image-based methods for infant pose estimation. Sci Rep 15, 12148 (2025). https://doi.org/10.1038/s41598-025-96206-0
    Please note that the performance of these models will depend on the specificity
    of your dataset. Please see the publication for more details on this matter.

    To run a model, install mmpose (https://github.com/open-mmlab/mmpose/tree/main)
    and configure everything as if you wanted to use ViTPose-large
    (https://mmpose.readthedocs.io/en/latest/model_zoo/body_2d_keypoint.html#
    topdown-heatmap-vitpose-on-coco), then swap out the .pth file for one provided
    here and change the filename accordingly.

  2. h

    mpii-human-pose-captions

    • huggingface.co
    Updated Apr 23, 2025
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Muhammad Saif Ullah Khan (2025). mpii-human-pose-captions [Dataset]. http://doi.org/10.57967/hf/1876
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Apr 23, 2025
    Authors
    Muhammad Saif Ullah Khan
    License

    https://choosealicense.com/licenses/bsd-2-clause/https://choosealicense.com/licenses/bsd-2-clause/

    Description

    Dataset Card for MPII Human Pose Descriptions

      Dataset Summary
    

    The MPII Human Pose Descriptions dataset extends the widely-used MPII Human Pose Dataset with rich textual annotations. These annotations are generated by various state-of-the-art language models (LLMs) and include detailed descriptions of the activities being performed, the count of people present, and their specific poses. The dataset consists of the same image splits as provided in MMPose, with 14644… See the full description on the dataset page: https://huggingface.co/datasets/saifkhichi96/mpii-human-pose-captions.

  3. P

    COCO-WholeBody Dataset

    • paperswithcode.com
    Updated Feb 19, 2021
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Sheng Jin; Lumin Xu; Jin Xu; Can Wang; Wentao Liu; Chen Qian; Wanli Ouyang; Ping Luo (2021). COCO-WholeBody Dataset [Dataset]. https://paperswithcode.com/dataset/coco-wholebody
    Explore at:
    Dataset updated
    Feb 19, 2021
    Authors
    Sheng Jin; Lumin Xu; Jin Xu; Can Wang; Wentao Liu; Chen Qian; Wanli Ouyang; Ping Luo
    Description

    COCO-WholeBody is an extension of COCO dataset with whole-body annotations. There are 4 types of bounding boxes (person box, face box, left-hand box, and right-hand box) and 133 keypoints (17 for body, 6 for feet, 68 for face and 42 for hands) annotations for each person in the image.

  4. Data from: Paving the Way Towards Kinematic Assessment Using Monocular...

    • zenodo.org
    • portaldelaciencia.uva.es
    • +1more
    zip
    Updated Mar 31, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Mario Medrano-Paredes; Mario Medrano-Paredes; Carmen Fernández-González; Carmen Fernández-González; Francisco-Javier Díaz-Pernas; Francisco-Javier Díaz-Pernas; Hichem Saoudi; Hichem Saoudi; Javier González-Alonso; Javier González-Alonso; Mario Martínez-Zarzuela; Mario Martínez-Zarzuela (2025). Paving the Way Towards Kinematic Assessment Using Monocular Video: A Benchmark of State-of-the-Art Deep-Learning-Based 3D Human Pose Estimators Against Inertial Sensors in Daily Living Activities [Dataset]. http://doi.org/10.5281/zenodo.15088423
    Explore at:
    zipAvailable download formats
    Dataset updated
    Mar 31, 2025
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Mario Medrano-Paredes; Mario Medrano-Paredes; Carmen Fernández-González; Carmen Fernández-González; Francisco-Javier Díaz-Pernas; Francisco-Javier Díaz-Pernas; Hichem Saoudi; Hichem Saoudi; Javier González-Alonso; Javier González-Alonso; Mario Martínez-Zarzuela; Mario Martínez-Zarzuela
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Time period covered
    Mar 26, 2025
    Description

    Advances in machine learning and wearable sensors offer new opportunities for capturing and analyzing human movement outside specialized laboratories. Accurately tracking and evaluating human movement under real-world conditions is essential for telemedicine, sports science, and rehabilitation. This work introduces a comprehensive benchmark comparing deep learning monocular video-based human pose estimation models with inertial measurement unit (IMU)-driven methods, leveraging VIDIMU dataset containing a total of 13 clinically relevant activities which were captured using both commodity video cameras and 5 IMUs. Joint angles derived from state-of-the-art deep learning frameworks (MotionAGFormer, MotionBERT, MMPose 2D-to-3D pose lifting, and NVIDIA BodyTrack included in Maxine-AR-SDK) were evaluated against joint angles computed from IMU data using OpenSim inverse kinematic methods. A graphical comparison of the angles estimated by each model shows the overall performance for each activity.

    The results, which also contains the evaluation of multiple metrics (RMSE, NMRSE, MAE, correlation and coefficient of determination) in table and plot format, highlight key trade-offs between video- and sensor-based approaches including costs, accessibility and precision across different daily life activities. This work establishes valuable guidelines for researchers and clinicians seeking to develop robust, cost-effective, and user-friendly solutions for telehealth and remote patient monitoring solutions, ultimately bridging the gap between AI-driven motion capture and accessible healthcare applications.

  5. Not seeing a result you expected?
    Learn how you can add new datasets to our index.

Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
Lennart Jahn; Lennart Jahn (2025). ViTPose with retraining on infant data [Dataset]. http://doi.org/10.5281/zenodo.14833182
Organization logo

ViTPose with retraining on infant data

Explore at:
binAvailable download formats
Dataset updated
Apr 23, 2025
Dataset provided by
Zenodohttp://zenodo.org/
Authors
Lennart Jahn; Lennart Jahn
License

Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically

Description

Retrained ViTPose Models

This dataset contains five retrained ViTPose-large models.
They were generated for
Jahn, L., Flügge, S., Zhang, D. et al. Comparison of marker-less 2D image-based methods for infant pose estimation. Sci Rep 15, 12148 (2025). https://doi.org/10.1038/s41598-025-96206-0
Please note that the performance of these models will depend on the specificity
of your dataset. Please see the publication for more details on this matter.

To run a model, install mmpose (https://github.com/open-mmlab/mmpose/tree/main)
and configure everything as if you wanted to use ViTPose-large
(https://mmpose.readthedocs.io/en/latest/model_zoo/body_2d_keypoint.html#
topdown-heatmap-vitpose-on-coco), then swap out the .pth file for one provided
here and change the filename accordingly.

Search
Clear search
Close search
Google apps
Main menu