Synthesising 2D Video from 3D Motion Data for Machine Learning Applications

Marion Mundt, Henrike Oberlack, Molly Goldacre, Julia Powles, Johannes Funken, Corey Morris, Wolfgang Potthast, Jacqueline Alderson

Publikation: Beitrag in FachzeitschriftZeitschriftenaufsätzeForschungBegutachtung

Abstract

To increase the utility of legacy, gold-standard, three-dimensional (3D) motion capture datasets for computer vision-based machine learning applications, this study proposed and validated a method to synthesise two-dimensional (2D) video image frames from historic 3D motion data. We applied the video-based human pose estimation model OpenPose to real (in situ) and synthesised 2D videos and compared anatomical landmark keypoint outputs, with trivial observed differences (2.11-3.49 mm). We further demonstrated the utility of the method in a downstream machine learning use-case in which we trained and then tested the validity of an artificial neural network (ANN) to estimate ground reaction forces (GRFs) using synthesised and real 2D videos. Training an ANN to estimate GRFs using eight OpenPose keypoints derived from synthesised 2D videos resulted in accurate waveform GRF estimations (r > 0.9; nRMSE < 14%). When compared with using the smaller number of real videos only, accuracy was improved by adding the synthetic views and enlarging the dataset. The results highlight the utility of the developed approach to enlarge small 2D video datasets, or to create 2D video images to accompany 3D motion capture datasets to make them accessible for machine learning applications.

OriginalspracheEnglisch
Aufsatznummer6522
ZeitschriftSensors
Jahrgang22
Ausgabenummer17
ISSN1424-8220
DOIs
PublikationsstatusVeröffentlicht - 29.08.2022

Fingerprint

Untersuchen Sie die Forschungsthemen von „Synthesising 2D Video from 3D Motion Data for Machine Learning Applications“. Zusammen bilden sie einen einzigartigen Fingerprint.

Zitation