{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,31]],"date-time":"2026-01-31T11:22:28Z","timestamp":1769858548181,"version":"3.49.0"},"reference-count":67,"publisher":"Association for Computing Machinery (ACM)","issue":"6","license":[{"start":{"date-parts":[[2024,11,19]],"date-time":"2024-11-19T00:00:00Z","timestamp":1731974400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-sa\/4.0\/"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":["ACM Trans. Graph."],"published-print":{"date-parts":[[2024,12,19]]},"abstract":"<jats:p>This paper introduces ELMO, a real-time upsampling motion capture framework designed for a single LiDAR sensor. Modeled as a conditional autoregressive transformer-based upsampling motion generator, ELMO achieves 60 fps motion capture from a 20 fps LiDAR point cloud sequence. The key feature of ELMO is the coupling of the self-attention mechanism with thoughtfully designed embedding modules for motion and point clouds, significantly elevating the motion quality. To facilitate accurate motion capture, we develop a one-time skeleton calibration model capable of predicting user skeleton off-sets from a single-frame point cloud. Additionally, we introduce a novel data augmentation technique utilizing a LiDAR simulator, which enhances global root tracking to improve environmental understanding. To demonstrate the effectiveness of our method, we compare ELMO with state-of-the-art methods in both image-based and point cloud-based motion capture. We further conduct an ablation study to validate our design principles. ELMO's fast inference time makes it well-suited for real-time applications, exemplified in our demo video featuring live streaming and interactive gaming scenarios. Furthermore, we contribute a high-quality LiDAR-mocap synchronized dataset comprising 20 different subjects performing a range of motions, which can serve as a valuable resource for future research. The dataset and evaluation code are available at https:\/\/movin3d.github.io\/ELMO_SIGASIA2024\/<\/jats:p>","DOI":"10.1145\/3687991","type":"journal-article","created":{"date-parts":[[2024,11,19]],"date-time":"2024-11-19T15:46:04Z","timestamp":1732031164000},"page":"1-14","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":2,"title":["ELMO: Enhanced Real-time LiDAR Motion Capture through Upsampling"],"prefix":"10.1145","volume":"43","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-7567-4339","authenticated-orcid":false,"given":"Deok-Kyeong","family":"Jang","sequence":"first","affiliation":[{"name":"MOVIN Inc., Seoul, South Korea"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4696-3465","authenticated-orcid":false,"given":"Dongseok","family":"Yang","sequence":"additional","affiliation":[{"name":"MOVIN Inc., Seoul, South Korea"}]},{"ORCID":"https:\/\/orcid.org\/0009-0006-1923-2540","authenticated-orcid":false,"given":"Deok-Yun","family":"Jang","sequence":"additional","affiliation":[{"name":"MOVIN Inc., Seoul, South Korea"},{"name":"GIST, Gwangju, South Korea"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2347-149X","authenticated-orcid":false,"given":"Byeoli","family":"Choi","sequence":"additional","affiliation":[{"name":"MOVIN Inc., Seoul, South Korea"},{"name":"KAIST, Daejeon, South Korea"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6604-4709","authenticated-orcid":false,"given":"Sung-Hee","family":"Lee","sequence":"additional","affiliation":[{"name":"KAIST, Daejeon, South Korea"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2447-5716","authenticated-orcid":false,"given":"Donghoon","family":"Shin","sequence":"additional","affiliation":[{"name":"MOVIN Inc., Seoul, South Korea"}]}],"member":"320","published-online":{"date-parts":[[2024,11,19]]},"reference":[{"key":"e_1_2_1_1_1","unstructured":"2009. OptiTrack Motion Capture Systems. https:\/\/www.optitrack.com\/"},{"key":"e_1_2_1_2_1","unstructured":"2011. Xsens Technologies B.V. https:\/\/www.xsens.com\/"},{"key":"e_1_2_1_3_1","unstructured":"2011. Xsens Technologies B.V. https:\/\/www.movella.com\/products\/motion-capture\/xsens-mvn-awinda"},{"key":"e_1_2_1_4_1","unstructured":"2023. Hesai QT128 Specification. https:\/\/www.hesaitech.com\/wp-content\/uploads\/QT128C2X_User_Manual_Q03-en-231210.pdf"},{"key":"e_1_2_1_5_1","doi-asserted-by":"publisher","DOI":"10.1145\/1360612.1360697"},{"key":"e_1_2_1_6_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01290"},{"key":"e_1_2_1_7_1","volume-title":"British Machine Vision Conference.","author":"Amin S.","year":"2013","unstructured":"S. Amin, Mykhaylo Andriluka, Marcus Rohrbach, and Bernt Schiele. 2013. Multiview Pictorial Structures for 3D Human Pose Estimation. In British Machine Vision Conference."},{"key":"e_1_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2011.6126356"},{"key":"e_1_2_1_9_1","volume-title":"Blazepose: On-device real-time body pose tracking. arXiv preprint arXiv:2006.10204","author":"Bazarevsky Valentin","year":"2020","unstructured":"Valentin Bazarevsky, Ivan Grishchenko, Karthik Raveendran, Tyler Zhu, Fan Zhang, and Matthias Grundmann. 2020. Blazepose: On-device real-time body pose tracking. arXiv preprint arXiv:2006.10204 (2020)."},{"key":"e_1_2_1_10_1","volume-title":"Black","author":"Bogo Federica","year":"2016","unstructured":"Federica Bogo, Angjoo Kanazawa, Christoph Lassner, Peter Gehler, Javier Romero, and Michael J. Black. 2016. Keep it SMPL: Automatic Estimation of 3D Human Pose and Shape from a Single Image. arXiv:1607.08128 [cs.CV]"},{"key":"e_1_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.1998.698581"},{"key":"e_1_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2013.464"},{"key":"e_1_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v37i1.25120"},{"key":"e_1_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2021.3098052"},{"key":"e_1_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.1145\/3610548.3618205"},{"key":"e_1_2_1_16_1","volume-title":"MuscleVAE: Model-Based Controllers of Muscle-Actuated Characters. In SIGGRAPH Asia 2023 Conference Papers. 1--11","author":"Feng Yusen","year":"2023","unstructured":"Yusen Feng, Xiyan Xu, and Libin Liu. 2023. MuscleVAE: Model-Based Controllers of Muscle-Actuated Characters. In SIGGRAPH Asia 2023 Conference Papers. 1--11."},{"key":"e_1_2_1_17_1","volume-title":"Proceedings of the 12th ACM SIGGRAPH Conference on Motion, Interaction and Games (Newcastle upon Tyne, United Kingdom) (MIG '19)","author":"Ferstl Ylva","year":"2019","unstructured":"Ylva Ferstl, Michael Neff, and Rachel McDonnell. 2019. Multi-Objective Adversarial Gesture Generation. In Proceedings of the 12th ACM SIGGRAPH Conference on Motion, Interaction and Games (Newcastle upon Tyne, United Kingdom) (MIG '19). Association for Computing Machinery, New York, NY, USA, Article 3, 10 pages."},{"key":"e_1_2_1_18_1","volume-title":"Humans in 4D: Reconstructing and Tracking Humans with Transformers. arXiv preprint arXiv:2305.20091","author":"Goel Shubham","year":"2023","unstructured":"Shubham Goel, Georgios Pavlakos, Jathushan Rajasegaran, Angjoo Kanazawa, and Jitendra Malik. 2023. Humans in 4D: Reconstructing and Tracking Humans with Transformers. arXiv preprint arXiv:2305.20091 (2023)."},{"key":"e_1_2_1_19_1","volume-title":"Weinberger (Eds.)","volume":"27","author":"Goodfellow Ian","year":"2014","unstructured":"Ian Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, and Yoshua Bengio. 2014. Generative Adversarial Nets. In Advances in Neural Information Processing Systems, Z. Ghahramani, M. Welling, C. Cortes, N. Lawrence, and K.Q. Weinberger (Eds.), Vol. 27. Curran Associates, Inc."},{"key":"e_1_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.1145\/3414685.3417836"},{"key":"e_1_2_1_21_1","doi-asserted-by":"crossref","unstructured":"Buzhen Huang Liang Pan Yuan Yang Jingyi Ju and Yangang Wang. 2022. Neural MoCon: Neural Motion Control for Physically Plausible Human Motion Capture. arXiv:2203.14065 [cs.CV]","DOI":"10.1109\/CVPR52688.2022.00631"},{"key":"e_1_2_1_22_1","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1145\/3272127.3275108","article-title":"Deep inertial poser: Learning to reconstruct human pose from sparse inertial measurements in real time","volume":"37","author":"Huang Yinghao","year":"2018","unstructured":"Yinghao Huang, Manuel Kaufmann, Emre Aksan, Michael J Black, Otmar Hilliges, and Gerard Pons-Moll. 2018. Deep inertial poser: Learning to reconstruct human pose from sparse inertial measurements in real time. ACM Transactions on Graphics (TOG) 37, 6 (2018), 1--15.","journal-title":"ACM Transactions on Graphics (TOG)"},{"key":"e_1_2_1_23_1","volume-title":"MOVIN: Real-time Motion Capture using a Single LiDAR. In Computer Graphics Forum","author":"Jang Deok-Kyeong","year":"2023","unstructured":"Deok-Kyeong Jang, Dongseok Yang, Deok-Yun Jang, Byeoli Choi, Taeil Jin, and Sung-Hee Lee. 2023a. MOVIN: Real-time Motion Capture using a Single LiDAR. In Computer Graphics Forum. Wiley Online Library, e14961."},{"key":"e_1_2_1_24_1","doi-asserted-by":"publisher","DOI":"10.1145\/3610548.3618252"},{"key":"e_1_2_1_25_1","volume-title":"Ego-Poser: Robust Real-Time Ego-Body Pose Estimation in Large Scenes. arXiv preprint arXiv:2308.06493","author":"Jiang Jiaxi","year":"2023","unstructured":"Jiaxi Jiang, Paul Streli, Manuel Meier, Andreas Fender, and Christian Holz. 2023. Ego-Poser: Robust Real-Time Ego-Body Pose Estimation in Large Scenes. arXiv preprint arXiv:2308.06493 (2023)."},{"key":"e_1_2_1_26_1","volume-title":"Transformer Inertial Poser: Attention-based Real-time Human Motion Reconstruction from Sparse IMUs. arXiv preprint arXiv:2203.15720","author":"Jiang Yifeng","year":"2022","unstructured":"Yifeng Jiang, Yuting Ye, Deepak Gopinath, Jungdam Won, Alexander W Winkler, and C Karen Liu. 2022. Transformer Inertial Poser: Attention-based Real-time Human Motion Reconstruction from Sparse IMUs. arXiv preprint arXiv:2203.15720 (2022)."},{"key":"e_1_2_1_27_1","volume-title":"Auto-encoding variational bayes. arXiv preprint arXiv:1312.6114","author":"Kingma Diederik P","year":"2013","unstructured":"Diederik P Kingma and Max Welling. 2013. Auto-encoding variational bayes. arXiv preprint arXiv:1312.6114 (2013)."},{"key":"e_1_2_1_28_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00530"},{"key":"e_1_2_1_29_1","doi-asserted-by":"crossref","unstructured":"Nikos Kolotouros Georgios Pavlakos and Kostas Daniilidis. 2019. Convolutional Mesh Regression for Single-Image Human Shape Reconstruction. arXiv:1905.03244 [cs.CV]","DOI":"10.1109\/CVPR.2019.00463"},{"key":"e_1_2_1_30_1","volume-title":"Garnett (Eds.)","volume":"32","author":"Lee Hsin-Ying","year":"2019","unstructured":"Hsin-Ying Lee, Xiaodong Yang, Ming-Yu Liu, Ting-Chun Wang, Yu-Ding Lu, Ming-Hsuan Yang, and Jan Kautz. 2019. Dancing to Music. In Advances in Neural Information Processing Systems, H. Wallach, H. Larochelle, A. Beygelzimer, F. d'Alch\u00e9-Buc, E. Fox, and R. Garnett (Eds.), Vol. 32. Curran Associates, Inc. https:\/\/proceedings.neurips.cc\/paper\/2019\/file\/7ca57a9f85a19a6e4b9a248c1daca185-Paper.pdf"},{"key":"e_1_2_1_31_1","volume-title":"SAME: Skeleton-Agnostic Motion Embedding for Character Animation. In SIGGRAPH Asia 2023 Conference Papers. 1--11","author":"Lee Sunmin","year":"2023","unstructured":"Sunmin Lee, Taeho Kang, Jungnam Park, Jehee Lee, and Jungdam Won. 2023a. SAME: Skeleton-Agnostic Motion Embedding for Character Animation. In SIGGRAPH Asia 2023 Conference Papers. 1--11."},{"key":"e_1_2_1_32_1","volume-title":"QuestEnvSim: Environment-Aware Simulated Motion Tracking from Sparse Sensors. arXiv preprint arXiv:2306.05666","author":"Lee Sunmin","year":"2023","unstructured":"Sunmin Lee, Sebastian Starke, Yuting Ye, Jungdam Won, and Alexander Winkler. 2023b. QuestEnvSim: Environment-Aware Simulated Motion Tracking from Sparse Sensors. arXiv preprint arXiv:2306.05666 (2023)."},{"key":"e_1_2_1_33_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.01243"},{"key":"e_1_2_1_34_1","volume-title":"Learning to Generate Diverse Dance Motions with Transformer. (08","author":"Li Jiaman","year":"2020","unstructured":"Jiaman Li, Yihang Yin, Hang Chu, Yi Zhou, Tingwu Wang, Sanja Fidler, and Hao Li. 2020. Learning to Generate Diverse Dance Motions with Transformer. (08 2020)."},{"key":"e_1_2_1_35_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01985"},{"key":"e_1_2_1_36_1","doi-asserted-by":"publisher","DOI":"10.1145\/3528223.3530157"},{"key":"e_1_2_1_37_1","volume-title":"ACE: Adversarial Correspondence Embedding for Cross Morphology Motion Retargeting from Human to Nonhuman Characters. arXiv preprint arXiv:2305.14792","author":"Li Tianyu","year":"2023","unstructured":"Tianyu Li, Jungdam Won, Alexander Clegg, Jeonghwan Kim, Akshara Rai, and Sehoon Ha. 2023b. ACE: Adversarial Correspondence Embedding for Cross Morphology Motion Retargeting from Human to Nonhuman Characters. arXiv preprint arXiv:2305.14792 (2023)."},{"key":"e_1_2_1_38_1","doi-asserted-by":"publisher","DOI":"10.1145\/3386569.3392422"},{"key":"e_1_2_1_39_1","doi-asserted-by":"publisher","DOI":"10.1145\/2816795.2818013"},{"key":"e_1_2_1_40_1","doi-asserted-by":"publisher","DOI":"10.1145\/3072959.3073596"},{"key":"e_1_2_1_41_1","doi-asserted-by":"crossref","unstructured":"Georgios Pavlakos Xiaowei Zhou Konstantinos G. Derpanis and Kostas Daniilidis. 2017. Coarse-to-Fine Volumetric Prediction for Single-Image 3D Human Pose. arXiv:1611.07828 [cs.CV]","DOI":"10.1109\/CVPR.2017.139"},{"key":"e_1_2_1_42_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.01080"},{"key":"e_1_2_1_43_1","doi-asserted-by":"publisher","DOI":"10.1145\/3625264"},{"key":"e_1_2_1_44_1","volume-title":"Guibas","author":"Qi Charles R.","year":"2017","unstructured":"Charles R. Qi, Hao Su, Kaichun Mo, and Leonidas J. Guibas. 2017. PointNet: Deep Learning on Point Sets for 3D Classification and Segmentation. arXiv:1612.00593 [cs.CV]"},{"key":"e_1_2_1_45_1","volume-title":"LiveHPS: LiDAR-based Scene-level Human Pose and Shape Estimation in Free Environment. arXiv preprint arXiv:2402.17171","author":"Ren Yiming","year":"2024","unstructured":"Yiming Ren, Xiao Han, Chengfeng Zhao, Jingya Wang, Lan Xu, Jingyi Yu, and Yuexin Ma. 2024. LiveHPS: LiDAR-based Scene-level Human Pose and Shape Estimation in Free Environment. arXiv preprint arXiv:2402.17171 (2024)."},{"key":"e_1_2_1_46_1","doi-asserted-by":"publisher","DOI":"10.1109\/TVCG.2023.3247088"},{"key":"e_1_2_1_47_1","volume-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. 574--584","author":"Shetty Karthik","year":"2023","unstructured":"Karthik Shetty, Annette Birkhold, Srikrishna Jaganathan, Norbert Strobel, Markus Kowarschik, Andreas Maier, and Bernhard Egger. 2023. PLIKS: A Pseudo-Linear Inverse Kinematic Solver for 3D Human Body Estimation. In Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. 574--584."},{"key":"e_1_2_1_48_1","volume-title":"Learning structured output representation using deep conditional generative models. Advances in neural information processing systems 28","author":"Sohn Kihyuk","year":"2015","unstructured":"Kihyuk Sohn, Honglak Lee, and Xinchen Yan. 2015. Learning structured output representation using deep conditional generative models. Advances in neural information processing systems 28 (2015)."},{"key":"e_1_2_1_49_1","volume-title":"Proceedings of the IEEE\/CVF International Conference on Computer Vision. 1292--1304","author":"Takmaz Ay\u00e7a","year":"2023","unstructured":"Ay\u00e7a Takmaz, Jonas Schult, Irem Kaftan, Mertcan Ak\u00e7ay, Bastian Leibe, Robert Sumner, Francis Engelmann, and Siyu Tang. 2023. 3D Segmentation of Humans in Point Clouds with Synthetic Data. In Proceedings of the IEEE\/CVF International Conference on Computer Vision. 1292--1304."},{"key":"e_1_2_1_50_1","volume-title":"RSMT: Real-time Stylized Motion Transition for Characters. arXiv preprint arXiv:2306.11970","author":"Tang Xiangjun","year":"2023","unstructured":"Xiangjun Tang, Linjun Wu, He Wang, Bo Hu, Xu Gong, Yuchen Liao, Songnan Li, Qilong Kou, and Xiaogang Jin. 2023. RSMT: Real-time Stylized Motion Transition for Characters. arXiv preprint arXiv:2306.11970 (2023)."},{"key":"e_1_2_1_51_1","unstructured":"Guy Tevet Sigal Raab Brian Gordon Yonatan Shafir Amit H Bermano and Daniel Cohen-Or. 2023. Human Motion Diffusion Model. In ICLR."},{"key":"e_1_2_1_52_1","volume-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition.","author":"Tseng Jonathan","year":"2023","unstructured":"Jonathan Tseng, Rodrigo Castellon, and C Karen Liu. 2023. EDGE: Editable Dance Generation From Music. In Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition."},{"key":"e_1_2_1_53_1","volume-title":"Computer graphics forum","author":"Marcard Timo Von","unstructured":"Timo Von Marcard, Bodo Rosenhahn, Michael J Black, and Gerard Pons-Moll. 2017. Sparse inertial poser: Automatic 3d human pose estimation from sparse imus. In Computer graphics forum, Vol. 36. Wiley Online Library, 349--360."},{"key":"e_1_2_1_54_1","first-page":"1","article-title":"Combining Recurrent Neural Networks and Adversarial Training for Human Motion Synthesis and Control","volume":"27","author":"Wang Zhiyong","year":"2021","unstructured":"Zhiyong Wang, Jinxiang Chai, and Shihong Xia. 2021. Combining Recurrent Neural Networks and Adversarial Training for Human Motion Synthesis and Control. IEEE Transactions on Visualization and Computer Graphics 27, 1 (jan 2021), 14--28.","journal-title":"IEEE Transactions on Visualization and Computer Graphics"},{"key":"e_1_2_1_55_1","unstructured":"Wen-Li Wei Jen-Chun Lin Tyng-Luh Liu and Hong-Yuan Mark Liao. 2022. Capturing Humans in Motion: Temporal-Attentive 3D Human Pose and Shape Estimation from Monocular Video. arXiv:2203.08534 [cs.CV]"},{"key":"e_1_2_1_56_1","doi-asserted-by":"publisher","DOI":"10.1145\/1833349.1778779"},{"key":"e_1_2_1_57_1","doi-asserted-by":"publisher","DOI":"10.1145\/3550469.3555411"},{"key":"e_1_2_1_58_1","doi-asserted-by":"publisher","DOI":"10.1145\/3528223.3530067"},{"key":"e_1_2_1_59_1","doi-asserted-by":"crossref","unstructured":"Sijie Yan Yuanjun Xiong and Dahua Lin. 2018. Spatial Temporal Graph Convolutional Networks for Skeleton-Based Action Recognition. arXiv:1801.07455 [cs.CV]","DOI":"10.1609\/aaai.v32i1.12328"},{"key":"e_1_2_1_60_1","volume-title":"Computer Graphics Forum","author":"Yang Dongseok","unstructured":"Dongseok Yang, Jiho Kang, Lingni Ma, Joseph Greer, Yuting Ye, and Sung-Hee Lee. 2024. DivaTrack: Diverse Bodies and Motions from Acceleration-Enhanced Three-Point Trackers. In Computer Graphics Forum. Wiley Online Library, e15057."},{"key":"e_1_2_1_61_1","volume-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. 21222--21232","author":"Ye Vickie","year":"2023","unstructured":"Vickie Ye, Georgios Pavlakos, Jitendra Malik, and Angjoo Kanazawa. 2023. Decoupling human and camera motion from videos in the wild. In Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition. 21222--21232."},{"key":"e_1_2_1_62_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICIP42928.2021.9506588"},{"key":"e_1_2_1_63_1","unstructured":"Xumin Yu Lulu Tang Yongming Rao Tiejun Huang Jie Zhou and Jiwen Lu. 2022. Point-BERT: Pre-training 3D Point Cloud Transformers with Masked Point Modeling. arXiv:2111.14819 [cs.CV]"},{"key":"e_1_2_1_64_1","volume-title":"MotionDiffuse: Text-Driven Human Motion Generation with Diffusion Model. arXiv preprint arXiv:2208.15001","author":"Zhang Mingyuan","year":"2022","unstructured":"Mingyuan Zhang, Zhongang Cai, Liang Pan, Fangzhou Hong, Xinying Guo, Lei Yang, and Ziwei Liu. 2022. MotionDiffuse: Text-Driven Human Motion Generation with Diffusion Model. arXiv preprint arXiv:2208.15001 (2022)."},{"key":"e_1_2_1_65_1","doi-asserted-by":"publisher","DOI":"10.1145\/3618397"},{"key":"e_1_2_1_66_1","volume-title":"MotionBERT: Unified Pretraining for Human Motion Analysis. arXiv preprint arXiv:2210.06551","author":"Zhu Wentao","year":"2022","unstructured":"Wentao Zhu, Xiaoxuan Ma, Zhaoyang Liu, Libin Liu, Wayne Wu, and Yizhou Wang. 2022. MotionBERT: Unified Pretraining for Human Motion Analysis. arXiv preprint arXiv:2210.06551 (2022)."},{"key":"e_1_2_1_67_1","doi-asserted-by":"publisher","DOI":"10.1145\/2601097.2601165"}],"container-title":["ACM Transactions on Graphics"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3687991","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3687991","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,19]],"date-time":"2025-06-19T01:09:58Z","timestamp":1750295398000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3687991"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,11,19]]},"references-count":67,"journal-issue":{"issue":"6","published-print":{"date-parts":[[2024,12,19]]}},"alternative-id":["10.1145\/3687991"],"URL":"https:\/\/doi.org\/10.1145\/3687991","relation":{},"ISSN":["0730-0301","1557-7368"],"issn-type":[{"value":"0730-0301","type":"print"},{"value":"1557-7368","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,11,19]]},"assertion":[{"value":"2024-11-19","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}