Timezone: »
The ability of forecasting future human motion is important for human-machine interaction systems to understand human behaviors and make interaction. In this work, we focus on developing models to predict future human motion from past observed video frames. Motivated by the observation that human motion is closely related to the action being performed, we propose to explore action context to guide motion prediction. Specifically, we construct an action-specific memory bank to store representative motion dynamics for each action category, and design a query-read process to retrieve some motion dynamics from the memory bank. The retrieved dynamics are consistent with the action depicted in the observed video frames and serve as a strong prior knowledge to guide motion prediction. We further formulate an action constraint loss to ensure the global semantic consistency of the predicted motion. Extensive experiments demonstrate the effectiveness of the proposed approach, and we achieve state-of-the-art performance on 3D human motion prediction.
Author Information
Jiangxin Sun (Sun Yat-sen University)
Zihang Lin (SUN YAT-SEN UNIVERSITY)
Xintong Han (Huya Inc)
Jian-Fang Hu (SUN YAT-SEN UNIVERSITY)
Jia Xu (Tencent AI Lab)
I am a principal researcher at Tencent AI Lab. Before returning to China, I was a senior research scientist in the Intel Visual Computing Lab, lead by the awesome Vladlen Koltun. I received my Ph.D. in Computer Sciences at the University of Wisconsin-Madison, with my thesis committee of Prof. Vikas Singh (advisor), Prof. Chuck Dyer, Prof. Jerry Zhu, Prof. Jude Shavlik, and Prof. Mark Craven. I was a visiting student in University of Toronto and in Toyota Technological Institute at Chicago, both working with Prof. Raquel Urtasun. Before graduate school, I obtained my B.S. degree from the Department of Computer Science and Technology at Nanjing University, China. My major interests include computer vision, deep learning, reinforcement learning, and robotics.
Wei-Shi Zheng (SUN YAT-SEN UNIVERSITY)
More from the Same Authors
-
2022 Poster: You Never Stop Dancing: Non-freezing Dance Generation via Bank-constrained Manifold Projection »
Jiangxin Sun · Chunyu Wang · Huang Hu · Hanjiang Lai · Zhi Jin · Jian-Fang Hu -
2022 Poster: Text-Adaptive Multiple Visual Prototype Matching for Video-Text Retrieval »
Chengzhi Lin · Ancong Wu · Junwei Liang · Jun Zhang · Wenhang Ge · Wei-Shi Zheng · Chunhua Shen -
2022 Spotlight: One Model to Edit Them All: Free-Form Text-Driven Image Manipulation with Semantic Modulations »
Yiming Zhu · Hongyu Liu · Yibing Song · Ziyang Yuan · Xintong Han · Chun Yuan · Qifeng Chen · Jue Wang -
2022 Spotlight: Lightning Talks 2A-4 »
Sarthak Mittal · Richard Grumitt · Zuoyu Yan · Lihao Wang · Dongsheng Wang · Alexander Korotin · Jiangxin Sun · Ankit Gupta · Vage Egiazarian · Tengfei Ma · Yi Zhou · Yishi Xu · Albert Gu · Biwei Dai · Chunyu Wang · Yoshua Bengio · Uros Seljak · Miaoge Li · Guillaume Lajoie · Yiqun Wang · Liangcai Gao · Lingxiao Li · Jonathan Berant · Huang Hu · Xiaoqing Zheng · Zhibin Duan · Hanjiang Lai · Evgeny Burnaev · Zhi Tang · Zhi Jin · Xuanjing Huang · Chaojie Wang · Yusu Wang · Jian-Fang Hu · Bo Chen · Chao Chen · Hao Zhou · Mingyuan Zhou -
2022 Spotlight: You Never Stop Dancing: Non-freezing Dance Generation via Bank-constrained Manifold Projection »
Jiangxin Sun · Chunyu Wang · Huang Hu · Hanjiang Lai · Zhi Jin · Jian-Fang Hu -
2022 Poster: One Model to Edit Them All: Free-Form Text-Driven Image Manipulation with Semantic Modulations »
Yiming Zhu · Hongyu Liu · Yibing Song · Ziyang Yuan · Xintong Han · Chun Yuan · Qifeng Chen · Jue Wang