Skip to content

[CVPR 2025] MG-MotionLLM: A Unified Framework for Motion Comprehension and Generation across Multiple Granularities

Notifications You must be signed in to change notification settings

CVI-SZU/MG-MotionLLM

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

5 Commits
 
 
 
 

Repository files navigation

(CVPR 2025) MG-MotionLLM: A Unified Framework for Motion Comprehension and Generation across Multiple Granularities

Bizhu Wu · Jinheng Xie · Keming Shen · Zhe Kong

Jianfeng Ren* · Ruibin Bai · Rong Qu · Linlin Shen*

*Corresponding Authors

arXiv

Description

MG-MotionLLM can address diverse motion-relevant tasks at multiple granularities by giving different instructions in a unified manner.

  • coarse-grained: e.g. text-to-motion and motion captioning (upper block)
  • fine-grained: e.g. motion-to-detailed text and motion localization (bottom block).
teaser

To achieve this, we propose multi-granularity training scheme with novel auxiliary tasks captures motion-related features at different levels, improving understanding across a wide range of tasks. Specifically, we pretrain the model with a total of 28 distinct motion-relevant tasks, including 12 existing classical coarse-grained tasks and 16 newly proposed fine-grained ones. Here, we display examples of prompt templates for a part of tasks used during training.

tasks_template

Visualization

We display some novel applications of our MG-MotionLLM.

  • text-driven fine-grained motion editing: Temporal Editing (left), Spatial Editing (middle), and Spatial-Temporal Editing (right).
edit
  • fine-grained captioning of both whole (up) and partial (bottom) motion sequences, and motion localization via fine-grained textual description (middle).
novel_apps

More Information (code, weights, etc)

For code, weights, etc, please see here.

Bibtex

If you use our code in your research, kindly cite our work:

@article{wu2025mg,
  title={MG-MotionLLM: A Unified Framework for Motion Comprehension and Generation across Multiple Granularities},
  author={Wu, Bizhu and Xie, Jinheng and Shen, Keming and Kong, Zhe and Ren, Jianfeng and Bai, Ruibin and Qu, Rong and Shen, Linlin},
  journal={arXiv preprint arXiv:2504.02478},
  year={2025}
}

About

[CVPR 2025] MG-MotionLLM: A Unified Framework for Motion Comprehension and Generation across Multiple Granularities

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published