JOURNAL ARTICLE

Fg-T2M: Fine-Grained Text-Driven Human Motion Generation via Diffusion Model

Abstract

Text-driven human motion generation in computer vision is both significant and challenging. However, current methods are limited to producing either deterministic or imprecise motion sequences, failing to effectively control the temporal and spatial relationships required to conform to a given text description. In this work, we propose a fine-grained method for generating high-quality, conditional human motion sequences supporting precise text description. Our approach consists of two key components: 1) a linguistics-structure assisted module that constructs accurate and complete language feature to fully utilize text information; and 2) a context-aware progressive reasoning module that learns neighborhood and overall semantic linguistics features from shallow and deep graph neural networks to achieve a multi-step inference. Experiments show that our approach outperforms text-driven motion generation methods on HumanML3D and KIT test sets and generates better visually confirmed motion to the text conditions.

Keywords:
Computer science Motion (physics) Diffusion Artificial intelligence Human motion Physics Thermodynamics

Metrics

30
Cited By
7.47
FWCI (Field Weighted Citation Impact)
42
Refs
0.97
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Citation History

Topics

Human Motion and Animation
Physical Sciences →  Engineering →  Control and Systems Engineering
Human Pose and Action Recognition
Physical Sciences →  Computer Science →  Computer Vision and Pattern Recognition
Generative Adversarial Networks and Image Synthesis
Physical Sciences →  Computer Science →  Computer Vision and Pattern Recognition

Related Documents

JOURNAL ARTICLE

Fg-T2M++: LLMs-Augmented Fine-Grained Text Driven Human Motion Generation

Yin WangLi MuYang LiuZhiying LengFrederick W. B. LiZ.F. ZhangXiaohui Liang

Journal:   International Journal of Computer Vision Year: 2025 Vol: 133 (7)Pages: 4277-4293
JOURNAL ARTICLE

Enhanced Fine-Grained Motion Diffusion for Text-Driven Human Motion Synthesis

Dongwei MaXiaoning SunHuaijiang SunShengxiang HuBin LiWeiqing LiJianfeng Lu

Journal:   Proceedings of the AAAI Conference on Artificial Intelligence Year: 2024 Vol: 38 (6)Pages: 5876-5884
JOURNAL ARTICLE

MotionDiffuse: Text-Driven Human Motion Generation With Diffusion Model

Mingyuan ZhangZhongang CaiLiang PanFangzhou HongXinying GuoLei YangZiwei Liu

Journal:   IEEE Transactions on Pattern Analysis and Machine Intelligence Year: 2024 Vol: 46 (6)Pages: 4115-4128
JOURNAL ARTICLE

Light-T2M: A Lightweight and Fast Model for Text-to-motion Generation

Ling-An ZengGuohong HuangGaojie WuWei‐Shi Zheng

Journal:   Proceedings of the AAAI Conference on Artificial Intelligence Year: 2025 Vol: 39 (9)Pages: 9797-9805
© 2026 ScienceGate Book Chapters — All rights reserved.