An aggregation of human motion understanding research, feel free to contribute.
Motion Generation Motion Editing Motion Stylization
Human-Object Interaction Human-Scene Interaction Human-Human Interaction
- (JEB 2025) McAllister et al: Behavioural energetics in human locomotion: how energy use influences how we move, McAllister et al.
- (ArXiv 2025) Gu et al: Humanoid Locomotion and Manipulation: Current Progress and Challenges in Control, Planning, and Learning, Gu et al.
- (ArXiv 2024) Zhao et al: Motion Generation Review: Exploring Deep Learning for Lifelike Animation with Manifold, Zhao et al.
- (T-PAMI 2023) Zhu et al: Human Motion Generation: A Survey, Zhu et al.
- (CVPR 2025) PersonalBooth: PersonaBooth: Personalized Text-to-Motion Generation, Kim et al.
- (CVPR 2025) MARDM: Rethinking Diffusion for Text-Driven Human Motion Generation, Meng et al.
- (CVPR 2025) StickMotion: StickMotion: Generating 3D Human Motions by Drawing a Stickman, Wang et al.
- (CVPR 2025) LLaMo: Human Motion Instruction Tuning, Li et al.
- (CVPR 2025) HOP: HOP: Heterogeneous Topology-based Multimodal Entanglement for Co-Speech Gesture Generation, Cheng et al.
- (CVPR 2025) AtoM: AToM: Aligning Text-to-Motion Model at Event-Level with GPT-4Vision Reward, Han et al.
- (CVPR 2025) EnergyMoGen: EnergyMoGen: Compositional Human Motion Generation with Energy-Based Diffusion Model in Latent Space, Zhang et al.
- (CVPR 2025) Languate of Motion: The Language of Motion: Unifying Verbal and Non-verbal Language of 3D Human Motion, Chen et al.
- (CVPR 2025) ScaMo: ScaMo: Exploring the Scaling Law in Autoregressive Motion Generation Model, Lu et al.
- (CVPR 2025) Move in 2D: Move-in-2D: 2D-Conditioned Human Motion Generation, Huang et al.
- (CVPR 2025) SOLAMI: SOLAMI: Social Vision-Language-Action Modeling for Immersive Interaction with 3D Autonomous Characters, Jiang et al.
- (CVPR 2025) MVLift: Lifting Motion to the 3D World via 2D Diffusion, Li et al.
- (ACM Sensys 2025) SHADE-AD: SHADE-AD: An LLM-Based Framework for Synthesizing Activity Data of Alzheimer’s Patients, Fu et al.
- (ICRA 2025) MotionGlot: MotionGlot: A Multi-Embodied Motion Generation Model, Harithas et al.
- (ICLR 2025) CLoSD: CLoSD: Closing the Loop between Simulation and Diffusion for multi-task character control, Tevet et al.
- (ICLR 2025) PedGen: Learning to Generate Diverse Pedestrian Movements from Web Videos with Noisy Labels, Liu et al.
- (ICLR 2025) HGM³: HGM³: Hierarchical Generative Masked Motion Modeling with Hard Token Mining, Jeong et al.
- (ICLR 2025) LaMP: LaMP: Language-Motion Pretraining for Motion Generation, Retrieval, and Captioning, Li et al.
- (ICLR 2025) MotionDreamer: MotionDreamer: One-to-Many Motion Synthesis with Localized Generative Masked Transformer, Wang et al.
- (ICLR 2025) Lyu et al: Towards Unified Human Motion-Language Understanding via Sparse Interpretable Characterization, Lyu et al.
- (ICLR 2025) DART: DART: A Diffusion-Based Autoregressive Motion Model for Real-Time Text-Driven Motion Control, Zhao et al.
- (ICLR 2025) Motion-Agent: Motion-Agent: A Conversational Framework for Human Motion Generation with LLMs, Wu et al.
- (IJCV 2025) Fg-T2M++:。 Fg-T2M++: LLMs-Augmented Fine-Grained Text Driven Human Motion Generation, Wang et al.
- (TVCG 2025) SPORT: SPORT: From Zero-Shot Prompts to Real-Time Motion Generation, Ji et al.
- (ArXiv 2025) HERO: HERO: Human Reaction Generation from Videos, Yu et al.
- (ArXiv 2025) BioMoDiffuse: BioMoDiffuse: Physics-Guided Biomechanical Diffusion for Controllable and Authentic Human Motion Synthesis, Kang et al.
- (ArXiv 2025) ExGes: ExGes: Expressive Human Motion Retrieval and Modulation for Audio-Driven Gesture Synthesis, Zhou et al.
- (ArXiv 2025) AnyTop: AnyTop: Character Animation Diffusion with Any Topology, Gat et al.
- (ArXiv 2025) MotionAnything: Motion Anything: Any to Motion Generation, Zhang et al.
- (ArXiv 2025) GCDance: GCDance: Genre-Controlled 3D Full Body Dance Generation Driven By Music, Liu et al.
- (ArXiv 2025) MotionLab: MotionLab: Unified Human Motion Generation and Editing via the Motion-Condition-Motion Paradigm, Guo et al.
- (ArXiv 2025) CASIM: CASIM: Composite Aware Semantic Injection for Text to Motion Generation, Chang et al.
- (ArXiv 2025) MotionPCM: MotionPCM: Real-Time Motion Synthesis with Phased Consistency Model, Jiang et al.
- (ArXiv 2025) GestureLSM: GestureLSM: Latent Shortcut based Co-Speech Gesture Generation with Spatial-Temporal Modeling, Liu et al.
- (ArXiv 2025) Free-T2M: Free-T2M: Frequency Enhanced Text-to-Motion Diffusion Model With Consistency Loss, Chen et al.
- (ArXiv 2025) LS-GAN: LS-GAN: Human Motion Synthesis with Latent-space GANs, Amballa et al.
- (ArXiv 2025) FlexMotion: FlexMotion: Lightweight, Physics-Aware, and Controllable Human Motion Generation, Tashakori et al.
- (ArXiv 2025) HiSTF Mamba: HiSTF Mamba: Hierarchical Spatiotemporal Fusion with Multi-Granular Body-Spatial Modeling for High-Fidelity Text-to-Motion Generation, Zhan et al.
- (ArXiv 2025) PackDiT: PackDiT: Joint Human Motion and Text Generation via Mutual Prompting, Jiang et al.
- (3DV 2025) Unimotion: Unimotion: Unifying 3D Human Motion Synthesis and Understanding, Li et al.
- (AAAI 2025) UniMuMo: UniMuMo: Unified Text, Music and Motion Generation, Yang et al.
- (AAAI 2025) ALERT-Motion: Autonomous LLM-Enhanced Adversarial Attack for Text-to-Motion, Miao et al.
- (AAAI 2025) MotionCraft: MotionCraft: Crafting Whole-Body Motion with Plug-and-Play Multimodal Controls, Bian et al.
- (AAAI 2025) Light-T2M: Light-T2M: A Lightweight and Fast Model for Text-to-Motion Generation, Zeng et al.
- (WACV 2025) ReinDiffuse: ReinDiffuse: Crafting Physically Plausible Motions with Reinforced Diffusion Model, Han et al.
- (WACV 2025) MoRAG: MoRAG -- Multi-Fusion Retrieval Augmented Generation for Human Motion, Shashank et al.
- (WACV 2025) Mandelli et al: Generation of Complex 3D Human Motion by Temporal and Spatial Composition of Diffusion Models, Mandelli et al.
- (ArXiv 2024) SemTalk: SemTalk: Holistic Co-speech Motion Generation with Frame-level Semantic Emphasis, Zhang et al.
- (ArXiv 2024) InterDance: InterDance: Reactive 3D Dance Generation with Realistic Duet Interactions, Li et al.
- (ArXiv 2024) Motion-2-to-3: Motion-2-to-3: Leveraging 2D Motion Data to Boost 3D Motion Generation, Pi et al.
- (ArXiv 2024) Mogo: Mogo: RQ Hierarchical Causal Transformer for High-Quality 3D Human Motion Generation, Fu et al.
- (ArXiv 2024) CoMA: CoMA: Compositional Human Motion Generation with Multi-modal Agents, Sun et al.
- (ArXiv 2024) SoPo: SoPo: Text-to-Motion Generation Using Semi-Online Preference Optimization, Tan et al.
- (ArXiv 2024) RMD: RMD: A Simple Baseline for More General Human Motion Generation via Training-free Retrieval-Augmented Motion Diffuse, Liao et al.
- (ArXiv 2024) BiPO: BiPO: Bidirectional Partial Occlusion Network for Text-to-Motion Synthesis, Hong et al.
- (ArXiv 2024) DisCoRD: DisCoRD: Discrete Tokens to Continuous Motion via Rectified Flow Decoding, Cho et al.
- (ArXiv 2024) MoTe: MoTe: Learning Motion-Text Diffusion Model for Multiple Generation Tasks, Wue et al.
- (ArXiv 2024) InfiniDreamer: InfiniDreamer: Arbitrarily Long Human Motion Generation via Segment Score Distillation, Zhuo et al.
- (ArXiv 2024) FTMoMamba: FTMoMamba: Motion Generation with Frequency and Text State Space Models, Li et al.
- (ArXiv 2024) KinMo: KinMo: Kinematic-aware Human Motion Understanding and Generation, Zhang et al.
- (ArXiv 2024) Morph: Morph: A Motion-free Physics Optimization Framework for Human Motion Generation, Li et al.
- (ArXiv 2024) KMM: KMM: Key Frame Mask Mamba for Extended Motion Generation, Zhang et al.
- (ArXiv 2024) MotionGPT-2: MotionGPT-2: A General-Purpose Motion-Language Model for Motion Generation and Understanding, Wang et al.
- (ArXiv 2024) Lodge++: Lodge++: High-quality and Long Dance Generation with Vivid Choreography Patterns, Li et al.
- (ArXiv 2024) MotionCLR: MotionCLR: Motion Generation and Training-Free Editing via Understanding Attention Mechanisms, Chen et al.
- (ArXiv 2024) LEAD: LEAD: Latent Realignment for Human Motion Diffusion, Andreou et al.
- (ArXiv 2024) Leite et al.: Enhancing Motion Variation in Text-to-Motion Models via Pose and Video Conditioned Editing, Leite et al.
- (ArXiv 2024) MotionRL: MotionRL: Align Text-to-Motion Generation to Human Preferences with Multi-Reward Reinforcement Learning, Liu et al.
- (ArXiv 2024) MotionLLM: MotionLLM: Understanding Human Behaviors from Human Motions and Videos, Chen et al.
- (ArXiv 2024) Wang et al: Quo Vadis, Motion Generation? From Large Language Models to Large Motion Models, Wang et al.
- (ArXiv 2024) T2M-X: T2M-X: Learning Expressive Text-to-Motion Generation from Partially Annotated Data, Liu et al.
- (ArXiv 2024) BAD: BAD: Bidirectional Auto-regressive Diffusion for Text-to-Motion Generation, Hosseyni et al.
- (ArXiv 2024) synNsync: Synergy and Synchrony in Couple Dances, Manukele et al.
- (EMNLP 2024) Dong et al: Word-Conditioned 3D American Sign Language Motion Generation, Dong et al.
- (NeurIPS D&B 2024) Text to blind motion: Text to Blind Motion, Kim et al.
- (NeurIPS 2024) UniMTS: UniMTS: Unified Pre-training for Motion Time Series, Zhang et al.
- (NeurIPS 2024) Christopher et al.: Constrained Synthesis with Projected Diffusion Models, Christopher et al.
- (NeurIPS 2024) MoMu-Diffusion: MoMu-Diffusion: On Learning Long-Term Motion-Music Synchronization and Correspondence, You et al.
- (NeurIPS 2024) MoGenTS: MoGenTS: Motion Generation based on Spatial-Temporal Joint Modeling, Yuan et al.
- (NeurIPS 2024) M3GPT: M3GPT: An Advanced Multimodal, Multitask Framework for Motion Comprehension and Generation, Luo et al.
- (NeurIPS Workshop 2024) Bikov et al: Fitness Aware Human Motion Generation with Fine-Tuning, Bikov et al.
- (NeurIPS Workshop 2024) DGFM: DGFM: Full Body Dance Generation Driven by Music Foundation Models, Liu et al.
- (ICPR 2024) FG-MDM: FG-MDM: Towards Zero-Shot Human Motion Generation via ChatGPT-Refined Descriptions, Shi et al.
- (ACM MM 2024) SynTalker: Enabling Synergistic Full-Body Control in Prompt-Based Co-Speech Motion Generation, Chen et al.
- (ACM MM 2024) L3EM: Towards Emotion-enriched Text-to-Motion Generation via LLM-guided Limb-level Emotion Manipulating. Yu et al.
- (ACM MM 2024) StableMoFusion: StableMoFusion: Towards Robust and Efficient Diffusion-based Motion Generation Framework, Huang et al.
- (ACM MM 2024) SATO: SATO: Stable Text-to-Motion Framework, Chen et al.
- (ICANN 2024) PIDM: PIDM: Personality-Aware Interaction Diffusion Model for Gesture Generation, Shibasaki et al.
- (HFES 2024) Macwan et al: High-Fidelity Worker Motion Simulation With Generative AI, Macwan et al.
- (ECCV 2024) Jin et al.: Local Action-Guided Motion Diffusion Model for Text-to-Motion Generation, Jin et al.
- (ECCV 2024) Motion Mamba: Motion Mamba: Efficient and Long Sequence Motion Generation, Zhong et al.
- (ECCV 2024) EMDM: EMDM: Efficient Motion Diffusion Model for Fast, High-Quality Human Motion Generation, Zhou et al.
- (ECCV 2024) CoMo: CoMo: Controllable Motion Generation through Language Guided Pose Code Editing, Huang et al.
- (ECCV 2024) CoMusion: CoMusion: Towards Consistent Stochastic Human Motion Prediction via Motion Diffusion, Sun et al.
- (ECCV 2024) Shan et al.: Towards Open Domain Text-Driven Synthesis of Multi-Person Motions, Shan et al.
- (ECCV 2024) ParCo: ParCo: Part-Coordinating Text-to-Motion Synthesis, Zou et al.
- (ECCV 2024) Sampieri et al.: Length-Aware Motion Synthesis via Latent Diffusion, Sampieri et al.
- (ECCV 2024) ChroAccRet: Chronologically Accurate Retrieval for Temporal Grounding of Motion-Language Models, Fujiwara et al.
- (ECCV 2024) MHC: Generating Physically Realistic and Directable Human Motions from Multi-Modal Inputs, Liu et al.
- (ECCV 2024) ProMotion: Plan, Posture and Go: Towards Open-vocabulary Text-to-Motion Generation, Liu et al.
- (ECCV 2024) FreeMotion: FreeMotion: MoCap-Free Human Motion Synthesis with Multimodal Large Language Models, Zhang et al.
- (ECCV 2024) Text Motion Translator: Text Motion Translator: A Bi-Directional Model for Enhanced 3D Human Motion Generation from Open-Vocabulary Descriptions, Qian et al.
- (ECCV 2024) FreeMotion: FreeMotion: A Unified Framework for Number-free Text-to-Motion Synthesis, Fan et al.
- (ECCV 2024) Kinematic Phrases: Bridging the Gap between Human Motion and Action Semantics via Kinematic Phrases, Liu et al.
- (ECCV 2024) MotionChain: MotionChain: Conversational Motion Controllers via Multimodal Prompts, Jiang et al.
- (ECCV 2024) SMooDi: SMooDi: Stylized Motion Diffusion Model, Zhong et al.
- (ECCV 2024) BAMM: BAMM: Bidirectional Autoregressive Motion Model, Pinyoanuntapong et al.
- (ECCV 2024) MotionLCM: MotionLCM: Real-time Controllable Motion Generation via Latent Consistency Model, Dai et al.
- (ECCV 2024) Ren et al.: Realistic Human Motion Generation with Cross-Diffusion Models, Ren et al.
- (ECCV 2024) M2D2M: M2D2M: Multi-Motion Generation from Text with Discrete Diffusion Models, Chi et al.
- (ECCV 2024) Large Motion Model: Large Motion Model for Unified Multi-Modal Motion Generation, Zhang et al.
- (ECCV 2024) TesMo: Generating Human Interaction Motions in Scenes with Text Control, Yi et al.
- (ECCV 2024) TLcontrol: TLcontrol: Trajectory and Language Control for Human Motion Synthesis, Wan et al.
- (ICME 2024) ExpGest: ExpGest: Expressive Speaker Generation Using Diffusion Model and Hybrid Audio-Text Guidance, Cheng et al.
- (ICME Workshop 2024) Chen et al: Anatomically-Informed Vector Quantization Variational Auto-Encoder for Text-to-Motion Generation, Chen et al.
- (ICML 2024) HumanTOMATO: HumanTOMATO: Text-aligned Whole-body Motion Generation, Lu et al.
- (ICML 2024) GPHLVM: Bringing Motion Taxonomies to Continuous Domains via GPLVM on Hyperbolic Manifolds, Jaquier et al.
- (SIGGRAPH 2024) CondMDI: Flexible Motion In-betweening with Diffusion Models, Cohan et al.
- (SIGGRAPH 2024) CAMDM: Taming Diffusion Probabilistic Models for Character Control, Chen et al.
- (SIGGRAPH 2024) LGTM: LGTM: Local-to-Global Text-Driven Human Motion Diffusion Models, Sun et al.
- (SIGGRAPH 2024) TEDi: TEDi: Temporally-Entangled Diffusion for Long-Term Motion Synthesis, Zhang et al.
- (SIGGRAPH 2024) A-MDM: Interactive Character Control with Auto-Regressive Motion Diffusion Models, Shi et al.
- (SIGGRAPH 2024) Starke et al.: Categorical Codebook Matching for Embodied Character Controllers, Starke et al.
- (SIGGRAPH 2024) SuperPADL: SuperPADL: Scaling Language-Directed Physics-Based Control with Progressive Supervised Distillation, Juravsky et al.
- (CVPR 2024) ProgMoGen: Programmable Motion Generation for Open-set Motion Control Tasks, Liu et al.
- (CVPR 2024) PACER+: PACER+: On-Demand Pedestrian Animation Controller in Driving Scenarios, Wang et al.
- (CVPR 2024) AMUSE: Emotional Speech-driven 3D Body Animation via Disentangled Latent Diffusion, Chhatre et al.
- (CVPR 2024) Liu et al.: Towards Variable and Coordinated Holistic Co-Speech Motion Generation, Liu et al.
- (CVPR 2024) MAS: MAS: Multi-view Ancestral Sampling for 3D motion generation using 2D diffusion, Kapon et al.
- (CVPR 2024) WANDR: WANDR: Intention-guided Human Motion Generation, Diomataris et al.
- (CVPR 2024) MoMask: MoMask: Generative Masked Modeling of 3D Human Motions, Guo et al.
- (CVPR 2024) ChapPose: ChatPose: Chatting about 3D Human Pose, Feng et al.
- (CVPR 2024) AvatarGPT: AvatarGPT: All-in-One Framework for Motion Understanding, Planning, Generation and Beyond, Zhou et al.
- (CVPR 2024) MMM: MMM: Generative Masked Motion Model, Pinyoanuntapong et al.
- (CVPR 2024) AAMDM: AAMDM: Accelerated Auto-regressive Motion Diffusion Model, Li et al.
- (CVPR 2024) OMG: OMG: Towards Open-vocabulary Motion Generation via Mixture of Controllers, Liang et al.
- (CVPR 2024) FlowMDM: FlowMDM: Seamless Human Motion Composition with Blended Positional Encodings, Barquero et al.
- (CVPR 2024) Digital Life Project: Digital Life Project: Autonomous 3D Characters with Social Intelligence, Cai et al.
- (CVPR 2024) EMAGE: EMAGE: Towards Unified Holistic Co-Speech Gesture Generation via Expressive Masked Audio Gesture Modeling, Liu et al.
- (CVPR Workshop 2024) STMC: Multi-Track Timeline Control for Text-Driven 3D Human Motion Generation, Petrovich et al.
- (CVPR Workshop 2024) InstructMotion: Exploring Text-to-Motion Generation with Human Preference, Sheng et al.
- (ICLR 2024) Single Motion Diffusion: Raab et al.
- (ICLR 2024) NeRM: NeRM: Learning Neural Representations for High-Framerate Human Motion Synthesis, Wei et al.
- (ICLR 2024) PriorMDM: PriorMDM: Human Motion Diffusion as a Generative Prior, Shafir et al.
- (ICLR 2024) OmniControl: OmniControl: Control Any Joint at Any Time for Human Motion Generation, Xie et al.
- (ICLR 2024) Adiya et al.: Bidirectional Temporal Diffusion Model for Temporally Consistent Human Animation, Adiya et al.
- (ICLR 2024) Duolando: Duolando: Follower GPT with Off-Policy Reinforcement Learning for Dance Accompaniment, Li et al.
- (AAAI 2024) HuTuDiffusion: HuTuMotion: Human-Tuned Navigation of Latent Motion Diffusion Models with Minimal Feedback, Han et al.
- (AAAI 2024) AMD: AMD: Anatomical Motion Diffusion with Interpretable Motion Decomposition and Fusion, Jing et al.
- (AAAI 2024) MotionMix: MotionMix: Weakly-Supervised Diffusion for Controllable Motion Generation, Hoang et al.
- (AAAI 2024) B2A-HDM: Towards Detailed Text-to-Motion Synthesis via Basic-to-Advanced Hierarchical Diffusion Model, Xie et al.
- (AAAI 2024) Everything2Motion: Everything2Motion: Synchronizing Diverse Inputs via a Unified Framework for Human Motion Synthesis, Fan et al.
- (AAAI 2024) MotionGPT: MotionGPT: Finetuned LLMs are General-Purpose Motion Generators, Zhang et al.
- (AAAI 2024) Dong et al: Enhanced Fine-grained Motion Diffusion for Text-driven Human Motion Synthesis, Dong et al.
- (AAAI 2024) UNIMASKM: A Unified Masked Autoencoder with Patchified Skeletons for Motion Synthesis, Mascaro et al.
- (AAAI 2024) B2A-HDM: Towards Detailed Text-to-Motion Synthesis via Basic-to-Advanced Hierarchical Diffusion Model, Xie et al.
- (TPAMI 2024) GUESS: GUESS: GradUally Enriching SyntheSis for Text-Driven Human Motion Generation, Gao et al.
- (WACV 2024) Xie et al.: Sign Language Production with Latent Motion Transformer, Xie et al.
- (NeurIPS 2023) GraphMotion: Act As You Wish: Fine-grained Control of Motion Diffusion Model with Hierarchical Semantic Graphs, Jin et al.
- (NeurIPS 2023) MotionGPT: MotionGPT: Human Motion as Foreign Language, Jiang et al.
- (NeurIPS 2023) FineMoGen: FineMoGen: Fine-Grained Spatio-Temporal Motion Generation and Editing, Zhang et al.
- (NeurIPS 2023) InsActor: InsActor: Instruction-driven Physics-based Characters, Ren et al.
- (ICCV 2023) AttT2M: AttT2M: Text-Driven Human Motion Generation with Multi-Perspective Attention Mechanism, Zhong et al.
- (ICCV 2023) TMR: TMR: Text-to-Motion Retrieval Using Contrastive 3D Human Motion Synthesis, Petrovich et al.
- (ICCV 2023) MAA: Make-An-Animation: Large-Scale Text-conditional 3D Human Motion Generation, Azadi et al.
- (ICCV 2023) PhysDiff: PhysDiff: Physics-Guided Human Motion Diffusion Model, Yuan et al.
- (ICCV 2023) ReMoDiffusion: ReMoDiffuse: Retrieval-Augmented Motion Diffusion Model, Zhang et al.
- (ICCV 2023) BelFusion: BeLFusion: Latent Diffusion for Behavior-Driven Human Motion Prediction, Barquero et al.
- (ICCV 2023) GMD: GMD: Guided Motion Diffusion for Controllable Human Motion Synthesis, Karunratanakul et al.
- (ICCV 2023) HMD-NeMo: HMD-NeMo: Online 3D Avatar Motion Generation From Sparse Observations, Aliakbarian et al.
- (ICCV 2023) SINC: SINC: Spatial Composition of 3D Human Motions for Simultaneous Action Generation, Athanasiou et al.
- (ICCV 2023) Kong et al.: Priority-Centric Human Motion Generation in Discrete Latent Space, Kong et al.
- (ICCV 2023) FgT2M: Fg-T2M: Fine-Grained Text-Driven Human Motion Generation via Diffusion Model, Wang et al.
- (ICCV 2023) EMS: Breaking The Limits of Text-conditioned 3D Motion Synthesis with Elaborative Descriptions, Qian et al.
- (SIGGRAPH 2023) GenMM: Example-based Motion Synthesis via Generative Motion Matching, Li et al.
- (SIGGRAPH 2023) GestureDiffuCLIP: GestureDiffuCLIP: Gesture Diffusion Model with CLIP Latents, Ao et al.
- (SIGGRAPH 2023) BodyFormer: BodyFormer: Semantics-guided 3D Body Gesture Synthesis with Transformer, Pang et al.
- (SIGGRAPH 2023) Alexanderson et al.: Listen, denoise, action! Audio-driven motion synthesis with diffusion models, Alexanderson et al.
- (CVPR 2023) AGroL: Avatars Grow Legs: Generating Smooth Human Motion from Sparse Tracking Inputs with Diffusion Model, Du et al.
- (CVPR 2023) TALKSHOW: Generating Holistic 3D Human Motion from Speech, Yi et al.
- (CVPR 2023) T2M-GPT: T2M-GPT: Generating Human Motion from Textual Descriptions with Discrete Representations, Zhang et al.
- (CVPR 2023) UDE: UDE: A Unified Driving Engine for Human Motion Generation, Zhou et al.
- (CVPR 2023) OOHMG: Being Comes from Not-being: Open-vocabulary Text-to-Motion Generation with Wordless Training, Lin et al.
- (CVPR 2023) EDGE: EDGE: Editable Dance Generation From Music, Tseng et al.
- (CVPR 2023) MLD: Executing your Commands via Motion Diffusion in Latent Space, Chen et al.
- (CVPR 2023) MoDi: MoDi: Unconditional Motion Synthesis from Diverse Data, Raab et al.
- (CVPR 2023) MoFusion: MoFusion: A Framework for Denoising-Diffusion-based Motion Synthesis, Dabral et al.
- (CVPR 2023) Mo et al.: Continuous Intermediate Token Learning with Implicit Motion Manifold for Keyframe Based Motion Interpolation, Mo et al.
- (ICLR 2023) HMDM: MDM: Human Motion Diffusion Model, Tevet et al.
- (TPAMI 2023) MotionDiffuse: MotionDiffuse: Text-Driven Human Motion Generation with Diffusion Model, Zhang et al.
- (TPAMI 2023) Bailando++: Bailando++: 3D Dance GPT with Choreographic Memory, Li et al.
- (ArXiv 2023) UDE-2: A Unified Framework for Multimodal, Multi-Part Human Motion Synthesis, Zhou et al.
- (ArXiv 2023) Motion Script: MotionScript: Natural Language Descriptions for Expressive 3D Human Motions, Yazdian et al.
- (NeurIPS 2022) NeMF: NeMF: Neural Motion Fields for Kinematic Animation, He et al.
- (SIGGRAPH Asia 2022) PADL: PADL: Language-Directed Physics-Based Character, Juravsky et al.
- (SIGGRAPH Asia 2022) Rhythmic Gesticulator: Rhythmic Gesticulator: Rhythm-Aware Co-Speech Gesture Synthesis with Hierarchical Neural Embeddings, Ao et al.
- (3DV 2022) TEACH: TEACH: Temporal Action Composition for 3D Human, Athanasiou et al.
- (ECCV 2022) Implicit Motion: Implicit Neural Representations for Variable Length Human Motion Generation, Cervantes et al.
- (ECCV 2022) Zhong et al.: Learning Uncoupled-Modulation CVAE for 3D Action-Conditioned Human Motion Synthesis, Zhong et al.
- (ECCV 2022) MotionCLIP: MotionCLIP: Exposing Human Motion Generation to CLIP Space, Tevet et al.
- (ECCV 2022) PoseGPT: PoseGPT: Quantizing human motion for large scale generative modeling, Lucas et al.
- (ECCV 2022) TEMOS: TEMOS: Generating diverse human motions from textual descriptions, Petrovich et al.
- (ECCV 2022) TM2T: TM2T: Stochastic and Tokenized Modeling for the Reciprocal Generation of 3D Human Motions and Texts, Guo et al.
- (SIGGRAPH 2022) AvatarCLIP: AvatarCLIP: Zero-Shot Text-Driven Generation and Animation of 3D Avatars, Hong et al.
- (SIGGRAPH 2022) DeepPhase: Deepphase: Periodic autoencoders for learning motion phase manifolds, Starke et al.
- (CVPR 2022) Guo et al.: Generating Diverse and Natural 3D Human Motions from Text, Guo et al.
- (CVPR 2022) Bailando: Bailando: 3D Dance Generation by Actor-Critic GPT with Choreographic Memory, Li et al.
- (ICCV 2021) ACTOR: Action-Conditioned 3D Human Motion Synthesis with Transformer VAE, Petrovich et al.
- (ICCV 2021) AIST++: AI Choreographer: Music Conditioned 3D Dance Generation with AIST++, Li et al.
- (SIGGRAPH 2021) Starke et al.: Neural animation layering for synthesizing martial arts movements, Starke et al.
- (CVPR 2021) MOJO: We are More than Our Joints: Predicting how 3D Bodies Move, Zhang et al.
- (ECCV 2020) DLow: DLow: Diversifying Latent Flows for Diverse Human Motion Prediction, Yuan et al.
- (SIGGRAPH 2020) Starke et al.: Local motion phases for learning multi-contact character movements, Starke et al.
- (CVPR 2025) AnyMoLe: AnyMoLe: Any Character Motion In-Betweening Leveraging Video Diffusion Models, Yun et al.
- (ArXiv 2025) Dai et al: Towards Synthesized and Editable Motion In-Betweening Through Part-Wise Phase Representation, Dai et al.
- (SIGGRAPH Asia 2024) MotionFix: MotionFix: Text-Driven 3D Human Motion Editing, Athanasiou et al.
- (NeurIPS 2024) CigTime: CigTime: Corrective Instruction Generation Through Inverse Motion Editing, Fang et al.
- (SIGGRAPH 2024) Iterative Motion Editing: Iterative Motion Editing with Natural Language, Goel et al.
- (CVPR 2024) DNO: DNO: Optimizing Diffusion Noise Can Serve As Universal Motion Priors, Karunratanakul et al.
- (ArXiv 2024) MulSMo: MulSMo: Multimodal Stylized Motion Generation by Bidirectional Control Flow, Li et al.
- (TSMC 2024) D-LORD: D-LORD for Motion Stylization, Gupta et al.
- (ECCV 2024) HUMOS: HUMOS: Human Motion Model Conditioned on Body Shape, Tripathi et al.
- (SIGGRAPH 2024) SMEAR: SMEAR: Stylized Motion Exaggeration with ARt-direction, Basset et al.
- (CVPR 2024) MCM-LDM: Arbitrary Motion Style Transfer with Multi-condition Motion Latent Diffusion Model, Song et al.
- (CVPR 2024) MoST: MoST: Motion Style Transformer between Diverse Action Contents, Kim et al.
- (ICLR 2024) GenMoStyle: Generative Human Motion Stylization in Latent Space, Guo et al.
- (CVPR 2025) SemGeoMo: SemGeoMo: Dynamic Contextual Human Motion Generation with Semantic and Geometric Guidance, Cong et al.
- (CVPR 2025) Phys-Reach-Grasp: Learning Physics-Based Full-Body Human Reaching and Grasping from Brief Walking References, Li et al.
- (CVPR 2025) ParaHome: ParaHome: Parameterizing Everyday Home Activities Towards 3D Generative Modeling of Human-Object Interactions, Kim et al.
- (CVPR 2025) InterMimic: InterMimic: Towards Universal Whole-Body Control for Physics-Based Human-Object Interactions, Xu et al.
- (CVPR 2025) CORE4D: CORE4D: A 4D Human-Object-Human Interaction Dataset for Collaborative Object REarrangement, Zhang et al.
- (CVPR 2025) SkillMimic: SkillMimic: Learning Reusable Basketball Skills from Demonstrations, Wang et al.
- (CVPR 2025) MobileH2R: MobileH2R: Learning Generalizable Human to Mobile Robot Handover Exclusively from Scalable and Diverse Synthetic Data, Wang et al.
- (AAAI 2025) DiffGrasp: Diffgrasp: Whole-Body Grasping Synthesis Guided by Object Motion Using a Diffusion Model, Zhang et al.
- (3DV 2025) Paschalidis et al: 3D Whole-body Grasp Synthesis with Directional Controllability, Paschalidis et al.
- (3DV 2025) InterTrack: InterTrack: Tracking Human Object Interaction without Object Templates, Xie et al.
- (3DV 2025) FORCE: FORCE: Dataset and Method for Intuitive Physics Guided Human-object Interaction, Zhang et al.
- (ArXiv 2025) Kaiwu: Kaiwu: A Multimodal Manipulation Dataset and Framework for Robot Learning and Human-Robot Interaction, Jiang et al.
- (ArXiv 2025) EigenActor: EigenActor: Variant Body-Object Interaction Generation Evolved from Invariant Action Basis Reasoning, Guo et al.
- (ArXiv 2024) SyncDiff: SyncDiff: Synchronized Motion Diffusion for Multi-Body Human-Object Interaction Synthesis, He et al.
- (ArXiv 2024) CHOICE: CHOICE: Coordinated Human-Object Interaction in Cluttered Environments for Pick-and-Place Actions, Lu et al.
- (ArXiv 2024) TriDi: TriDi: Trilateral Diffusion of 3D Humans, Objects and Interactions, Petrov et al.
- (ArXiv 2024) OOD-HOI: OOD-HOI: Text-Driven 3D Whole-Body Human-Object Interactions Generation Beyond Training Domains, Zhang et al.
- (ArXiv 2024) COLLAGE: COLLAGE: Collaborative Human-Agent Interaction Generation using Hierarchical Latent Diffusion and Language Models, Daiya et al.
- (ArXiv 2024) SMGDiff: SMGDiff: Soccer Motion Generation using diffusion probabilistic models, Yang et al.
- (ArXiv 2024) Wu et al: Human-Object Interaction from Human-Level Instructions, Wu et al.
- (NeurIPS 2024) HumanVLA: HumanVLA: Towards Vision-Language Directed Object Rearrangement by Physical Humanoid, Xu et al.
- (NeurIPS 2024) OmniGrasp: Grasping Diverse Objects with Simulated Humanoids, Luo et al.
- (NeurIPS 2024) EgoChoir: EgoChoir: Capturing 3D Human-Object Interaction Regions from Egocentric Views, Yang et al.
- (NeurIPS 2024) CooHOI: CooHOI: Learning Cooperative Human-Object Interaction with Manipulated Object Dynamics, Gao et al.
- (NeurIPS 2024) InterDreamer: InterDreamer: Zero-Shot Text to 3D Dynamic Human-Object Interaction, Xu et al.
- (NeurIPS 2024) PiMForce: Posture-Informed Muscular Force Learning for Robust Hand Pressure Estimation, Seo et al.
- (ECCV 2024) InterFusion: InterFusion: Text-Driven Generation of 3D Human-Object Interaction, Dai et al.
- (ECCV 2024) CHOIS: Controllable Human-Object Interaction Synthesis, Li et al.
- (ECCV 2024) F-HOI: F-HOI: Toward Fine-grained Semantic-Aligned 3D Human-Object Interactions, Yang et al.
- (ECCV 2024) HIMO: HIMO: A New Benchmark for Full-Body Human Interacting with Multiple Objects, Lv et al.
- (SIGGRAPH 2024) PhysicsPingPong: Strategy and Skill Learning for Physics-based Table Tennis Animation, Wang et al.
- (CVPR 2024) NIFTY: NIFTY: Neural Object Interaction Fields for Guided Human Motion Synthesis, Kulkarni et al.
- (CVPR 2024) HOI Animator: HOIAnimator: Generating Text-Prompt Human-Object Animations using Novel Perceptive Diffusion Models, Son et al.
- (CVPR 2024) CG-HOI: CG-HOI: Contact-Guided 3D Human-Object Interaction Generation, Diller et al.
- (IJCV 2024) InterCap: InterCap: Joint Markerless 3D Tracking of Humans and Objects in Interaction, Huang et al.
- (3DV 2024) Phys-Fullbody-Grasp: Physically Plausible Full-Body Hand-Object Interaction Synthesis, Braun et al.
- (3DV 2024) GRIP: GRIP: Generating Interaction Poses Using Spatial Cues and Latent Consistency, Taheri et al.
- (AAAI 2024) FAVOR: Favor: Full-Body AR-driven Virtual Object Rearrangement Guided by Instruction Text, Li et al.
- (SIGGRAPH Asia 2023) OMOMO: Object Motion Guided Human Motion Synthesis, Li et al.
- (ICCV 2023) CHAIRS: Full-Body Articulated Human-Object Interaction, Jiang et al.
- (ICCV 2023) HGHOI: Hierarchical Generation of Human-Object Interactions with Diffusion Probabilistic Models, Pi et al.
- (ICCV 2023) InterDiff: InterDiff: Generating 3D Human-Object Interactions with Physics-Informed Diffusion, Xu et al.
- (CVPR 2023) Object Pop Up: Object pop-up: Can we infer 3D objects and their poses from human interactions alone? Petrov et al.
- (CVPR 2023) ARCTIC: A Dataset for Dexterous Bimanual Hand-Object Manipulation, Fan et al.
- (ECCV 2022) TOCH: TOCH: Spatio-Temporal Object-to-Hand Correspondence for Motion Refinement, Zhou et al.
- (ECCV 2022) COUCH: COUCH: Towards Controllable Human-Chair Interactions, Zhang et al.
- (ECCV 2022) SAGA: SAGA: Stochastic Whole-Body Grasping with Contact, Wu et al.
- (CVPR 2022) GOAL: GOAL: Generating 4D Whole-Body Motion for Hand-Object Grasping, Taheri et al.
- (CVPR 2022) BEHAVE: BEHAVE: Dataset and Method for Tracking Human Object Interactions, Bhatnagar et al.
- (ECCV 2020) GRAB: GRAB: A Dataset of Whole-Body Human Grasping of Objects, Taheri et al.
- (CVPR 2025) EnvPoser: EnvPoser: Environment-aware Realistic Human Motion Estimation from Sparse Observations with Uncertainty Modeling. Xia et al.
- (ICLR 2025) Sitcom-Crafter: Sitcom-Crafter: A Plot-Driven Human Motion Generation System in 3D Scenes, Chen et al.
- (3DV 2025) Paschalidis et al: 3D Whole-body Grasp Synthesis with Directional Controllability, Paschalidis et al.
- (WACV 2025) GHOST: GHOST: Grounded Human Motion Generation with Open Vocabulary Scene-and-Text Contexts, Milacski et al.
- (ArXiv 2025) Gao et al: Jointly Understand Your Command and Intention: Reciprocal Co-Evolution between Scene-Aware 3D Human Motion Synthesis and Analysis, Gao et al.
- (ArXiv 2024) ZeroHSI: ZeroHSI: Zero-Shot 4D Human-Scene Interaction by Video Generation, Li et al.
- (ArXiv 2024) Mimicking-Bench: Mimicking-Bench: A Benchmark for Generalizable Humanoid-Scene Interaction Learning via Human Mimicking, Liu et al.
- (ArXiv 2024) SCENIC: SCENIC: Scene-aware Semantic Navigation with Instruction-guided Control, Zhang et al.
- (ArXiv 2024) Diffusion Implicit Policy: Diffusion Implicit Policy for Unpaired Scene-aware Motion synthesis, Gong et al.
- (ArXiv 2024) SIMS: SIMS: Simulating Human-Scene Interactions with Real World Script Planning, Wang et al.
- (ArXiv 2024) LaserHuman: LaserHuman: Language-guided Scene-aware Human Motion Generation in Free Environment, Cong et al.
- (SIGGRAPH Asia 2024) LINGO: Autonomous Character-Scene Interaction Synthesis from Text Instruction, Jiang et al.
- (NeurIPS 2024) DiMoP3D: Harmonizing Stochasticity and Determinism: Scene-responsive Diverse Human Motion Prediction, Lou et al.
- (ECCV 2024) Liu et al.: Revisit Human-Scene Interaction via Space Occupancy, Liu et al.
- (ECCV 2024) TesMo: Generating Human Interaction Motions in Scenes with Text Control, Yi et al.
- (ECCV 2024 Workshop) SAST: Massively Multi-Person 3D Human Motion Forecasting with Scene Context, Mueller et al.
- (Eurographics 2024) Kang et al: Learning Climbing Controllers for Physics-Based Characters, Kang et al.
- (CVPR 2024) Afford-Motion: Move as You Say, Interact as You Can: Language-guided Human Motion Generation with Scene Affordance, Wang et al.
- (CVPR 2024) GenZI: GenZI: Zero-Shot 3D Human-Scene Interaction Generation, Li et al.
- (CVPR 2024) Cen et al.: Generating Human Motion in 3D Scenes from Text Descriptions, Cen et al.
- (CVPR 2024) TRUMANS: Scaling Up Dynamic Human-Scene Interaction Modeling, Jiang et al.
- (ICLR 2024) UniHSI: UniHSI: Unified Human-Scene Interaction via Prompted Chain-of-Contacts, Xiao et al.
- (3DV 2024) Purposer: Purposer: Putting Human Motion Generation in Context, Ugrinovic et al.
- (3DV 2024) InterScene: Synthesizing Physically Plausible Human Motions in 3D Scenes, Pan et al.
- (3DV 2024) Mir et al: Generating Continual Human Motion in Diverse 3D Scenes, Mir et al.
- (ICCV 2023) DIMOS: DIMOS: Synthesizing Diverse Human Motions in 3D Indoor Scenes, Zhao et al.
- (ICCV 2023) LAMA: Locomotion-Action-Manipulation: Synthesizing Human-Scene Interactions in Complex 3D Environments, Lee et al.
- (ICCV 2023) Narrator: Narrator: Towards Natural Control of Human-Scene Interaction Generation via Relationship Reasoning, Xuan et al.
- (CVPR 2023) CIMI4D: CIMI4D: A Large Multimodal Climbing Motion Dataset under Human-Scene Interactions, Yan et al.
- (CVPR 2023) Scene-Ego: Scene-aware Egocentric 3D Human Pose Estimation, Wang et al.
- (CVPR 2023) SLOPER4D: SLOPER4D: A Scene-Aware Dataset for Global 4D Human Pose Estimation in Urban Environments, Dai et al.
- (CVPR 2023) CIRCLE: CIRCLE: Capture in Rich Contextual Environments, Araujo et al.
- (CVPR 2023) SceneDiffuser: Diffusion-based Generation, Optimization, and Planning in 3D Scenes, Huang et al.
- (SIGGRAPH 2023) PMP: PMP: Learning to Physically Interact with Environments using Part-wise Motion Priors, Bae et al.
- (SIGGRAPH 2023) QuestEnvSim: QuestEnvSim: Environment-Aware Simulated Motion Tracking from Sparse Sensors, Lee et al.
- (SIGGRAPH 2023) Hassan et al.: Synthesizing Physical Character-Scene Interactions, Hassan et al.
- (NeurIPS 2022) Mao et al.: Contact-Aware Human Motion Forecasting, Mao et al.
- (NeurIPS 2022) HUMANISE: HUMANISE: Language-conditioned Human Motion Generation in 3D Scenes, Wang et al.
- (NeurIPS 2022) EmbodiedPose: Embodied Scene-aware Human Pose Estimation, Luo et al.
- (ECCV 2022) GIMO: GIMO: Gaze-Informed Human Motion Prediction in Context, Zheng et al.
- (ECCV 2022) COINS: COINS: Compositional Human-Scene Interaction Synthesis with Semantic Control, Zhao et al.
- (CVPR 2022) Wang et al.: Towards Diverse and Natural Scene-aware 3D Human Motion Synthesis, Wang et al.
- (CVPR 2022) GAMMA: The Wanderings of Odysseus in 3D Scenes, Zhang et al.
- (ICCV 2021) SAMP: Stochastic Scene-Aware Motion Prediction, Hassan et al.
- (ICCV 2021) LEMO: Learning Motion Priors for 4D Human Body Capture in 3D Scenes, Zhang et al.
- (3DV 2020) PLACE: PLACE: Proximity Learning of Articulation and Contact in 3D Environments, Zhang et al.
- (SIGGRAPH 2020) Starke et al.: Local motion phases for learning multi-contact character movements, Starke et al.
- (CVPR 2020) PSI: Generating 3D People in Scenes without People, Zhang et al.
- (SIGGRAPH Asia 2019) NSM: Neural State Machine for Character-Scene Interactions, Starke et al.
- (ICCV 2019) PROX: Resolving 3D Human Pose Ambiguities with 3D Scene Constraints, Hassan et al.
- (CVPR 2025) TIMotion: TIMotion: Temporal and Interactive Framework for Efficient Human-Human Motion Generation, Wang et al.
- (ICLR 2025) Think Then React: Think Then React: Towards Unconstrained Action-to-Reaction Motion Generation, Tan et al.
- (ICLR 2025) Ready-to-React: Ready-to-React: Online Reaction Policy for Two-Character Interaction Generation, Cen et al.
- (ICLR 2025) InterMask: InterMask: 3D Human Interaction Generation via Collaborative Masked Modelling, Javed et al.
- (3DV 2025) Interactive Humanoid: Interactive Humanoid: Online Full-Body Motion Reaction Synthesis with Social Affordance Canonicalization and Forecasting, Liu et al.
- (ArXiv 2025) Invisible Strings: Invisible Strings: Revealing Latent Dancer-to-Dancer Interactions with Graph Neural Networks, Zerkowski et al.
- (ArXiv 2025) Leader and Follower: Leader and Follower: Interactive Motion Generation under Trajectory Constraints, Wang et al.
- (ArXiv 2024) Two in One: Two-in-One: Unified Multi-Person Interactive Motion Generation by Latent Diffusion Transformer, Li et al.
- (ArXiv 2024) It Takes Two: It Takes Two: Real-time Co-Speech Two-person’s Interaction Generation via Reactive Auto-regressive Diffusion Model, Shi et al.
- (ArXiv 2024) COLLAGE: COLLAGE: Collaborative Human-Agent Interaction Generation using Hierarchical Latent Diffusion and Language Models, Daiya et al.
- (NeurIPS 2024) Harmony4D: Harmony4D: A Video Dataset for In-The-Wild Close Human Interactions, Khirodkar et al.
- (NeurIPS 2024) InterControl: InterControl: Generate Human Motion Interactions by Controlling Every Joint, Wang et al.
- (ACM MM 2024) PhysReaction: PhysReaction: Physically Plausible Real-Time Humanoid Reaction Synthesis via Forward Dynamics Guided 4D Imitation, Liu et al.
- (ECCV 2024) Shan et al: Towards Open Domain Text-Driven Synthesis of Multi-Person Motions, Shan et al.
- (ECCV 2024) ReMoS: ReMoS: 3D Motion-Conditioned Reaction Synthesis for Two-Person Interactions, Ghosh et al.
- (CVPR 2024) Inter-X: Inter-X: Towards Versatile Human-Human Interaction Analysis, Xu et al.
- (CVPR 2024) ReGenNet: ReGenNet: Towards Human Action-Reaction Synthesis, Xu et al.
- (CVPR 2024) Fang et al.: Capturing Closely Interacted Two-Person Motions with Reaction Priors, Fan et al.
- (CVPR Workshop 2024) in2IN: in2IN: Leveraging Individual Information to Generate Human INteractions, Ruiz-Ponce et al.
- (IJCV 2024) InterGen: InterGen: Diffusion-based Multi-human Motion Generation under Complex Interactions, Liang et al.
- (ICCV 2023) ActFormer: ActFormer: A GAN-based Transformer towards General Action-Conditioned 3D Human Motion Generation, Xu et al.
- (ICCV 2023) Tanaka et al.: Role-aware Interaction Generation from Textual Description, Tanaka et al.
- (CVPR 2023) Hi4D: Hi4D: 4D Instance Segmentation of Close Human Interaction, Yin et al.
- (CVPR 2022) ExPI: Multi-Person Extreme Motion Prediction, Guo et al.
- (CVPR 2020) CHI3D: Three-Dimensional Reconstruction of Human Interactions, Fieraru et al.
- (CVPR 2025) AtoM: AToM: Aligning Text-to-Motion Model at Event-Level with GPT-4Vision Reward, Han et al.
- (CVPR 2025) CORE4D: CORE4D: A 4D Human-Object-Human Interaction Dataset for Collaborative Object REarrangement, Zhang et al.
- (ICLR 2025) MotionCritic: Aligning Human Motion Generation with Human Perceptions, Wang et al.
- (ICLR 2025) LocoVR: LocoVR: Multiuser Indoor Locomotion Dataset in Virtual Reality, Takeyama et al.
- (ICLR 2025) PMR: Pedestrian Motion Reconstruction: A Large-scale Benchmark via Mixed Reality Rendering with Multiple Perspectives and Modalities, Wang et al.
- (AAAI 2025) EMHI: EMHI: A Multimodal Egocentric Human Motion Dataset with HMD and Body-Worn IMUs, Fan et al.
- (ArXiv 2025) SGA-INTERACT: SGA-INTERACT: A3DSkeleton-based Benchmark for Group Activity Understanding in Modern Basketball Tactic, Yang et al.
- (ArXiv 2025) Kaiwu: Kaiwu: A Multimodal Manipulation Dataset and Framework for Robot Learning and Human-Robot Interaction, Jiang et al.
- (ArXiv 2025) Motion-X++: Motion-X++: A Large-Scale Multimodal 3D Whole-body Human Motion Dataset, Zhang et al.
- (ArXiv 2024) Mimicking-Bench: Mimicking-Bench: A Benchmark for Generalizable Humanoid-Scene Interaction Learning via Human Mimicking, Liu et al.
- (ArXiv 2024) LaserHuman: LaserHuman: Language-guided Scene-aware Human Motion Generation in Free Environment, Cong et al.
- (ArXiv 2024) SCENIC: SCENIC: Scene-aware Semantic Navigation with Instruction-guided Control, Zhang et al.
- (ArXiv 2024) synNsync: Synergy and Synchrony in Couple Dances, Manukele et al.
- (ArXiv 2024) MotionBank: MotionBank: A Large-scale Video Motion Benchmark with Disentangled Rule-based Annotations, Xu et al.
- (Github 2024) CMP & CMR: AnimationGPT: An AIGC tool for generating game combat motion assets, Liao et al.
- (Scientific Data 2024) Evans et al: Synchronized Video, Motion Capture and Force Plate Dataset for Validating Markerless Human Movement Analysis, Evans et al.
- (SIGGRAPH Asia 2024) LINGO: Autonomous Character-Scene Interaction Synthesis from Text Instruction, Jiang et al.
- (NeurIPS 2024) Harmony4D: Harmony4D: A Video Dataset for In-The-Wild Close Human Interactions, Khirodkar et al.
- (NeurIPS D&B 2024) EgoSim: EgoSim: An Egocentric Multi-view Simulator for Body-worn Cameras during Human Motion, Hollidt et al.
- (NeurIPS D&B 2024) Muscles in Time: Muscles in Time: Learning to Understand Human Motion by Simulating Muscle Activations, Schneider et al.
- (NeurIPS D&B 2024) Text to blind motion: Text to blind motion, Kim et al.
- (ACM MM 2024) CLaM: CLaM: An Open-Source Library for Performance Evaluation of Text-driven Human Motion Generation, Chen et al.
- (ECCV 2024) AddBiomechanics: AddBiomechanics Dataset: Capturing the Physics of Human Motion at Scale, Werling et al.
- (ECCV 2024) LiveHPS++: LiveHPS++: Robust and Coherent Motion Capture in Dynamic Free Environment, Ren et al.
- (ECCV 2024) SignAvatars: SignAvatars: A Large-scale 3D Sign Language Holistic Motion Dataset and Benchmark, Yu et al.
- (ECCV 2024) Nymeria: Nymeria: A massive collection of multimodal egocentric daily motion in the wild, Ma et al.
- (Multibody System Dynamics 2024) Human3.6M+: Using musculoskeletal models to generate physically-consistent data for 3D human pose, kinematic, dynamic, and muscle estimation, Nasr et al.
- (CVPR 2024) Inter-X: Inter-X: Towards Versatile Human-Human Interaction Analysis, Xu et al.
- (CVPR 2024) HardMo: HardMo: ALarge-Scale Hardcase Dataset for Motion Capture, Liao et al.
- (CVPR 2024) MMVP: MMVP: A Multimodal MoCap Dataset with Vision and Pressure Sensors, Zhang et al.
- (CVPR 2024) RELI11D: RELI11D: A Comprehensive Multimodal Human Motion Dataset and Method, Yan et al.
- (SIGGRAPH Asia 2023) GroundLink: GroundLink: A Dataset Unifying Human Body Movement and Ground Reaction Dynamics, Han et al.
- (NeurIPS D&B 2023) HOH: HOH: Markerless Multimodal Human-Object-Human Handover Dataset with Large Object Count, Wiederhold et al.
- (NeurIPS D&B 2023) Motion-X: Motion-X: A Large-scale 3D Expressive Whole-body Human Motion Dataset, Lin et al.
- (NeurIPS D&B 2023) Humans in Kitchens: Humans in Kitchens: A Dataset for Multi-Person Human Motion Forecasting with Scene Context, Tanke et al.
- (ICCV 2023) CHAIRS: Full-Body Articulated Human-Object Interaction, Jiang et al.
- (CVPR 2023) CIMI4D: CIMI4D: A Large Multimodal Climbing Motion Dataset under Human-Scene Interactions, Yan et al.
- (CVPR 2023) FLAG3D: FLAG3D: A 3D Fitness Activity Dataset with Language Instruction, Tang et al.
- (CVPR 2023) Hi4D: Hi4D: 4D Instance Segmentation of Close Human Interaction, Yin et al.
- (CVPR 2023) CIRCLE: CIRCLE: Capture in Rich Contextual Environments, Araujo et al.
- (NeurIPS 2022) MoCapAct: MoCapAct: A Multi-Task Dataset for Simulated Humanoid Control, Wagener et al.
- (ACM MM 2022) ForcePose: Learning to Estimate External Forces of Human Motion in Video, Louis et al.
- (ECCV 2022) BEAT: BEAT: A Large-Scale Semantic and Emotional Multi-Modal Dataset for Conversational Gestures Synthesis, Liu et al.
- (ECCV 2022) BRACE: BRACE: The Breakdancing Competition Dataset for Dance Motion Synthesis, Moltisanti et al.
- (ECCV 2022) EgoBody: Egobody: Human body shape and motion of interacting people from head-mounted devices, Zhang et al.
- (ECCV 2022) GIMO: GIMO: Gaze-Informed Human Motion Prediction in Context, Zheng et al.
- (ECCV 2022) HuMMan: HuMMan: Multi-Modal 4D Human Dataset for Versatile Sensing and Modeling, Cai et al.
- (CVPR 2022) ExPI: Multi-Person Extreme Motion Prediction, Guo et al.
- (CVPR 2022) HumanML3D: Generating Diverse and Natural 3D Human Motions from Text, Guo et al.
- (CVPR 2022) BEHAVE: BEHAVE: Dataset and Method for Tracking Human Object Interactions, Bhatnagar et al.
- (ICCV 2021) AIST++: AI Choreographer: Music Conditioned 3D Dance Generation with AIST++, Li et al.
- (CVPR 2021) Fit3D: AIFit: Automatic 3D Human-Interpretable Feedback Models for Fitness Training, Fieraru et al.
- (CVPR 2021) BABEL: BABEL: Bodies, Action, and Behavior with English Labels, Punnakkal et al.
- (AAAI 2021) HumanSC3D: Learning complex 3d human self-contact, Fieraru et al.
- (CVPR 2020) CHI3D: Three-Dimensional Reconstruction of Human Interactions, Fieraru et al.
- (ICCV 2019) PROX: Resolving 3D Human Pose Ambiguities with 3D Scene Constraints, Hassan et al.
- (ICCV 2019) AMASS: AMASS: Archive of Motion Capture As Surface Shapes, Mahmood et al.
- (CVPR 2025) InterMimic: InterMimic: Towards Universal Whole-Body Control for Physics-Based Human-Object Interactions, Xu et al.
- (CVPR 2025) SkillMimic: SkillMimic: Learning Reusable Basketball Skills from Demonstrations, Wang et al.
- (ICRA 2025) HOVER: HOVER: Versatile Neural Whole-Body Controller for Humanoid Robots, He et al.
- (ICRA 2025) PIM: Learning Humanoid Locomotion with Perceptive Internal Model, Long et al.
- (ICRA 2025) Think on your feet: Think on your feet: Seamless Transition between Human-like Locomotion in Response to Changing Commands, Huang et al.
- (ICLR 2025) Puppeteer: Hierarchical World Models as Visual Whole-Body Humanoid Controllers, Hansen et al.
- (ICLR 2025) FB-CPR: Zero-Shot Whole-Body Humanoid Control via Behavioral Foundation Models, Tirinzoni et al.
- (ICLR 2025) MPC2: Motion Control of High-Dimensional Musculoskeletal System with Hierarchical Model-Based Planning, Wei et al.
- (ICLR 2025) CLoSD: CLoSD: Closing the Loop between Simulation and Diffusion for multi-task character control, Tevet et al.
- (ICLR 2025) HiLo: HiLo: Learning Whole-Body Human-like Locomotion with Motion Tracking Controller, Zhang et al.
- (Github 2025) MobilityGen: MobilityGen.
- (ArXiv 2025) GMP: Natural Humanoid Robot Locomotion with Generative Motion Prior, Zhang et al.
- (ArXiv 2025) Sun et al: Learning Perceptive Humanoid Locomotion over Challenging Terrain, Sun et al.
- (ArXiv 2025) HWC-Loco: HWC-Loco: AHierarchical Whole-Body Control Approach to Robust Humanoid Locomotion, Lin et al.
- (ArXiv 2025) Lin et al: Sim-to-Real Reinforcement Learning for Vision-Based Dexterous Manipulation on Humanoids, Lin et al.
- (ArXiv 2025) COMPASS: COMPASS: Cross-embOdiment Mobility Policy via ResiduAl RL and Skill Synthesis, Liu et al.
- (ArXiv 2025) VB-COM: VB-Com: Learning Vision-Blind Composite Humanoid Locomotion Against Deficient Perception, Ren et al.
- (ArXiv 2025) ModSkill: ModSkill: Physical Character Skill Modularization, Huang et al.
- (ArXiv 2025) Humanoid-VLA: Humanoid-VLA: Towards Universal Humanoid Control with Visual Integration, Ding et al.
- (ArXiv 2025) Li et al: Human-Like Robot Impedance Regulation Skill Learning from Human-Human Demonstrations, Li et al.
- (ArXiv 2025) HumanUP: Learning Getting-Up Policies for Real-World Humanoid Robots, He et al.
- (ArXiv 2025) RHINO: RHINO: Learning Real-Time Humanoid-Human-Object Interaction from Human Demonstrations, Chen et al.
- (ArXiv 2025) HOMIE: HOMIE: Humanoid Loco-Manipulation with Isomorphic Exoskeleton Cockpit, Ben et al.
- (ArXiv 2025) BeamDojo: BeamDojo: Learning Agile Humanoid Locomotion on Sparse Footholds, Wang et al.
- (ArXiv 2025) HoST: Learning Humanoid Standing-up Control across Diverse Postures, Huang et al.
- (ArXiv 2025) Embrace Collisions: Embrace Collisions: Humanoid Shadowing for Deployable Contact-Agnostics Motion, Zhuang et al.
- (ArXiv 2025) ToddlerBot: ToddlerBot: Open-Source ML-Compatible Humanoid Platform for Loco-Manipulation, Shi et al.
- (ArXiv 2025) ASAP: ASAP: Aligning Simulation and Real-World Physics for Learning Agile Humanoid Whole-Body Skills, He et al.
- (ArXiv 2025) Gu et al: Humanoid Locomotion and Manipulation: Current Progress and Challenges in Control, Planning, and Learning, Gu et al.
- (ArXiv 2024) UH-1: Learning from Massive Human Videos for Universal Humanoid Pose Control, Mao et al.
- (ArXiv 2024) Mimicking-Bench: Mimicking-Bench: A Benchmark for Generalizable Humanoid-Scene Interaction Learning via Human Mimicking, Liu et al.
- (ArXiv 2024) Exbody2: Exbody2: Advanced Expressive Humanoid Whole-Body Control, Ji et al.
- (ArXiv 2024) SIMS: SIMS: Simulating Human-Scene Interactions with Real World Script Planning, Wang et al.
- (ArXiv 2024) Humanoidlympics: Humanoidlympics: Sports Environments for Physically Simulated Humanoids, Luo et al.
- (ArXiv 2024) PhySHOI: PhysHOI: Physics-Based Imitation of Dynamic Human-Object Interaction, Wang et al.
- (RA-L 2024) Liu et al: Human-Humanoid Robots Cross-Embodiment Behavior-Skill Transfer Using Decomposed Adversarial Learning from Demonstration, Liu et al.
- (SIGGRAPH Asia 2024) PDP: PDP: Physics-Based Character Animation via Diffusion Policy, Truong et al.
- (SIGGRAPH Asia 2024) MaskedMimic: MaskedMimic: Unified Physics-Based Character Control Through Masked Motion Inpainting, Tessler et al.
- (NeurIPS 2024) HumanVLA: HumanVLA: Towards Vision-Language Directed Object Rearrangement by Physical Humanoid, Xu et al.
- (NeurIPS 2024) OmniGrasp: Grasping Diverse Objects with Simulated Humanoids, Luo et al.
- (NeurIPS 2024) InterControl: InterControl: Generate Human Motion Interactions by Controlling Every Joint, Wang et al.
- (NeurIPS 2024) CooHOI: CooHOI: Learning Cooperative Human-Object Interaction with Manipulated Object Dynamics, Gao et al.
- (NeurIPS 2024) Radosavovic et al.: Humanoid Locomotion as Next Token Prediction, Radosavovic et al.
- (CoRL 2024) HARMON: Harmon: Whole-Body Motion Generation of Humanoid Robots from Language Descriptions, Jiang et al.
- (CoRL 2024) OKAMI: OKAMI: Teaching Humanoid Robots Manipulation Skills through Single Video Imitation, Li et al.
- (CoRL 2024) HumanPlus: HumanPlus: Humanoid Shadowing and Imitation from Humans, Fu et al.
- (CoRL 2024) OmniH2O: OmniH2O: Universal and Dexterous Human-to-Humanoid Whole-Body Teleoperation and Learning, He et al.
- (Humanoids 2024) Self-Aware: Know your limits! Optimize the behavior of bipedal robots through self-awareness, Mascaro et al.
- (ACM MM 2024) PhysReaction: PhysReaction: Physically Plausible Real-Time Humanoid Reaction Synthesis via Forward Dynamics Guided 4D Imitation, Liu et al.
- (IROS 2024) H2O: Learning Human-to-Humanoid Real-Time Whole-Body Teleoperation, He et al.
- (ECCV 2024) MHC: Generating Physically Realistic and Directable Human Motions from Multi-Modal Inputs, Shrestha et al.
- (ICML 2024) DIDI: DIDI: Diffusion-Guided Diversity for Offline Behavioral Generation, Liu et al.
- (SIGGRAPH 2024) MoConVQ: MoConVQ: Unified Physics-Based Motion Control via Scalable Discrete Representations, Yao et al.
- (SIGGRAPH 2024) PhysicsPingPong: Strategy and Skill Learning for Physics-based Table Tennis Animation, Wang et al.
- (SIGGRAPH 2024) SuperPADL: SuperPADL: Scaling Language-Directed Physics-Based Control with Progressive Supervised Distillation, Juravsky et al.
- (CVPR 2024) SimXR: Real-Time Simulated Avatar from Head-Mounted Sensors, Luo et al.
- (CVPR 2024) AnySkill: AnySkill: Learning Open-Vocabulary Physical Skill for Interactive Agents, Cui et al.
- (ICLR 2024) PULSE: Universal Humanoid Motion Representations for Physics-Based Control, Luo et al.
- (ICLR 2024) H-GAP: H-GAP: Humanoid Control with a Generalist Planner, Jiang et al.
- (ICLR 2024) UniHSI: UniHSI: Unified Human-Scene Interaction via Prompted Chain-of-Contacts, Xiao et al.
- (3DV 2024) Phys-Fullbody-Grasp: Physically Plausible Full-Body Hand-Object Interaction Synthesis, Braun et al.
- (RSS 2024) ExBody: Expressive Whole-Body Control for Humanoid Robots, Cheng et al.
- (SIGGRAPH Asia 2023) Fatigued Movements: Discovering Fatigued Movements for Virtual Character Animation, Cheema et al.
- (SIGGRAPH Asia 2023) CASE: C·ASE: Learning Conditional Adversarial Skill Embeddings for Physics-based Characters, Dou et al.
- (SIGGRAPH Asia 2023) AdaptNet: AdaptNet: Policy Adaptation for Physics-Based Character Control, Xu et al.
- (SIGGRAPH Asia 2023) NCP: Neural Categorical Priors for Physics-Based Character Control, Zhu et al.
- (SIGGRAPH Asia 2023) DROP: DROP: Dynamics Responses from Human Motion Prior and Projective Dynamics, Jiang et al.
- (NeurIPS 2023) InsActor: InsActor: Instruction-driven Physics-based Characters, Ren et al.
- (CoRL 2023) Humanoid4Parkour: Humanoid Parkour Learning, Zhuang et al.
- (CoRL Workshop 2023) Words into Action: Words into Action: Learning Diverse Humanoid Robot Behaviors using Language Guided Iterative Motion Refinement, Kumar et al.
- (ICCV 2023) PHC: Perpetual Humanoid Control for Real-time Simulated Avatars, Luo et al.
- (CVPR 2023) Trace and Pace: Trace and Pace: Controllable Pedestrian Animation via Guided Trajectory Diffusion, Rempe et al.
- (SIGGRAPH 2023) Vid2Player3D: DiffMimic: Efficient Motion Mimicking with Differentiable Physics, Zhang et al.
- (SIGGRAPH 2023) QuestEnvSim: QuestEnvSim: Environment-Aware Simulated Motion Tracking from Sparse Sensors, Lee et al.
- (SIGGRAPH 2023) Hassan et al.: Synthesizing Physical Character-Scene Interactions, Hassan et al.
- (SIGGRAPH 2023) CALM: CALM: Conditional Adversarial Latent Models for Directable Virtual Characters, Tessler et al.
- (SIGGRAPH 2023) Composite Motion: Composite Motion Learning with Task Control, Xu et al.
- (ICLR 2023) DiffMimic: DiffMimic: Efficient Motion Mimicking with Differentiable Physics, Ren et al.
- (NeurIPS 2022) EmbodiedPose: Embodied Scene-aware Human Pose Estimation, Luo et al.
- (NeurIPS 2022) MoCapAct: MoCapAct: A Multi-Task Dataset for Simulated Humanoid Control, Wagener et al.
- (SIGGRAPH Asia 2022) Gopinath et al.: Motion In-betweening for Physically Simulated Characters, Gopinath et al.
- (SIGGRAPH Asia 2022) AIP: AIP: Adversarial Interaction Priors for Multi-Agent Physics-based Character Control, Younes et al.
- (SIGGRAPH Asia 2022) ControlVAE: ControlVAE: Model-Based Learning of Generative Controllers for Physics-Based Characters, Yao et al.
- (SIGGRAPH Asia 2022) QuestSim: QuestSim: Human Motion Tracking from Sparse Sensors with Simulated Avatars, Winkler et al.
- (SIGGRAPH Asia 2022) PADL: PADL: Language-Directed Physics-Based Character, Juravsky et al.
- (SIGGRAPH Asia 2022) Wang et al.: Differentiable Simulation of Inertial Musculotendons, Wang et al.
- (SIGGRAPH 2022) ASE: ASE: Large-Scale Reusable Adversarial Skill Embeddings for Physically Simulated Characters, Peng et al.
- (Journal of Neuro-Engineering and Rehabilitation 2021) Learn to Move: Deep Reinforcement Learning for Modeling Human Locomotion Control in Neuromechanical Simulation, Peng et al.
- (NeurIPS 2021) KinPoly: Dynamics-Regulated Kinematic Policy for Egocentric Pose Estimation, Luo et al.
- (SIGGRAPH 2021) AMP: AMP: Adversarial Motion Priors for Stylized Physics-Based Character Control, Peng et al.
- (CVPR 2021) SimPoE: SimPoE: Simulated Character Control for 3D Human Pose Estimation, Yuan et al.
- (NeurIPS 2020) RFC: Residual Force Control for Agile Human Behavior Imitation and Extended Motion Synthesis, Yuan et al.
- (ICLR 2020) Yuan et al.: Diverse Trajectory Forecasting with Determinantal Point Processes, Yuan et al.
- (ICCV 2019) Ego-Pose: Ego-Pose Estimation and Forecasting as Real-Time PD Control, Yuan et al.
- (SIGGRAPH 2018) DeepMimic: DeepMimic: Example-Guided Deep Reinforcement Learning of Physics-Based Character Skills, Peng et al.
- (CVPR 2025) HDyS: Homogeneous Dynamics Space for Heterogeneous Humans, Liu et al.
- (ICLR 2025) ImDy: ImDy: Human Inverse Dynamics from Imitated Observations, Liu et al.
- (ICLR 2025) MPC2: Motion Control of High-Dimensional Musculoskeletal System with Hierarchical Model-Based Planning, Wei et al.
- (ACM Sensys 2025) SHADE-AD: SHADE-AD: An LLM-Based Framework for Synthesizing Activity Data of Alzheimer’s Patients, Fu et al.
- (JEB 2025) McAllister et al: Behavioural energetics in human locomotion: how energy use influences how we move, McAllister et al.
- (WACV 2025) OpenCapBench: A Benchmark to Bridge Pose Estimation and Biomechanics, Gozlan et al.
- (ArXiv 2025) Wu et al: Muscle Activation Estimation by Optimizing the Musculoskeletal Model for Personalized Strength and Conditioning Training, Wu et al.
- (ArXiv 2025) Cotton et al: Biomechanical Reconstruction with Confidence Intervals from Multiview Markerless Motion Capture, Cotton et al.
- (BiorXiv 2024) Lai et al: Mapping Grip Force to Muscular Activity Towards Understanding Upper Limb Musculoskeletal Intent using a Novel Grip Strength Model, Lai et al.
- (IROS 2024) Shahriari et al: Enhancing Robustness in Manipulability Assessment: The Pseudo-Ellipsoid Approach, Shahriari et al.
- (SIGGRAPH Asia 2024) BioDesign: Motion-Driven Neural Optimizer for Prophylactic Braces Made by Distributed Microstructures, Han et al.
- (Scientific Data 2024) Evans et al: Synchronized Video, Motion Capture and Force Plate Dataset for Validating Markerless Human Movement Analysis, Evans et al.
- (NeurIPS D&B 2024) Muscles in Time: Muscles in Time: Learning to Understand Human Motion by Simulating Muscle Activations, Schneider et al.
- (CoRL 2024) Wei et al: Safe Bayesian Optimization for the Control of High-Dimensional Embodied Systems, Wei et al.
- (HFES 2024) Macwan et al: High-Fidelity Worker Motion Simulation With Generative AI, Macwan et al.
- (ECCV 2024) AddBiomechanics: AddBiomechanics Dataset: Capturing the Physics of Human Motion at Scale, Werling et al.
- (ECCV 2024) MANIKIN: MANIKIN: Biomechanically Accurate Neural Inverse Kinematics for Human Motion Estimation, Jiang et al.
- (TOG 2024) NICER: NICER: A New and Improved Consumed Endurance and Recovery Metric to Quantify Muscle Fatigue of Mid-Air Interactions, Li et al.
- (ICML 2024) DynSyn: DynSyn: Dynamical Synergistic Representation for Efficient Learning and Control in Overactuated Embodied Systems, He et al.
- (Multibody System Dynamics 2024) Human3.6M+: Using musculoskeletal models to generate physically-consistent data for 3D human pose, kinematic, dynamic, and muscle estimation, Nasr et al.
- (CVPR 2024) HIT: HIT: Estimating Internal Human Implicit Tissues from the Body Surface, Keller et al.
- (Frontiers in Neuroscience 2024) Dai et al: Full-body pose reconstruction and correction in virtual reality for rehabilitation training, Dai et al.
- (ICRA 2024) He et al.: Self Model for Embodied Intelligence: Modeling Full-Body Human Musculoskeletal System and Locomotion Control with Hierarchical Low-Dimensional Representation, He et al.
- (SIGGRAPH Asia 2023) Fatigued Movements: Discovering Fatigued Movements for Virtual Character Animation, Cheema et al.
- (SIGGRAPH Asia 2023) SKEL: From skin to skeleton: Towards biomechanically accurate 3d digital humans, Keller et al.
- (SIGGRAPH Asia 2023) MuscleVAE: MuscleVAE: Model-Based Controllers of Muscle-Actuated Characters, Feng et al.
- (SIGGRAPH 2023) Bidirectional GaitNet: Bidirectional GaitNet, Park et al.
- (SIGGRAPH 2023) Lee et al.: Anatomically Detailed Simulation of Human Torso, Lee et al.
- (ICCV 2023) MiA: Muscles in Action, Chiquer et al.
- (CVPR 2022) OSSO: OSSO: Obtaining Skeletal Shape from Outside, Keller et al.
- (Scientific Data 2022) Xing et al: Functional movement screen dataset collected with two Azure Kinect depth sensors, Xing et al.
- (NCA 2020) Zell et al: Learning inverse dynamics for human locomotion analysis, Zell et al.
- (ECCV 2020) Zell et al: Weakly-supervised learning of human dynamics, Zell et al.
- (SIGGRAPH 2019) LRLE: Synthesis of biologically realistic human motion using joint torque actuation, Jiang et al.
- (TII 2018) Pham et al: Multicontact Interaction Force Sensing From Whole-Body Motion Capture, Pham et al.
- (ICCV Workshop 2017) Zell et al: Learning-based inverse dynamics of human motion, Zell et al.
- (CVPR Workshop 2017) Zell et al: Joint 3d human motion capture and physical analysis from monocular videos, Zell et al.
- (AIST 2017) HuGaDb: HuGaDB: Human Gait Database for Activity Recognition from Wearable Inertial Sensor Networks, Chereshnev et al.
- (SIGGRAPH 2016) Lv et al: Data-driven inverse dynamics for human motion, Lv et al.