Skip to content

An aggregation of human motion understanding research.

Notifications You must be signed in to change notification settings

Foruck/Awesome-Human-Motion

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 

Repository files navigation

Awesome Human Motion

An aggregation of human motion understanding research, feel free to contribute.

Reviews & Surveys

Motion Generation Motion Editing Motion Stylization

Human-Object Interaction Human-Scene Interaction Human-Human Interaction

Datasets Humanoid Bio-stuff

Reviews & Surveys

  • (JEB 2025) McAllister et al: Behavioural energetics in human locomotion: how energy use influences how we move, McAllister et al.
  • (ArXiv 2025) Gu et al: Humanoid Locomotion and Manipulation: Current Progress and Challenges in Control, Planning, and Learning, Gu et al.
  • (ArXiv 2024) Zhao et al: Motion Generation Review: Exploring Deep Learning for Lifelike Animation with Manifold, Zhao et al.
  • (T-PAMI 2023) Zhu et al: Human Motion Generation: A Survey, Zhu et al.

Motion Generation, Text/Speech/Music-Driven

    2025

    • (CVPR 2025) PersonalBooth: PersonaBooth: Personalized Text-to-Motion Generation, Kim et al.
    • (CVPR 2025) MARDM: Rethinking Diffusion for Text-Driven Human Motion Generation, Meng et al.
    • (CVPR 2025) StickMotion: StickMotion: Generating 3D Human Motions by Drawing a Stickman, Wang et al.
    • (CVPR 2025) LLaMo: Human Motion Instruction Tuning, Li et al.
    • (CVPR 2025) HOP: HOP: Heterogeneous Topology-based Multimodal Entanglement for Co-Speech Gesture Generation, Cheng et al.
    • (CVPR 2025) AtoM: AToM: Aligning Text-to-Motion Model at Event-Level with GPT-4Vision Reward, Han et al.
    • (CVPR 2025) EnergyMoGen: EnergyMoGen: Compositional Human Motion Generation with Energy-Based Diffusion Model in Latent Space, Zhang et al.
    • (CVPR 2025) Languate of Motion: The Language of Motion: Unifying Verbal and Non-verbal Language of 3D Human Motion, Chen et al.
    • (CVPR 2025) ScaMo: ScaMo: Exploring the Scaling Law in Autoregressive Motion Generation Model, Lu et al.
    • (CVPR 2025) Move in 2D: Move-in-2D: 2D-Conditioned Human Motion Generation, Huang et al.
    • (CVPR 2025) SOLAMI: SOLAMI: Social Vision-Language-Action Modeling for Immersive Interaction with 3D Autonomous Characters, Jiang et al.
    • (CVPR 2025) MVLift: Lifting Motion to the 3D World via 2D Diffusion, Li et al.
    • (ACM Sensys 2025) SHADE-AD: SHADE-AD: An LLM-Based Framework for Synthesizing Activity Data of Alzheimer’s Patients, Fu et al.
    • (ICRA 2025) MotionGlot: MotionGlot: A Multi-Embodied Motion Generation Model, Harithas et al.
    • (ICLR 2025) CLoSD: CLoSD: Closing the Loop between Simulation and Diffusion for multi-task character control, Tevet et al.
    • (ICLR 2025) PedGen: Learning to Generate Diverse Pedestrian Movements from Web Videos with Noisy Labels, Liu et al.
    • (ICLR 2025) HGM³: HGM³: Hierarchical Generative Masked Motion Modeling with Hard Token Mining, Jeong et al.
    • (ICLR 2025) LaMP: LaMP: Language-Motion Pretraining for Motion Generation, Retrieval, and Captioning, Li et al.
    • (ICLR 2025) MotionDreamer: MotionDreamer: One-to-Many Motion Synthesis with Localized Generative Masked Transformer, Wang et al.
    • (ICLR 2025) Lyu et al: Towards Unified Human Motion-Language Understanding via Sparse Interpretable Characterization, Lyu et al.
    • (ICLR 2025) DART: DART: A Diffusion-Based Autoregressive Motion Model for Real-Time Text-Driven Motion Control, Zhao et al.
    • (ICLR 2025) Motion-Agent: Motion-Agent: A Conversational Framework for Human Motion Generation with LLMs, Wu et al.
    • (IJCV 2025) Fg-T2M++:。 Fg-T2M++: LLMs-Augmented Fine-Grained Text Driven Human Motion Generation, Wang et al.
    • (TVCG 2025) SPORT: SPORT: From Zero-Shot Prompts to Real-Time Motion Generation, Ji et al.
    • (ArXiv 2025) HERO: HERO: Human Reaction Generation from Videos, Yu et al.
    • (ArXiv 2025) BioMoDiffuse: BioMoDiffuse: Physics-Guided Biomechanical Diffusion for Controllable and Authentic Human Motion Synthesis, Kang et al.
    • (ArXiv 2025) ExGes: ExGes: Expressive Human Motion Retrieval and Modulation for Audio-Driven Gesture Synthesis, Zhou et al.
    • (ArXiv 2025) AnyTop: AnyTop: Character Animation Diffusion with Any Topology, Gat et al.
    • (ArXiv 2025) MotionAnything: Motion Anything: Any to Motion Generation, Zhang et al.
    • (ArXiv 2025) GCDance: GCDance: Genre-Controlled 3D Full Body Dance Generation Driven By Music, Liu et al.
    • (ArXiv 2025) MotionLab: MotionLab: Unified Human Motion Generation and Editing via the Motion-Condition-Motion Paradigm, Guo et al.
    • (ArXiv 2025) CASIM: CASIM: Composite Aware Semantic Injection for Text to Motion Generation, Chang et al.
    • (ArXiv 2025) MotionPCM: MotionPCM: Real-Time Motion Synthesis with Phased Consistency Model, Jiang et al.
    • (ArXiv 2025) GestureLSM: GestureLSM: Latent Shortcut based Co-Speech Gesture Generation with Spatial-Temporal Modeling, Liu et al.
    • (ArXiv 2025) Free-T2M: Free-T2M: Frequency Enhanced Text-to-Motion Diffusion Model With Consistency Loss, Chen et al.
    • (ArXiv 2025) LS-GAN: LS-GAN: Human Motion Synthesis with Latent-space GANs, Amballa et al.
    • (ArXiv 2025) FlexMotion: FlexMotion: Lightweight, Physics-Aware, and Controllable Human Motion Generation, Tashakori et al.
    • (ArXiv 2025) HiSTF Mamba: HiSTF Mamba: Hierarchical Spatiotemporal Fusion with Multi-Granular Body-Spatial Modeling for High-Fidelity Text-to-Motion Generation, Zhan et al.
    • (ArXiv 2025) PackDiT: PackDiT: Joint Human Motion and Text Generation via Mutual Prompting, Jiang et al.
    • (3DV 2025) Unimotion: Unimotion: Unifying 3D Human Motion Synthesis and Understanding, Li et al.
    • (AAAI 2025) UniMuMo: UniMuMo: Unified Text, Music and Motion Generation, Yang et al.
    • (AAAI 2025) ALERT-Motion: Autonomous LLM-Enhanced Adversarial Attack for Text-to-Motion, Miao et al.
    • (AAAI 2025) MotionCraft: MotionCraft: Crafting Whole-Body Motion with Plug-and-Play Multimodal Controls, Bian et al.
    • (AAAI 2025) Light-T2M: Light-T2M: A Lightweight and Fast Model for Text-to-Motion Generation, Zeng et al.
    • (WACV 2025) ReinDiffuse: ReinDiffuse: Crafting Physically Plausible Motions with Reinforced Diffusion Model, Han et al.
    • (WACV 2025) MoRAG: MoRAG -- Multi-Fusion Retrieval Augmented Generation for Human Motion, Shashank et al.
    • (WACV 2025) Mandelli et al: Generation of Complex 3D Human Motion by Temporal and Spatial Composition of Diffusion Models, Mandelli et al.

    2024

    • (ArXiv 2024) SemTalk: SemTalk: Holistic Co-speech Motion Generation with Frame-level Semantic Emphasis, Zhang et al.
    • (ArXiv 2024) InterDance: InterDance: Reactive 3D Dance Generation with Realistic Duet Interactions, Li et al.
    • (ArXiv 2024) Motion-2-to-3: Motion-2-to-3: Leveraging 2D Motion Data to Boost 3D Motion Generation, Pi et al.
    • (ArXiv 2024) Mogo: Mogo: RQ Hierarchical Causal Transformer for High-Quality 3D Human Motion Generation, Fu et al.
    • (ArXiv 2024) CoMA: CoMA: Compositional Human Motion Generation with Multi-modal Agents, Sun et al.
    • (ArXiv 2024) SoPo: SoPo: Text-to-Motion Generation Using Semi-Online Preference Optimization, Tan et al.
    • (ArXiv 2024) RMD: RMD: A Simple Baseline for More General Human Motion Generation via Training-free Retrieval-Augmented Motion Diffuse, Liao et al.
    • (ArXiv 2024) BiPO: BiPO: Bidirectional Partial Occlusion Network for Text-to-Motion Synthesis, Hong et al.
    • (ArXiv 2024) DisCoRD: DisCoRD: Discrete Tokens to Continuous Motion via Rectified Flow Decoding, Cho et al.
    • (ArXiv 2024) MoTe: MoTe: Learning Motion-Text Diffusion Model for Multiple Generation Tasks, Wue et al.
    • (ArXiv 2024) InfiniDreamer: InfiniDreamer: Arbitrarily Long Human Motion Generation via Segment Score Distillation, Zhuo et al.
    • (ArXiv 2024) FTMoMamba: FTMoMamba: Motion Generation with Frequency and Text State Space Models, Li et al.
    • (ArXiv 2024) KinMo: KinMo: Kinematic-aware Human Motion Understanding and Generation, Zhang et al.
    • (ArXiv 2024) Morph: Morph: A Motion-free Physics Optimization Framework for Human Motion Generation, Li et al.
    • (ArXiv 2024) KMM: KMM: Key Frame Mask Mamba for Extended Motion Generation, Zhang et al.
    • (ArXiv 2024) MotionGPT-2: MotionGPT-2: A General-Purpose Motion-Language Model for Motion Generation and Understanding, Wang et al.
    • (ArXiv 2024) Lodge++: Lodge++: High-quality and Long Dance Generation with Vivid Choreography Patterns, Li et al.
    • (ArXiv 2024) MotionCLR: MotionCLR: Motion Generation and Training-Free Editing via Understanding Attention Mechanisms, Chen et al.
    • (ArXiv 2024) LEAD: LEAD: Latent Realignment for Human Motion Diffusion, Andreou et al.
    • (ArXiv 2024) Leite et al.: Enhancing Motion Variation in Text-to-Motion Models via Pose and Video Conditioned Editing, Leite et al.
    • (ArXiv 2024) MotionRL: MotionRL: Align Text-to-Motion Generation to Human Preferences with Multi-Reward Reinforcement Learning, Liu et al.
    • (ArXiv 2024) MotionLLM: MotionLLM: Understanding Human Behaviors from Human Motions and Videos, Chen et al.
    • (ArXiv 2024) Wang et al: Quo Vadis, Motion Generation? From Large Language Models to Large Motion Models, Wang et al.
    • (ArXiv 2024) T2M-X: T2M-X: Learning Expressive Text-to-Motion Generation from Partially Annotated Data, Liu et al.
    • (ArXiv 2024) BAD: BAD: Bidirectional Auto-regressive Diffusion for Text-to-Motion Generation, Hosseyni et al.
    • (ArXiv 2024) synNsync: Synergy and Synchrony in Couple Dances, Manukele et al.
    • (EMNLP 2024) Dong et al: Word-Conditioned 3D American Sign Language Motion Generation, Dong et al.
    • (NeurIPS D&B 2024) Text to blind motion: Text to Blind Motion, Kim et al.
    • (NeurIPS 2024) UniMTS: UniMTS: Unified Pre-training for Motion Time Series, Zhang et al.
    • (NeurIPS 2024) Christopher et al.: Constrained Synthesis with Projected Diffusion Models, Christopher et al.
    • (NeurIPS 2024) MoMu-Diffusion: MoMu-Diffusion: On Learning Long-Term Motion-Music Synchronization and Correspondence, You et al.
    • (NeurIPS 2024) MoGenTS: MoGenTS: Motion Generation based on Spatial-Temporal Joint Modeling, Yuan et al.
    • (NeurIPS 2024) M3GPT: M3GPT: An Advanced Multimodal, Multitask Framework for Motion Comprehension and Generation, Luo et al.
    • (NeurIPS Workshop 2024) Bikov et al: Fitness Aware Human Motion Generation with Fine-Tuning, Bikov et al.
    • (NeurIPS Workshop 2024) DGFM: DGFM: Full Body Dance Generation Driven by Music Foundation Models, Liu et al.
    • (ICPR 2024) FG-MDM: FG-MDM: Towards Zero-Shot Human Motion Generation via ChatGPT-Refined Descriptions, Shi et al.
    • (ACM MM 2024) SynTalker: Enabling Synergistic Full-Body Control in Prompt-Based Co-Speech Motion Generation, Chen et al.
    • (ACM MM 2024) L3EM: Towards Emotion-enriched Text-to-Motion Generation via LLM-guided Limb-level Emotion Manipulating. Yu et al.
    • (ACM MM 2024) StableMoFusion: StableMoFusion: Towards Robust and Efficient Diffusion-based Motion Generation Framework, Huang et al.
    • (ACM MM 2024) SATO: SATO: Stable Text-to-Motion Framework, Chen et al.
    • (ICANN 2024) PIDM: PIDM: Personality-Aware Interaction Diffusion Model for Gesture Generation, Shibasaki et al.
    • (HFES 2024) Macwan et al: High-Fidelity Worker Motion Simulation With Generative AI, Macwan et al.
    • (ECCV 2024) Jin et al.: Local Action-Guided Motion Diffusion Model for Text-to-Motion Generation, Jin et al.
    • (ECCV 2024) Motion Mamba: Motion Mamba: Efficient and Long Sequence Motion Generation, Zhong et al.
    • (ECCV 2024) EMDM: EMDM: Efficient Motion Diffusion Model for Fast, High-Quality Human Motion Generation, Zhou et al.
    • (ECCV 2024) CoMo: CoMo: Controllable Motion Generation through Language Guided Pose Code Editing, Huang et al.
    • (ECCV 2024) CoMusion: CoMusion: Towards Consistent Stochastic Human Motion Prediction via Motion Diffusion, Sun et al.
    • (ECCV 2024) Shan et al.: Towards Open Domain Text-Driven Synthesis of Multi-Person Motions, Shan et al.
    • (ECCV 2024) ParCo: ParCo: Part-Coordinating Text-to-Motion Synthesis, Zou et al.
    • (ECCV 2024) Sampieri et al.: Length-Aware Motion Synthesis via Latent Diffusion, Sampieri et al.
    • (ECCV 2024) ChroAccRet: Chronologically Accurate Retrieval for Temporal Grounding of Motion-Language Models, Fujiwara et al.
    • (ECCV 2024) MHC: Generating Physically Realistic and Directable Human Motions from Multi-Modal Inputs, Liu et al.
    • (ECCV 2024) ProMotion: Plan, Posture and Go: Towards Open-vocabulary Text-to-Motion Generation, Liu et al.
    • (ECCV 2024) FreeMotion: FreeMotion: MoCap-Free Human Motion Synthesis with Multimodal Large Language Models, Zhang et al.
    • (ECCV 2024) Text Motion Translator: Text Motion Translator: A Bi-Directional Model for Enhanced 3D Human Motion Generation from Open-Vocabulary Descriptions, Qian et al.
    • (ECCV 2024) FreeMotion: FreeMotion: A Unified Framework for Number-free Text-to-Motion Synthesis, Fan et al.
    • (ECCV 2024) Kinematic Phrases: Bridging the Gap between Human Motion and Action Semantics via Kinematic Phrases, Liu et al.
    • (ECCV 2024) MotionChain: MotionChain: Conversational Motion Controllers via Multimodal Prompts, Jiang et al.
    • (ECCV 2024) SMooDi: SMooDi: Stylized Motion Diffusion Model, Zhong et al.
    • (ECCV 2024) BAMM: BAMM: Bidirectional Autoregressive Motion Model, Pinyoanuntapong et al.
    • (ECCV 2024) MotionLCM: MotionLCM: Real-time Controllable Motion Generation via Latent Consistency Model, Dai et al.
    • (ECCV 2024) Ren et al.: Realistic Human Motion Generation with Cross-Diffusion Models, Ren et al.
    • (ECCV 2024) M2D2M: M2D2M: Multi-Motion Generation from Text with Discrete Diffusion Models, Chi et al.
    • (ECCV 2024) Large Motion Model: Large Motion Model for Unified Multi-Modal Motion Generation, Zhang et al.
    • (ECCV 2024) TesMo: Generating Human Interaction Motions in Scenes with Text Control, Yi et al.
    • (ECCV 2024) TLcontrol: TLcontrol: Trajectory and Language Control for Human Motion Synthesis, Wan et al.
    • (ICME 2024) ExpGest: ExpGest: Expressive Speaker Generation Using Diffusion Model and Hybrid Audio-Text Guidance, Cheng et al.
    • (ICME Workshop 2024) Chen et al: Anatomically-Informed Vector Quantization Variational Auto-Encoder for Text-to-Motion Generation, Chen et al.
    • (ICML 2024) HumanTOMATO: HumanTOMATO: Text-aligned Whole-body Motion Generation, Lu et al.
    • (ICML 2024) GPHLVM: Bringing Motion Taxonomies to Continuous Domains via GPLVM on Hyperbolic Manifolds, Jaquier et al.
    • (SIGGRAPH 2024) CondMDI: Flexible Motion In-betweening with Diffusion Models, Cohan et al.
    • (SIGGRAPH 2024) CAMDM: Taming Diffusion Probabilistic Models for Character Control, Chen et al.
    • (SIGGRAPH 2024) LGTM: LGTM: Local-to-Global Text-Driven Human Motion Diffusion Models, Sun et al.
    • (SIGGRAPH 2024) TEDi: TEDi: Temporally-Entangled Diffusion for Long-Term Motion Synthesis, Zhang et al.
    • (SIGGRAPH 2024) A-MDM: Interactive Character Control with Auto-Regressive Motion Diffusion Models, Shi et al.
    • (SIGGRAPH 2024) Starke et al.: Categorical Codebook Matching for Embodied Character Controllers, Starke et al.
    • (SIGGRAPH 2024) SuperPADL: SuperPADL: Scaling Language-Directed Physics-Based Control with Progressive Supervised Distillation, Juravsky et al.
    • (CVPR 2024) ProgMoGen: Programmable Motion Generation for Open-set Motion Control Tasks, Liu et al.
    • (CVPR 2024) PACER+: PACER+: On-Demand Pedestrian Animation Controller in Driving Scenarios, Wang et al.
    • (CVPR 2024) AMUSE: Emotional Speech-driven 3D Body Animation via Disentangled Latent Diffusion, Chhatre et al.
    • (CVPR 2024) Liu et al.: Towards Variable and Coordinated Holistic Co-Speech Motion Generation, Liu et al.
    • (CVPR 2024) MAS: MAS: Multi-view Ancestral Sampling for 3D motion generation using 2D diffusion, Kapon et al.
    • (CVPR 2024) WANDR: WANDR: Intention-guided Human Motion Generation, Diomataris et al.
    • (CVPR 2024) MoMask: MoMask: Generative Masked Modeling of 3D Human Motions, Guo et al.
    • (CVPR 2024) ChapPose: ChatPose: Chatting about 3D Human Pose, Feng et al.
    • (CVPR 2024) AvatarGPT: AvatarGPT: All-in-One Framework for Motion Understanding, Planning, Generation and Beyond, Zhou et al.
    • (CVPR 2024) MMM: MMM: Generative Masked Motion Model, Pinyoanuntapong et al.
    • (CVPR 2024) AAMDM: AAMDM: Accelerated Auto-regressive Motion Diffusion Model, Li et al.
    • (CVPR 2024) OMG: OMG: Towards Open-vocabulary Motion Generation via Mixture of Controllers, Liang et al.
    • (CVPR 2024) FlowMDM: FlowMDM: Seamless Human Motion Composition with Blended Positional Encodings, Barquero et al.
    • (CVPR 2024) Digital Life Project: Digital Life Project: Autonomous 3D Characters with Social Intelligence, Cai et al.
    • (CVPR 2024) EMAGE: EMAGE: Towards Unified Holistic Co-Speech Gesture Generation via Expressive Masked Audio Gesture Modeling, Liu et al.
    • (CVPR Workshop 2024) STMC: Multi-Track Timeline Control for Text-Driven 3D Human Motion Generation, Petrovich et al.
    • (CVPR Workshop 2024) InstructMotion: Exploring Text-to-Motion Generation with Human Preference, Sheng et al.
    • (ICLR 2024) Single Motion Diffusion: Raab et al.
    • (ICLR 2024) NeRM: NeRM: Learning Neural Representations for High-Framerate Human Motion Synthesis, Wei et al.
    • (ICLR 2024) PriorMDM: PriorMDM: Human Motion Diffusion as a Generative Prior, Shafir et al.
    • (ICLR 2024) OmniControl: OmniControl: Control Any Joint at Any Time for Human Motion Generation, Xie et al.
    • (ICLR 2024) Adiya et al.: Bidirectional Temporal Diffusion Model for Temporally Consistent Human Animation, Adiya et al.
    • (ICLR 2024) Duolando: Duolando: Follower GPT with Off-Policy Reinforcement Learning for Dance Accompaniment, Li et al.
    • (AAAI 2024) HuTuDiffusion: HuTuMotion: Human-Tuned Navigation of Latent Motion Diffusion Models with Minimal Feedback, Han et al.
    • (AAAI 2024) AMD: AMD: Anatomical Motion Diffusion with Interpretable Motion Decomposition and Fusion, Jing et al.
    • (AAAI 2024) MotionMix: MotionMix: Weakly-Supervised Diffusion for Controllable Motion Generation, Hoang et al.
    • (AAAI 2024) B2A-HDM: Towards Detailed Text-to-Motion Synthesis via Basic-to-Advanced Hierarchical Diffusion Model, Xie et al.
    • (AAAI 2024) Everything2Motion: Everything2Motion: Synchronizing Diverse Inputs via a Unified Framework for Human Motion Synthesis, Fan et al.
    • (AAAI 2024) MotionGPT: MotionGPT: Finetuned LLMs are General-Purpose Motion Generators, Zhang et al.
    • (AAAI 2024) Dong et al: Enhanced Fine-grained Motion Diffusion for Text-driven Human Motion Synthesis, Dong et al.
    • (AAAI 2024) UNIMASKM: A Unified Masked Autoencoder with Patchified Skeletons for Motion Synthesis, Mascaro et al.
    • (AAAI 2024) B2A-HDM: Towards Detailed Text-to-Motion Synthesis via Basic-to-Advanced Hierarchical Diffusion Model, Xie et al.
    • (TPAMI 2024) GUESS: GUESS: GradUally Enriching SyntheSis for Text-Driven Human Motion Generation, Gao et al.
    • (WACV 2024) Xie et al.: Sign Language Production with Latent Motion Transformer, Xie et al.

    2023

    • (NeurIPS 2023) GraphMotion: Act As You Wish: Fine-grained Control of Motion Diffusion Model with Hierarchical Semantic Graphs, Jin et al.
    • (NeurIPS 2023) MotionGPT: MotionGPT: Human Motion as Foreign Language, Jiang et al.
    • (NeurIPS 2023) FineMoGen: FineMoGen: Fine-Grained Spatio-Temporal Motion Generation and Editing, Zhang et al.
    • (NeurIPS 2023) InsActor: InsActor: Instruction-driven Physics-based Characters, Ren et al.
    • (ICCV 2023) AttT2M: AttT2M: Text-Driven Human Motion Generation with Multi-Perspective Attention Mechanism, Zhong et al.
    • (ICCV 2023) TMR: TMR: Text-to-Motion Retrieval Using Contrastive 3D Human Motion Synthesis, Petrovich et al.
    • (ICCV 2023) MAA: Make-An-Animation: Large-Scale Text-conditional 3D Human Motion Generation, Azadi et al.
    • (ICCV 2023) PhysDiff: PhysDiff: Physics-Guided Human Motion Diffusion Model, Yuan et al.
    • (ICCV 2023) ReMoDiffusion: ReMoDiffuse: Retrieval-Augmented Motion Diffusion Model, Zhang et al.
    • (ICCV 2023) BelFusion: BeLFusion: Latent Diffusion for Behavior-Driven Human Motion Prediction, Barquero et al.
    • (ICCV 2023) GMD: GMD: Guided Motion Diffusion for Controllable Human Motion Synthesis, Karunratanakul et al.
    • (ICCV 2023) HMD-NeMo: HMD-NeMo: Online 3D Avatar Motion Generation From Sparse Observations, Aliakbarian et al.
    • (ICCV 2023) SINC: SINC: Spatial Composition of 3D Human Motions for Simultaneous Action Generation, Athanasiou et al.
    • (ICCV 2023) Kong et al.: Priority-Centric Human Motion Generation in Discrete Latent Space, Kong et al.
    • (ICCV 2023) FgT2M: Fg-T2M: Fine-Grained Text-Driven Human Motion Generation via Diffusion Model, Wang et al.
    • (ICCV 2023) EMS: Breaking The Limits of Text-conditioned 3D Motion Synthesis with Elaborative Descriptions, Qian et al.
    • (SIGGRAPH 2023) GenMM: Example-based Motion Synthesis via Generative Motion Matching, Li et al.
    • (SIGGRAPH 2023) GestureDiffuCLIP: GestureDiffuCLIP: Gesture Diffusion Model with CLIP Latents, Ao et al.
    • (SIGGRAPH 2023) BodyFormer: BodyFormer: Semantics-guided 3D Body Gesture Synthesis with Transformer, Pang et al.
    • (SIGGRAPH 2023) Alexanderson et al.: Listen, denoise, action! Audio-driven motion synthesis with diffusion models, Alexanderson et al.
    • (CVPR 2023) AGroL: Avatars Grow Legs: Generating Smooth Human Motion from Sparse Tracking Inputs with Diffusion Model, Du et al.
    • (CVPR 2023) TALKSHOW: Generating Holistic 3D Human Motion from Speech, Yi et al.
    • (CVPR 2023) T2M-GPT: T2M-GPT: Generating Human Motion from Textual Descriptions with Discrete Representations, Zhang et al.
    • (CVPR 2023) UDE: UDE: A Unified Driving Engine for Human Motion Generation, Zhou et al.
    • (CVPR 2023) OOHMG: Being Comes from Not-being: Open-vocabulary Text-to-Motion Generation with Wordless Training, Lin et al.
    • (CVPR 2023) EDGE: EDGE: Editable Dance Generation From Music, Tseng et al.
    • (CVPR 2023) MLD: Executing your Commands via Motion Diffusion in Latent Space, Chen et al.
    • (CVPR 2023) MoDi: MoDi: Unconditional Motion Synthesis from Diverse Data, Raab et al.
    • (CVPR 2023) MoFusion: MoFusion: A Framework for Denoising-Diffusion-based Motion Synthesis, Dabral et al.
    • (CVPR 2023) Mo et al.: Continuous Intermediate Token Learning with Implicit Motion Manifold for Keyframe Based Motion Interpolation, Mo et al.
    • (ICLR 2023) HMDM: MDM: Human Motion Diffusion Model, Tevet et al.
    • (TPAMI 2023) MotionDiffuse: MotionDiffuse: Text-Driven Human Motion Generation with Diffusion Model, Zhang et al.
    • (TPAMI 2023) Bailando++: Bailando++: 3D Dance GPT with Choreographic Memory, Li et al.
    • (ArXiv 2023) UDE-2: A Unified Framework for Multimodal, Multi-Part Human Motion Synthesis, Zhou et al.
    • (ArXiv 2023) Motion Script: MotionScript: Natural Language Descriptions for Expressive 3D Human Motions, Yazdian et al.

    2022 and earlier

    • (NeurIPS 2022) NeMF: NeMF: Neural Motion Fields for Kinematic Animation, He et al.
    • (SIGGRAPH Asia 2022) PADL: PADL: Language-Directed Physics-Based Character, Juravsky et al.
    • (SIGGRAPH Asia 2022) Rhythmic Gesticulator: Rhythmic Gesticulator: Rhythm-Aware Co-Speech Gesture Synthesis with Hierarchical Neural Embeddings, Ao et al.
    • (3DV 2022) TEACH: TEACH: Temporal Action Composition for 3D Human, Athanasiou et al.
    • (ECCV 2022) Implicit Motion: Implicit Neural Representations for Variable Length Human Motion Generation, Cervantes et al.
    • (ECCV 2022) Zhong et al.: Learning Uncoupled-Modulation CVAE for 3D Action-Conditioned Human Motion Synthesis, Zhong et al.
    • (ECCV 2022) MotionCLIP: MotionCLIP: Exposing Human Motion Generation to CLIP Space, Tevet et al.
    • (ECCV 2022) PoseGPT: PoseGPT: Quantizing human motion for large scale generative modeling, Lucas et al.
    • (ECCV 2022) TEMOS: TEMOS: Generating diverse human motions from textual descriptions, Petrovich et al.
    • (ECCV 2022) TM2T: TM2T: Stochastic and Tokenized Modeling for the Reciprocal Generation of 3D Human Motions and Texts, Guo et al.
    • (SIGGRAPH 2022) AvatarCLIP: AvatarCLIP: Zero-Shot Text-Driven Generation and Animation of 3D Avatars, Hong et al.
    • (SIGGRAPH 2022) DeepPhase: Deepphase: Periodic autoencoders for learning motion phase manifolds, Starke et al.
    • (CVPR 2022) Guo et al.: Generating Diverse and Natural 3D Human Motions from Text, Guo et al.
    • (CVPR 2022) Bailando: Bailando: 3D Dance Generation by Actor-Critic GPT with Choreographic Memory, Li et al.
    • (ICCV 2021) ACTOR: Action-Conditioned 3D Human Motion Synthesis with Transformer VAE, Petrovich et al.
    • (ICCV 2021) AIST++: AI Choreographer: Music Conditioned 3D Dance Generation with AIST++, Li et al.
    • (SIGGRAPH 2021) Starke et al.: Neural animation layering for synthesizing martial arts movements, Starke et al.
    • (CVPR 2021) MOJO: We are More than Our Joints: Predicting how 3D Bodies Move, Zhang et al.
    • (ECCV 2020) DLow: DLow: Diversifying Latent Flows for Diverse Human Motion Prediction, Yuan et al.
    • (SIGGRAPH 2020) Starke et al.: Local motion phases for learning multi-contact character movements, Starke et al.

Motion Editing

  • (CVPR 2025) AnyMoLe: AnyMoLe: Any Character Motion In-Betweening Leveraging Video Diffusion Models, Yun et al.
  • (ArXiv 2025) Dai et al: Towards Synthesized and Editable Motion In-Betweening Through Part-Wise Phase Representation, Dai et al.
  • (SIGGRAPH Asia 2024) MotionFix: MotionFix: Text-Driven 3D Human Motion Editing, Athanasiou et al.
  • (NeurIPS 2024) CigTime: CigTime: Corrective Instruction Generation Through Inverse Motion Editing, Fang et al.
  • (SIGGRAPH 2024) Iterative Motion Editing: Iterative Motion Editing with Natural Language, Goel et al.
  • (CVPR 2024) DNO: DNO: Optimizing Diffusion Noise Can Serve As Universal Motion Priors, Karunratanakul et al.

Motion Stylization

  • (ArXiv 2024) MulSMo: MulSMo: Multimodal Stylized Motion Generation by Bidirectional Control Flow, Li et al.
  • (TSMC 2024) D-LORD: D-LORD for Motion Stylization, Gupta et al.
  • (ECCV 2024) HUMOS: HUMOS: Human Motion Model Conditioned on Body Shape, Tripathi et al.
  • (SIGGRAPH 2024) SMEAR: SMEAR: Stylized Motion Exaggeration with ARt-direction, Basset et al.
  • (CVPR 2024) MCM-LDM: Arbitrary Motion Style Transfer with Multi-condition Motion Latent Diffusion Model, Song et al.
  • (CVPR 2024) MoST: MoST: Motion Style Transformer between Diverse Action Contents, Kim et al.
  • (ICLR 2024) GenMoStyle: Generative Human Motion Stylization in Latent Space, Guo et al.

Human-Object Interaction

    2025

    • (CVPR 2025) SemGeoMo: SemGeoMo: Dynamic Contextual Human Motion Generation with Semantic and Geometric Guidance, Cong et al.
    • (CVPR 2025) Phys-Reach-Grasp: Learning Physics-Based Full-Body Human Reaching and Grasping from Brief Walking References, Li et al.
    • (CVPR 2025) ParaHome: ParaHome: Parameterizing Everyday Home Activities Towards 3D Generative Modeling of Human-Object Interactions, Kim et al.
    • (CVPR 2025) InterMimic: InterMimic: Towards Universal Whole-Body Control for Physics-Based Human-Object Interactions, Xu et al.
    • (CVPR 2025) CORE4D: CORE4D: A 4D Human-Object-Human Interaction Dataset for Collaborative Object REarrangement, Zhang et al.
    • (CVPR 2025) SkillMimic: SkillMimic: Learning Reusable Basketball Skills from Demonstrations, Wang et al.
    • (CVPR 2025) MobileH2R: MobileH2R: Learning Generalizable Human to Mobile Robot Handover Exclusively from Scalable and Diverse Synthetic Data, Wang et al.
    • (AAAI 2025) DiffGrasp: Diffgrasp: Whole-Body Grasping Synthesis Guided by Object Motion Using a Diffusion Model, Zhang et al.
    • (3DV 2025) Paschalidis et al: 3D Whole-body Grasp Synthesis with Directional Controllability, Paschalidis et al.
    • (3DV 2025) InterTrack: InterTrack: Tracking Human Object Interaction without Object Templates, Xie et al.
    • (3DV 2025) FORCE: FORCE: Dataset and Method for Intuitive Physics Guided Human-object Interaction, Zhang et al.
    • (ArXiv 2025) Kaiwu: Kaiwu: A Multimodal Manipulation Dataset and Framework for Robot Learning and Human-Robot Interaction, Jiang et al.
    • (ArXiv 2025) EigenActor: EigenActor: Variant Body-Object Interaction Generation Evolved from Invariant Action Basis Reasoning, Guo et al.

    2024

    • (ArXiv 2024) SyncDiff: SyncDiff: Synchronized Motion Diffusion for Multi-Body Human-Object Interaction Synthesis, He et al.
    • (ArXiv 2024) CHOICE: CHOICE: Coordinated Human-Object Interaction in Cluttered Environments for Pick-and-Place Actions, Lu et al.
    • (ArXiv 2024) TriDi: TriDi: Trilateral Diffusion of 3D Humans, Objects and Interactions, Petrov et al.
    • (ArXiv 2024) OOD-HOI: OOD-HOI: Text-Driven 3D Whole-Body Human-Object Interactions Generation Beyond Training Domains, Zhang et al.
    • (ArXiv 2024) COLLAGE: COLLAGE: Collaborative Human-Agent Interaction Generation using Hierarchical Latent Diffusion and Language Models, Daiya et al.
    • (ArXiv 2024) SMGDiff: SMGDiff: Soccer Motion Generation using diffusion probabilistic models, Yang et al.
    • (ArXiv 2024) Wu et al: Human-Object Interaction from Human-Level Instructions, Wu et al.
    • (NeurIPS 2024) HumanVLA: HumanVLA: Towards Vision-Language Directed Object Rearrangement by Physical Humanoid, Xu et al.
    • (NeurIPS 2024) OmniGrasp: Grasping Diverse Objects with Simulated Humanoids, Luo et al.
    • (NeurIPS 2024) EgoChoir: EgoChoir: Capturing 3D Human-Object Interaction Regions from Egocentric Views, Yang et al.
    • (NeurIPS 2024) CooHOI: CooHOI: Learning Cooperative Human-Object Interaction with Manipulated Object Dynamics, Gao et al.
    • (NeurIPS 2024) InterDreamer: InterDreamer: Zero-Shot Text to 3D Dynamic Human-Object Interaction, Xu et al.
    • (NeurIPS 2024) PiMForce: Posture-Informed Muscular Force Learning for Robust Hand Pressure Estimation, Seo et al.
    • (ECCV 2024) InterFusion: InterFusion: Text-Driven Generation of 3D Human-Object Interaction, Dai et al.
    • (ECCV 2024) CHOIS: Controllable Human-Object Interaction Synthesis, Li et al.
    • (ECCV 2024) F-HOI: F-HOI: Toward Fine-grained Semantic-Aligned 3D Human-Object Interactions, Yang et al.
    • (ECCV 2024) HIMO: HIMO: A New Benchmark for Full-Body Human Interacting with Multiple Objects, Lv et al.
    • (SIGGRAPH 2024) PhysicsPingPong: Strategy and Skill Learning for Physics-based Table Tennis Animation, Wang et al.
    • (CVPR 2024) NIFTY: NIFTY: Neural Object Interaction Fields for Guided Human Motion Synthesis, Kulkarni et al.
    • (CVPR 2024) HOI Animator: HOIAnimator: Generating Text-Prompt Human-Object Animations using Novel Perceptive Diffusion Models, Son et al.
    • (CVPR 2024) CG-HOI: CG-HOI: Contact-Guided 3D Human-Object Interaction Generation, Diller et al.
    • (IJCV 2024) InterCap: InterCap: Joint Markerless 3D Tracking of Humans and Objects in Interaction, Huang et al.
    • (3DV 2024) Phys-Fullbody-Grasp: Physically Plausible Full-Body Hand-Object Interaction Synthesis, Braun et al.
    • (3DV 2024) GRIP: GRIP: Generating Interaction Poses Using Spatial Cues and Latent Consistency, Taheri et al.
    • (AAAI 2024) FAVOR: Favor: Full-Body AR-driven Virtual Object Rearrangement Guided by Instruction Text, Li et al.

    2023 and earlier

    • (SIGGRAPH Asia 2023) OMOMO: Object Motion Guided Human Motion Synthesis, Li et al.
    • (ICCV 2023) CHAIRS: Full-Body Articulated Human-Object Interaction, Jiang et al.
    • (ICCV 2023) HGHOI: Hierarchical Generation of Human-Object Interactions with Diffusion Probabilistic Models, Pi et al.
    • (ICCV 2023) InterDiff: InterDiff: Generating 3D Human-Object Interactions with Physics-Informed Diffusion, Xu et al.
    • (CVPR 2023) Object Pop Up: Object pop-up: Can we infer 3D objects and their poses from human interactions alone? Petrov et al.
    • (CVPR 2023) ARCTIC: A Dataset for Dexterous Bimanual Hand-Object Manipulation, Fan et al.
    • (ECCV 2022) TOCH: TOCH: Spatio-Temporal Object-to-Hand Correspondence for Motion Refinement, Zhou et al.
    • (ECCV 2022) COUCH: COUCH: Towards Controllable Human-Chair Interactions, Zhang et al.
    • (ECCV 2022) SAGA: SAGA: Stochastic Whole-Body Grasping with Contact, Wu et al.
    • (CVPR 2022) GOAL: GOAL: Generating 4D Whole-Body Motion for Hand-Object Grasping, Taheri et al.
    • (CVPR 2022) BEHAVE: BEHAVE: Dataset and Method for Tracking Human Object Interactions, Bhatnagar et al.
    • (ECCV 2020) GRAB: GRAB: A Dataset of Whole-Body Human Grasping of Objects, Taheri et al.

Human-Scene Interaction

    2025

    • (CVPR 2025) EnvPoser: EnvPoser: Environment-aware Realistic Human Motion Estimation from Sparse Observations with Uncertainty Modeling. Xia et al.
    • (ICLR 2025) Sitcom-Crafter: Sitcom-Crafter: A Plot-Driven Human Motion Generation System in 3D Scenes, Chen et al.
    • (3DV 2025) Paschalidis et al: 3D Whole-body Grasp Synthesis with Directional Controllability, Paschalidis et al.
    • (WACV 2025) GHOST: GHOST: Grounded Human Motion Generation with Open Vocabulary Scene-and-Text Contexts, Milacski et al.
    • (ArXiv 2025) Gao et al: Jointly Understand Your Command and Intention: Reciprocal Co-Evolution between Scene-Aware 3D Human Motion Synthesis and Analysis, Gao et al.

    2024

    • (ArXiv 2024) ZeroHSI: ZeroHSI: Zero-Shot 4D Human-Scene Interaction by Video Generation, Li et al.
    • (ArXiv 2024) Mimicking-Bench: Mimicking-Bench: A Benchmark for Generalizable Humanoid-Scene Interaction Learning via Human Mimicking, Liu et al.
    • (ArXiv 2024) SCENIC: SCENIC: Scene-aware Semantic Navigation with Instruction-guided Control, Zhang et al.
    • (ArXiv 2024) Diffusion Implicit Policy: Diffusion Implicit Policy for Unpaired Scene-aware Motion synthesis, Gong et al.
    • (ArXiv 2024) SIMS: SIMS: Simulating Human-Scene Interactions with Real World Script Planning, Wang et al.
    • (ArXiv 2024) LaserHuman: LaserHuman: Language-guided Scene-aware Human Motion Generation in Free Environment, Cong et al.
    • (SIGGRAPH Asia 2024) LINGO: Autonomous Character-Scene Interaction Synthesis from Text Instruction, Jiang et al.
    • (NeurIPS 2024) DiMoP3D: Harmonizing Stochasticity and Determinism: Scene-responsive Diverse Human Motion Prediction, Lou et al.
    • (ECCV 2024) Liu et al.: Revisit Human-Scene Interaction via Space Occupancy, Liu et al.
    • (ECCV 2024) TesMo: Generating Human Interaction Motions in Scenes with Text Control, Yi et al.
    • (ECCV 2024 Workshop) SAST: Massively Multi-Person 3D Human Motion Forecasting with Scene Context, Mueller et al.
    • (Eurographics 2024) Kang et al: Learning Climbing Controllers for Physics-Based Characters, Kang et al.
    • (CVPR 2024) Afford-Motion: Move as You Say, Interact as You Can: Language-guided Human Motion Generation with Scene Affordance, Wang et al.
    • (CVPR 2024) GenZI: GenZI: Zero-Shot 3D Human-Scene Interaction Generation, Li et al.
    • (CVPR 2024) Cen et al.: Generating Human Motion in 3D Scenes from Text Descriptions, Cen et al.
    • (CVPR 2024) TRUMANS: Scaling Up Dynamic Human-Scene Interaction Modeling, Jiang et al.
    • (ICLR 2024) UniHSI: UniHSI: Unified Human-Scene Interaction via Prompted Chain-of-Contacts, Xiao et al.
    • (3DV 2024) Purposer: Purposer: Putting Human Motion Generation in Context, Ugrinovic et al.
    • (3DV 2024) InterScene: Synthesizing Physically Plausible Human Motions in 3D Scenes, Pan et al.
    • (3DV 2024) Mir et al: Generating Continual Human Motion in Diverse 3D Scenes, Mir et al.

    2023 and earlier

    • (ICCV 2023) DIMOS: DIMOS: Synthesizing Diverse Human Motions in 3D Indoor Scenes, Zhao et al.
    • (ICCV 2023) LAMA: Locomotion-Action-Manipulation: Synthesizing Human-Scene Interactions in Complex 3D Environments, Lee et al.
    • (ICCV 2023) Narrator: Narrator: Towards Natural Control of Human-Scene Interaction Generation via Relationship Reasoning, Xuan et al.
    • (CVPR 2023) CIMI4D: CIMI4D: A Large Multimodal Climbing Motion Dataset under Human-Scene Interactions, Yan et al.
    • (CVPR 2023) Scene-Ego: Scene-aware Egocentric 3D Human Pose Estimation, Wang et al.
    • (CVPR 2023) SLOPER4D: SLOPER4D: A Scene-Aware Dataset for Global 4D Human Pose Estimation in Urban Environments, Dai et al.
    • (CVPR 2023) CIRCLE: CIRCLE: Capture in Rich Contextual Environments, Araujo et al.
    • (CVPR 2023) SceneDiffuser: Diffusion-based Generation, Optimization, and Planning in 3D Scenes, Huang et al.
    • (SIGGRAPH 2023) PMP: PMP: Learning to Physically Interact with Environments using Part-wise Motion Priors, Bae et al.
    • (SIGGRAPH 2023) QuestEnvSim: QuestEnvSim: Environment-Aware Simulated Motion Tracking from Sparse Sensors, Lee et al.
    • (SIGGRAPH 2023) Hassan et al.: Synthesizing Physical Character-Scene Interactions, Hassan et al.
    • (NeurIPS 2022) Mao et al.: Contact-Aware Human Motion Forecasting, Mao et al.
    • (NeurIPS 2022) HUMANISE: HUMANISE: Language-conditioned Human Motion Generation in 3D Scenes, Wang et al.
    • (NeurIPS 2022) EmbodiedPose: Embodied Scene-aware Human Pose Estimation, Luo et al.
    • (ECCV 2022) GIMO: GIMO: Gaze-Informed Human Motion Prediction in Context, Zheng et al.
    • (ECCV 2022) COINS: COINS: Compositional Human-Scene Interaction Synthesis with Semantic Control, Zhao et al.
    • (CVPR 2022) Wang et al.: Towards Diverse and Natural Scene-aware 3D Human Motion Synthesis, Wang et al.
    • (CVPR 2022) GAMMA: The Wanderings of Odysseus in 3D Scenes, Zhang et al.
    • (ICCV 2021) SAMP: Stochastic Scene-Aware Motion Prediction, Hassan et al.
    • (ICCV 2021) LEMO: Learning Motion Priors for 4D Human Body Capture in 3D Scenes, Zhang et al.
    • (3DV 2020) PLACE: PLACE: Proximity Learning of Articulation and Contact in 3D Environments, Zhang et al.
    • (SIGGRAPH 2020) Starke et al.: Local motion phases for learning multi-contact character movements, Starke et al.
    • (CVPR 2020) PSI: Generating 3D People in Scenes without People, Zhang et al.
    • (SIGGRAPH Asia 2019) NSM: Neural State Machine for Character-Scene Interactions, Starke et al.
    • (ICCV 2019) PROX: Resolving 3D Human Pose Ambiguities with 3D Scene Constraints, Hassan et al.

Human-Human Interaction

  • (CVPR 2025) TIMotion: TIMotion: Temporal and Interactive Framework for Efficient Human-Human Motion Generation, Wang et al.
  • (ICLR 2025) Think Then React: Think Then React: Towards Unconstrained Action-to-Reaction Motion Generation, Tan et al.
  • (ICLR 2025) Ready-to-React: Ready-to-React: Online Reaction Policy for Two-Character Interaction Generation, Cen et al.
  • (ICLR 2025) InterMask: InterMask: 3D Human Interaction Generation via Collaborative Masked Modelling, Javed et al.
  • (3DV 2025) Interactive Humanoid: Interactive Humanoid: Online Full-Body Motion Reaction Synthesis with Social Affordance Canonicalization and Forecasting, Liu et al.
  • (ArXiv 2025) Invisible Strings: Invisible Strings: Revealing Latent Dancer-to-Dancer Interactions with Graph Neural Networks, Zerkowski et al.
  • (ArXiv 2025) Leader and Follower: Leader and Follower: Interactive Motion Generation under Trajectory Constraints, Wang et al.
  • (ArXiv 2024) Two in One: Two-in-One: Unified Multi-Person Interactive Motion Generation by Latent Diffusion Transformer, Li et al.
  • (ArXiv 2024) It Takes Two: It Takes Two: Real-time Co-Speech Two-person’s Interaction Generation via Reactive Auto-regressive Diffusion Model, Shi et al.
  • (ArXiv 2024) COLLAGE: COLLAGE: Collaborative Human-Agent Interaction Generation using Hierarchical Latent Diffusion and Language Models, Daiya et al.
  • (NeurIPS 2024) Harmony4D: Harmony4D: A Video Dataset for In-The-Wild Close Human Interactions, Khirodkar et al.
  • (NeurIPS 2024) InterControl: InterControl: Generate Human Motion Interactions by Controlling Every Joint, Wang et al.
  • (ACM MM 2024) PhysReaction: PhysReaction: Physically Plausible Real-Time Humanoid Reaction Synthesis via Forward Dynamics Guided 4D Imitation, Liu et al.
  • (ECCV 2024) Shan et al: Towards Open Domain Text-Driven Synthesis of Multi-Person Motions, Shan et al.
  • (ECCV 2024) ReMoS: ReMoS: 3D Motion-Conditioned Reaction Synthesis for Two-Person Interactions, Ghosh et al.
  • (CVPR 2024) Inter-X: Inter-X: Towards Versatile Human-Human Interaction Analysis, Xu et al.
  • (CVPR 2024) ReGenNet: ReGenNet: Towards Human Action-Reaction Synthesis, Xu et al.
  • (CVPR 2024) Fang et al.: Capturing Closely Interacted Two-Person Motions with Reaction Priors, Fan et al.
  • (CVPR Workshop 2024) in2IN: in2IN: Leveraging Individual Information to Generate Human INteractions, Ruiz-Ponce et al.
  • (IJCV 2024) InterGen: InterGen: Diffusion-based Multi-human Motion Generation under Complex Interactions, Liang et al.
  • (ICCV 2023) ActFormer: ActFormer: A GAN-based Transformer towards General Action-Conditioned 3D Human Motion Generation, Xu et al.
  • (ICCV 2023) Tanaka et al.: Role-aware Interaction Generation from Textual Description, Tanaka et al.
  • (CVPR 2023) Hi4D: Hi4D: 4D Instance Segmentation of Close Human Interaction, Yin et al.
  • (CVPR 2022) ExPI: Multi-Person Extreme Motion Prediction, Guo et al.
  • (CVPR 2020) CHI3D: Three-Dimensional Reconstruction of Human Interactions, Fieraru et al.

Datasets & Benchmarks

    2025

    • (CVPR 2025) AtoM: AToM: Aligning Text-to-Motion Model at Event-Level with GPT-4Vision Reward, Han et al.
    • (CVPR 2025) CORE4D: CORE4D: A 4D Human-Object-Human Interaction Dataset for Collaborative Object REarrangement, Zhang et al.
    • (ICLR 2025) MotionCritic: Aligning Human Motion Generation with Human Perceptions, Wang et al.
    • (ICLR 2025) LocoVR: LocoVR: Multiuser Indoor Locomotion Dataset in Virtual Reality, Takeyama et al.
    • (ICLR 2025) PMR: Pedestrian Motion Reconstruction: A Large-scale Benchmark via Mixed Reality Rendering with Multiple Perspectives and Modalities, Wang et al.
    • (AAAI 2025) EMHI: EMHI: A Multimodal Egocentric Human Motion Dataset with HMD and Body-Worn IMUs, Fan et al.
    • (ArXiv 2025) SGA-INTERACT: SGA-INTERACT: A3DSkeleton-based Benchmark for Group Activity Understanding in Modern Basketball Tactic, Yang et al.
    • (ArXiv 2025) Kaiwu: Kaiwu: A Multimodal Manipulation Dataset and Framework for Robot Learning and Human-Robot Interaction, Jiang et al.
    • (ArXiv 2025) Motion-X++: Motion-X++: A Large-Scale Multimodal 3D Whole-body Human Motion Dataset, Zhang et al.

    2024

    • (ArXiv 2024) Mimicking-Bench: Mimicking-Bench: A Benchmark for Generalizable Humanoid-Scene Interaction Learning via Human Mimicking, Liu et al.
    • (ArXiv 2024) LaserHuman: LaserHuman: Language-guided Scene-aware Human Motion Generation in Free Environment, Cong et al.
    • (ArXiv 2024) SCENIC: SCENIC: Scene-aware Semantic Navigation with Instruction-guided Control, Zhang et al.
    • (ArXiv 2024) synNsync: Synergy and Synchrony in Couple Dances, Manukele et al.
    • (ArXiv 2024) MotionBank: MotionBank: A Large-scale Video Motion Benchmark with Disentangled Rule-based Annotations, Xu et al.
    • (Github 2024) CMP & CMR: AnimationGPT: An AIGC tool for generating game combat motion assets, Liao et al.
    • (Scientific Data 2024) Evans et al: Synchronized Video, Motion Capture and Force Plate Dataset for Validating Markerless Human Movement Analysis, Evans et al.
    • (SIGGRAPH Asia 2024) LINGO: Autonomous Character-Scene Interaction Synthesis from Text Instruction, Jiang et al.
    • (NeurIPS 2024) Harmony4D: Harmony4D: A Video Dataset for In-The-Wild Close Human Interactions, Khirodkar et al.
    • (NeurIPS D&B 2024) EgoSim: EgoSim: An Egocentric Multi-view Simulator for Body-worn Cameras during Human Motion, Hollidt et al.
    • (NeurIPS D&B 2024) Muscles in Time: Muscles in Time: Learning to Understand Human Motion by Simulating Muscle Activations, Schneider et al.
    • (NeurIPS D&B 2024) Text to blind motion: Text to blind motion, Kim et al.
    • (ACM MM 2024) CLaM: CLaM: An Open-Source Library for Performance Evaluation of Text-driven Human Motion Generation, Chen et al.
    • (ECCV 2024) AddBiomechanics: AddBiomechanics Dataset: Capturing the Physics of Human Motion at Scale, Werling et al.
    • (ECCV 2024) LiveHPS++: LiveHPS++: Robust and Coherent Motion Capture in Dynamic Free Environment, Ren et al.
    • (ECCV 2024) SignAvatars: SignAvatars: A Large-scale 3D Sign Language Holistic Motion Dataset and Benchmark, Yu et al.
    • (ECCV 2024) Nymeria: Nymeria: A massive collection of multimodal egocentric daily motion in the wild, Ma et al.
    • (Multibody System Dynamics 2024) Human3.6M+: Using musculoskeletal models to generate physically-consistent data for 3D human pose, kinematic, dynamic, and muscle estimation, Nasr et al.
    • (CVPR 2024) Inter-X: Inter-X: Towards Versatile Human-Human Interaction Analysis, Xu et al.
    • (CVPR 2024) HardMo: HardMo: ALarge-Scale Hardcase Dataset for Motion Capture, Liao et al.
    • (CVPR 2024) MMVP: MMVP: A Multimodal MoCap Dataset with Vision and Pressure Sensors, Zhang et al.
    • (CVPR 2024) RELI11D: RELI11D: A Comprehensive Multimodal Human Motion Dataset and Method, Yan et al.

    2023 and earlier

    • (SIGGRAPH Asia 2023) GroundLink: GroundLink: A Dataset Unifying Human Body Movement and Ground Reaction Dynamics, Han et al.
    • (NeurIPS D&B 2023) HOH: HOH: Markerless Multimodal Human-Object-Human Handover Dataset with Large Object Count, Wiederhold et al.
    • (NeurIPS D&B 2023) Motion-X: Motion-X: A Large-scale 3D Expressive Whole-body Human Motion Dataset, Lin et al.
    • (NeurIPS D&B 2023) Humans in Kitchens: Humans in Kitchens: A Dataset for Multi-Person Human Motion Forecasting with Scene Context, Tanke et al.
    • (ICCV 2023) CHAIRS: Full-Body Articulated Human-Object Interaction, Jiang et al.
    • (CVPR 2023) CIMI4D: CIMI4D: A Large Multimodal Climbing Motion Dataset under Human-Scene Interactions, Yan et al.
    • (CVPR 2023) FLAG3D: FLAG3D: A 3D Fitness Activity Dataset with Language Instruction, Tang et al.
    • (CVPR 2023) Hi4D: Hi4D: 4D Instance Segmentation of Close Human Interaction, Yin et al.
    • (CVPR 2023) CIRCLE: CIRCLE: Capture in Rich Contextual Environments, Araujo et al.
    • (NeurIPS 2022) MoCapAct: MoCapAct: A Multi-Task Dataset for Simulated Humanoid Control, Wagener et al.
    • (ACM MM 2022) ForcePose: Learning to Estimate External Forces of Human Motion in Video, Louis et al.
    • (ECCV 2022) BEAT: BEAT: A Large-Scale Semantic and Emotional Multi-Modal Dataset for Conversational Gestures Synthesis, Liu et al.
    • (ECCV 2022) BRACE: BRACE: The Breakdancing Competition Dataset for Dance Motion Synthesis, Moltisanti et al.
    • (ECCV 2022) EgoBody: Egobody: Human body shape and motion of interacting people from head-mounted devices, Zhang et al.
    • (ECCV 2022) GIMO: GIMO: Gaze-Informed Human Motion Prediction in Context, Zheng et al.
    • (ECCV 2022) HuMMan: HuMMan: Multi-Modal 4D Human Dataset for Versatile Sensing and Modeling, Cai et al.
    • (CVPR 2022) ExPI: Multi-Person Extreme Motion Prediction, Guo et al.
    • (CVPR 2022) HumanML3D: Generating Diverse and Natural 3D Human Motions from Text, Guo et al.
    • (CVPR 2022) BEHAVE: BEHAVE: Dataset and Method for Tracking Human Object Interactions, Bhatnagar et al.
    • (ICCV 2021) AIST++: AI Choreographer: Music Conditioned 3D Dance Generation with AIST++, Li et al.
    • (CVPR 2021) Fit3D: AIFit: Automatic 3D Human-Interpretable Feedback Models for Fitness Training, Fieraru et al.
    • (CVPR 2021) BABEL: BABEL: Bodies, Action, and Behavior with English Labels, Punnakkal et al.
    • (AAAI 2021) HumanSC3D: Learning complex 3d human self-contact, Fieraru et al.
    • (CVPR 2020) CHI3D: Three-Dimensional Reconstruction of Human Interactions, Fieraru et al.
    • (ICCV 2019) PROX: Resolving 3D Human Pose Ambiguities with 3D Scene Constraints, Hassan et al.
    • (ICCV 2019) AMASS: AMASS: Archive of Motion Capture As Surface Shapes, Mahmood et al.

Humanoid, Simulated or Real

    2025

    • (CVPR 2025) InterMimic: InterMimic: Towards Universal Whole-Body Control for Physics-Based Human-Object Interactions, Xu et al.
    • (CVPR 2025) SkillMimic: SkillMimic: Learning Reusable Basketball Skills from Demonstrations, Wang et al.
    • (ICRA 2025) HOVER: HOVER: Versatile Neural Whole-Body Controller for Humanoid Robots, He et al.
    • (ICRA 2025) PIM: Learning Humanoid Locomotion with Perceptive Internal Model, Long et al.
    • (ICRA 2025) Think on your feet: Think on your feet: Seamless Transition between Human-like Locomotion in Response to Changing Commands, Huang et al.
    • (ICLR 2025) Puppeteer: Hierarchical World Models as Visual Whole-Body Humanoid Controllers, Hansen et al.
    • (ICLR 2025) FB-CPR: Zero-Shot Whole-Body Humanoid Control via Behavioral Foundation Models, Tirinzoni et al.
    • (ICLR 2025) MPC2: Motion Control of High-Dimensional Musculoskeletal System with Hierarchical Model-Based Planning, Wei et al.
    • (ICLR 2025) CLoSD: CLoSD: Closing the Loop between Simulation and Diffusion for multi-task character control, Tevet et al.
    • (ICLR 2025) HiLo: HiLo: Learning Whole-Body Human-like Locomotion with Motion Tracking Controller, Zhang et al.
    • (Github 2025) MobilityGen: MobilityGen.
    • (ArXiv 2025) GMP: Natural Humanoid Robot Locomotion with Generative Motion Prior, Zhang et al.
    • (ArXiv 2025) Sun et al: Learning Perceptive Humanoid Locomotion over Challenging Terrain, Sun et al.
    • (ArXiv 2025) HWC-Loco: HWC-Loco: AHierarchical Whole-Body Control Approach to Robust Humanoid Locomotion, Lin et al.
    • (ArXiv 2025) Lin et al: Sim-to-Real Reinforcement Learning for Vision-Based Dexterous Manipulation on Humanoids, Lin et al.
    • (ArXiv 2025) COMPASS: COMPASS: Cross-embOdiment Mobility Policy via ResiduAl RL and Skill Synthesis, Liu et al.
    • (ArXiv 2025) VB-COM: VB-Com: Learning Vision-Blind Composite Humanoid Locomotion Against Deficient Perception, Ren et al.
    • (ArXiv 2025) ModSkill: ModSkill: Physical Character Skill Modularization, Huang et al.
    • (ArXiv 2025) Humanoid-VLA: Humanoid-VLA: Towards Universal Humanoid Control with Visual Integration, Ding et al.
    • (ArXiv 2025) Li et al: Human-Like Robot Impedance Regulation Skill Learning from Human-Human Demonstrations, Li et al.
    • (ArXiv 2025) HumanUP: Learning Getting-Up Policies for Real-World Humanoid Robots, He et al.
    • (ArXiv 2025) RHINO: RHINO: Learning Real-Time Humanoid-Human-Object Interaction from Human Demonstrations, Chen et al.
    • (ArXiv 2025) HOMIE: HOMIE: Humanoid Loco-Manipulation with Isomorphic Exoskeleton Cockpit, Ben et al.
    • (ArXiv 2025) BeamDojo: BeamDojo: Learning Agile Humanoid Locomotion on Sparse Footholds, Wang et al.
    • (ArXiv 2025) HoST: Learning Humanoid Standing-up Control across Diverse Postures, Huang et al.
    • (ArXiv 2025) Embrace Collisions: Embrace Collisions: Humanoid Shadowing for Deployable Contact-Agnostics Motion, Zhuang et al.
    • (ArXiv 2025) ToddlerBot: ToddlerBot: Open-Source ML-Compatible Humanoid Platform for Loco-Manipulation, Shi et al.
    • (ArXiv 2025) ASAP: ASAP: Aligning Simulation and Real-World Physics for Learning Agile Humanoid Whole-Body Skills, He et al.
    • (ArXiv 2025) Gu et al: Humanoid Locomotion and Manipulation: Current Progress and Challenges in Control, Planning, and Learning, Gu et al.

    2024

    • (ArXiv 2024) UH-1: Learning from Massive Human Videos for Universal Humanoid Pose Control, Mao et al.
    • (ArXiv 2024) Mimicking-Bench: Mimicking-Bench: A Benchmark for Generalizable Humanoid-Scene Interaction Learning via Human Mimicking, Liu et al.
    • (ArXiv 2024) Exbody2: Exbody2: Advanced Expressive Humanoid Whole-Body Control, Ji et al.
    • (ArXiv 2024) SIMS: SIMS: Simulating Human-Scene Interactions with Real World Script Planning, Wang et al.
    • (ArXiv 2024) Humanoidlympics: Humanoidlympics: Sports Environments for Physically Simulated Humanoids, Luo et al.
    • (ArXiv 2024) PhySHOI: PhysHOI: Physics-Based Imitation of Dynamic Human-Object Interaction, Wang et al.
    • (RA-L 2024) Liu et al: Human-Humanoid Robots Cross-Embodiment Behavior-Skill Transfer Using Decomposed Adversarial Learning from Demonstration, Liu et al.
    • (SIGGRAPH Asia 2024) PDP: PDP: Physics-Based Character Animation via Diffusion Policy, Truong et al.
    • (SIGGRAPH Asia 2024) MaskedMimic: MaskedMimic: Unified Physics-Based Character Control Through Masked Motion Inpainting, Tessler et al.
    • (NeurIPS 2024) HumanVLA: HumanVLA: Towards Vision-Language Directed Object Rearrangement by Physical Humanoid, Xu et al.
    • (NeurIPS 2024) OmniGrasp: Grasping Diverse Objects with Simulated Humanoids, Luo et al.
    • (NeurIPS 2024) InterControl: InterControl: Generate Human Motion Interactions by Controlling Every Joint, Wang et al.
    • (NeurIPS 2024) CooHOI: CooHOI: Learning Cooperative Human-Object Interaction with Manipulated Object Dynamics, Gao et al.
    • (NeurIPS 2024) Radosavovic et al.: Humanoid Locomotion as Next Token Prediction, Radosavovic et al.
    • (CoRL 2024) HARMON: Harmon: Whole-Body Motion Generation of Humanoid Robots from Language Descriptions, Jiang et al.
    • (CoRL 2024) OKAMI: OKAMI: Teaching Humanoid Robots Manipulation Skills through Single Video Imitation, Li et al.
    • (CoRL 2024) HumanPlus: HumanPlus: Humanoid Shadowing and Imitation from Humans, Fu et al.
    • (CoRL 2024) OmniH2O: OmniH2O: Universal and Dexterous Human-to-Humanoid Whole-Body Teleoperation and Learning, He et al.
    • (Humanoids 2024) Self-Aware: Know your limits! Optimize the behavior of bipedal robots through self-awareness, Mascaro et al.
    • (ACM MM 2024) PhysReaction: PhysReaction: Physically Plausible Real-Time Humanoid Reaction Synthesis via Forward Dynamics Guided 4D Imitation, Liu et al.
    • (IROS 2024) H2O: Learning Human-to-Humanoid Real-Time Whole-Body Teleoperation, He et al.
    • (ECCV 2024) MHC: Generating Physically Realistic and Directable Human Motions from Multi-Modal Inputs, Shrestha et al.
    • (ICML 2024) DIDI: DIDI: Diffusion-Guided Diversity for Offline Behavioral Generation, Liu et al.
    • (SIGGRAPH 2024) MoConVQ: MoConVQ: Unified Physics-Based Motion Control via Scalable Discrete Representations, Yao et al.
    • (SIGGRAPH 2024) PhysicsPingPong: Strategy and Skill Learning for Physics-based Table Tennis Animation, Wang et al.
    • (SIGGRAPH 2024) SuperPADL: SuperPADL: Scaling Language-Directed Physics-Based Control with Progressive Supervised Distillation, Juravsky et al.
    • (CVPR 2024) SimXR: Real-Time Simulated Avatar from Head-Mounted Sensors, Luo et al.
    • (CVPR 2024) AnySkill: AnySkill: Learning Open-Vocabulary Physical Skill for Interactive Agents, Cui et al.
    • (ICLR 2024) PULSE: Universal Humanoid Motion Representations for Physics-Based Control, Luo et al.
    • (ICLR 2024) H-GAP: H-GAP: Humanoid Control with a Generalist Planner, Jiang et al.
    • (ICLR 2024) UniHSI: UniHSI: Unified Human-Scene Interaction via Prompted Chain-of-Contacts, Xiao et al.
    • (3DV 2024) Phys-Fullbody-Grasp: Physically Plausible Full-Body Hand-Object Interaction Synthesis, Braun et al.
    • (RSS 2024) ExBody: Expressive Whole-Body Control for Humanoid Robots, Cheng et al.

    2023 and earlier

    • (SIGGRAPH Asia 2023) Fatigued Movements: Discovering Fatigued Movements for Virtual Character Animation, Cheema et al.
    • (SIGGRAPH Asia 2023) CASE: C·ASE: Learning Conditional Adversarial Skill Embeddings for Physics-based Characters, Dou et al.
    • (SIGGRAPH Asia 2023) AdaptNet: AdaptNet: Policy Adaptation for Physics-Based Character Control, Xu et al.
    • (SIGGRAPH Asia 2023) NCP: Neural Categorical Priors for Physics-Based Character Control, Zhu et al.
    • (SIGGRAPH Asia 2023) DROP: DROP: Dynamics Responses from Human Motion Prior and Projective Dynamics, Jiang et al.
    • (NeurIPS 2023) InsActor: InsActor: Instruction-driven Physics-based Characters, Ren et al.
    • (CoRL 2023) Humanoid4Parkour: Humanoid Parkour Learning, Zhuang et al.
    • (CoRL Workshop 2023) Words into Action: Words into Action: Learning Diverse Humanoid Robot Behaviors using Language Guided Iterative Motion Refinement, Kumar et al.
    • (ICCV 2023) PHC: Perpetual Humanoid Control for Real-time Simulated Avatars, Luo et al.
    • (CVPR 2023) Trace and Pace: Trace and Pace: Controllable Pedestrian Animation via Guided Trajectory Diffusion, Rempe et al.
    • (SIGGRAPH 2023) Vid2Player3D: DiffMimic: Efficient Motion Mimicking with Differentiable Physics, Zhang et al.
    • (SIGGRAPH 2023) QuestEnvSim: QuestEnvSim: Environment-Aware Simulated Motion Tracking from Sparse Sensors, Lee et al.
    • (SIGGRAPH 2023) Hassan et al.: Synthesizing Physical Character-Scene Interactions, Hassan et al.
    • (SIGGRAPH 2023) CALM: CALM: Conditional Adversarial Latent Models for Directable Virtual Characters, Tessler et al.
    • (SIGGRAPH 2023) Composite Motion: Composite Motion Learning with Task Control, Xu et al.
    • (ICLR 2023) DiffMimic: DiffMimic: Efficient Motion Mimicking with Differentiable Physics, Ren et al.
    • (NeurIPS 2022) EmbodiedPose: Embodied Scene-aware Human Pose Estimation, Luo et al.
    • (NeurIPS 2022) MoCapAct: MoCapAct: A Multi-Task Dataset for Simulated Humanoid Control, Wagener et al.
    • (SIGGRAPH Asia 2022) Gopinath et al.: Motion In-betweening for Physically Simulated Characters, Gopinath et al.
    • (SIGGRAPH Asia 2022) AIP: AIP: Adversarial Interaction Priors for Multi-Agent Physics-based Character Control, Younes et al.
    • (SIGGRAPH Asia 2022) ControlVAE: ControlVAE: Model-Based Learning of Generative Controllers for Physics-Based Characters, Yao et al.
    • (SIGGRAPH Asia 2022) QuestSim: QuestSim: Human Motion Tracking from Sparse Sensors with Simulated Avatars, Winkler et al.
    • (SIGGRAPH Asia 2022) PADL: PADL: Language-Directed Physics-Based Character, Juravsky et al.
    • (SIGGRAPH Asia 2022) Wang et al.: Differentiable Simulation of Inertial Musculotendons, Wang et al.
    • (SIGGRAPH 2022) ASE: ASE: Large-Scale Reusable Adversarial Skill Embeddings for Physically Simulated Characters, Peng et al.
    • (Journal of Neuro-Engineering and Rehabilitation 2021) Learn to Move: Deep Reinforcement Learning for Modeling Human Locomotion Control in Neuromechanical Simulation, Peng et al.
    • (NeurIPS 2021) KinPoly: Dynamics-Regulated Kinematic Policy for Egocentric Pose Estimation, Luo et al.
    • (SIGGRAPH 2021) AMP: AMP: Adversarial Motion Priors for Stylized Physics-Based Character Control, Peng et al.
    • (CVPR 2021) SimPoE: SimPoE: Simulated Character Control for 3D Human Pose Estimation, Yuan et al.
    • (NeurIPS 2020) RFC: Residual Force Control for Agile Human Behavior Imitation and Extended Motion Synthesis, Yuan et al.
    • (ICLR 2020) Yuan et al.: Diverse Trajectory Forecasting with Determinantal Point Processes, Yuan et al.
    • (ICCV 2019) Ego-Pose: Ego-Pose Estimation and Forecasting as Real-Time PD Control, Yuan et al.
    • (SIGGRAPH 2018) DeepMimic: DeepMimic: Example-Guided Deep Reinforcement Learning of Physics-Based Character Skills, Peng et al.

Bio-stuff: Human Anatomy, Biomechanics, Physiology

  • (CVPR 2025) HDyS: Homogeneous Dynamics Space for Heterogeneous Humans, Liu et al.
  • (ICLR 2025) ImDy: ImDy: Human Inverse Dynamics from Imitated Observations, Liu et al.
  • (ICLR 2025) MPC2: Motion Control of High-Dimensional Musculoskeletal System with Hierarchical Model-Based Planning, Wei et al.
  • (ACM Sensys 2025) SHADE-AD: SHADE-AD: An LLM-Based Framework for Synthesizing Activity Data of Alzheimer’s Patients, Fu et al.
  • (JEB 2025) McAllister et al: Behavioural energetics in human locomotion: how energy use influences how we move, McAllister et al.
  • (WACV 2025) OpenCapBench: A Benchmark to Bridge Pose Estimation and Biomechanics, Gozlan et al.
  • (ArXiv 2025) Wu et al: Muscle Activation Estimation by Optimizing the Musculoskeletal Model for Personalized Strength and Conditioning Training, Wu et al.
  • (ArXiv 2025) Cotton et al: Biomechanical Reconstruction with Confidence Intervals from Multiview Markerless Motion Capture, Cotton et al.
  • (BiorXiv 2024) Lai et al: Mapping Grip Force to Muscular Activity Towards Understanding Upper Limb Musculoskeletal Intent using a Novel Grip Strength Model, Lai et al.
  • (IROS 2024) Shahriari et al: Enhancing Robustness in Manipulability Assessment: The Pseudo-Ellipsoid Approach, Shahriari et al.
  • (SIGGRAPH Asia 2024) BioDesign: Motion-Driven Neural Optimizer for Prophylactic Braces Made by Distributed Microstructures, Han et al.
  • (Scientific Data 2024) Evans et al: Synchronized Video, Motion Capture and Force Plate Dataset for Validating Markerless Human Movement Analysis, Evans et al.
  • (NeurIPS D&B 2024) Muscles in Time: Muscles in Time: Learning to Understand Human Motion by Simulating Muscle Activations, Schneider et al.
  • (CoRL 2024) Wei et al: Safe Bayesian Optimization for the Control of High-Dimensional Embodied Systems, Wei et al.
  • (HFES 2024) Macwan et al: High-Fidelity Worker Motion Simulation With Generative AI, Macwan et al.
  • (ECCV 2024) AddBiomechanics: AddBiomechanics Dataset: Capturing the Physics of Human Motion at Scale, Werling et al.
  • (ECCV 2024) MANIKIN: MANIKIN: Biomechanically Accurate Neural Inverse Kinematics for Human Motion Estimation, Jiang et al.
  • (TOG 2024) NICER: NICER: A New and Improved Consumed Endurance and Recovery Metric to Quantify Muscle Fatigue of Mid-Air Interactions, Li et al.
  • (ICML 2024) DynSyn: DynSyn: Dynamical Synergistic Representation for Efficient Learning and Control in Overactuated Embodied Systems, He et al.
  • (Multibody System Dynamics 2024) Human3.6M+: Using musculoskeletal models to generate physically-consistent data for 3D human pose, kinematic, dynamic, and muscle estimation, Nasr et al.
  • (CVPR 2024) HIT: HIT: Estimating Internal Human Implicit Tissues from the Body Surface, Keller et al.
  • (Frontiers in Neuroscience 2024) Dai et al: Full-body pose reconstruction and correction in virtual reality for rehabilitation training, Dai et al.
  • (ICRA 2024) He et al.: Self Model for Embodied Intelligence: Modeling Full-Body Human Musculoskeletal System and Locomotion Control with Hierarchical Low-Dimensional Representation, He et al.
  • (SIGGRAPH Asia 2023) Fatigued Movements: Discovering Fatigued Movements for Virtual Character Animation, Cheema et al.
  • (SIGGRAPH Asia 2023) SKEL: From skin to skeleton: Towards biomechanically accurate 3d digital humans, Keller et al.
  • (SIGGRAPH Asia 2023) MuscleVAE: MuscleVAE: Model-Based Controllers of Muscle-Actuated Characters, Feng et al.
  • (SIGGRAPH 2023) Bidirectional GaitNet: Bidirectional GaitNet, Park et al.
  • (SIGGRAPH 2023) Lee et al.: Anatomically Detailed Simulation of Human Torso, Lee et al.
  • (ICCV 2023) MiA: Muscles in Action, Chiquer et al.
  • (CVPR 2022) OSSO: OSSO: Obtaining Skeletal Shape from Outside, Keller et al.
  • (Scientific Data 2022) Xing et al: Functional movement screen dataset collected with two Azure Kinect depth sensors, Xing et al.
  • (NCA 2020) Zell et al: Learning inverse dynamics for human locomotion analysis, Zell et al.
  • (ECCV 2020) Zell et al: Weakly-supervised learning of human dynamics, Zell et al.
  • (SIGGRAPH 2019) LRLE: Synthesis of biologically realistic human motion using joint torque actuation, Jiang et al.
  • (TII 2018) Pham et al: Multicontact Interaction Force Sensing From Whole-Body Motion Capture, Pham et al.
  • (ICCV Workshop 2017) Zell et al: Learning-based inverse dynamics of human motion, Zell et al.
  • (CVPR Workshop 2017) Zell et al: Joint 3d human motion capture and physical analysis from monocular videos, Zell et al.
  • (AIST 2017) HuGaDb: HuGaDB: Human Gait Database for Activity Recognition from Wearable Inertial Sensor Networks, Chereshnev et al.
  • (SIGGRAPH 2016) Lv et al: Data-driven inverse dynamics for human motion, Lv et al.