Skip to content

Foruck/Awesome-Human-Motion

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

70 Commits
 
 

Repository files navigation

Awesome Human Motion

An aggregation of human motion understanding research.

Reviews & Surveys

Motion Generation Motion Editing Motion Stylization

Human-Object Interaction Human-Scene Interaction Human-Human Interaction

Datasets Humanoid Bio-stuff

Reviews & Surveys

  • Zhao et al. Motion Generation Review: Exploring Deep Learning for Lifelike Animation with Manifold, Zhao et al. ArXiv 2024.
  • Zhu et al. Human Motion Generation: A Survey, Zhu et al. T-PAMI 2023.

Motion Generation, Text/Speech/Music-Driven

  • LS-GAN. LS-GAN: HumanMotion Synthesis with Latent-space GANs, Amballa et al. ArXiv 2025.
  • SemTalk. SemTalk: Holistic Co-speech Motion Generation with Frame-level Semantic Emphasis, Zhang et al. ArXiv 2024.
  • InterDance. InterDance: Reactive 3D Dance Generation with Realistic Duet Interactions, Li et al. ArXiv 2024.
  • Unimotion. Unimotion: Unifying 3D Human Motion Synthesis and Understanding, Li et al. 3DV 2025.
  • MotionCraft. MotionCraft: Crafting Whole-Body Motion with Plug-and-Play Multimodal Controls, Bian et al. AAAI 2025.
  • ReinDiffuse. ReinDiffuse: Crafting Physically Plausible Motions with Reinforced Diffusion Model, Han et al. WACV 2025.
  • ScaMo. ScaMo: Exploring the Scaling Law in Autoregressive Motion Generation Model, Lu et al. ArXiv 2024.
  • EnergyMoGen. EnergyMoGen: Compositional Human Motion Generation with Energy-Based Diffusion Model in Latent Space, Zhang et al. ArXiv 2024.
  • Move in 2D. Move-in-2D: 2D-Conditioned Human Motion Generation, Huang et al. ArXiv 2024.
  • Motion-2-to-3. Motion-2-to-3: Leveraging 2D Motion Data to Boost 3D Motion Generation, Pi et al. ArXiv 2024.
  • Light-T2M. Light-T2M: A Lightweight and Fast Model for Text-to-motion Generation, Zeng et al. AAAI 2025.
  • Languate of Motion. The Language of Motion: Unifying Verbal and Non-verbal Language of 3D Human Motion, Chen et al. ArXiv 2024.
  • Mogo. Mogo: RQ Hierarchical Causal Transformer for High-Quality 3D Human Motion Generation, Fu et al. ArXiv 2024.
  • CoMA. CoMA: Compositional Human Motion Generation with Multi-modal Agents, Sun et al. ArXiv 2024.
  • SoPo. SoPo: Text-to-Motion Generation Using Semi-Online Preference Optimization, Tan et al. ArXiv 2024.
  • RMD. RMD: A Simple Baseline for More General Human Motion Generation via Training-free Retrieval-Augmented Motion Diffuse, Liao et al. ArXiv 2024.
  • BiPO. BiPO: Bidirectional Partial Occlusion Network for Text-to-Motion Synthesis, Hong et al. ArXiv 2024.
  • AtoM. AToM: Aligning Text-to-Motion Model at Event-Level with GPT-4Vision Reward, Han et al. ArXiv 2024.
  • MVLift. Lifting Motion to the 3D World via 2D Diffusion, Li et al. ArXiv 2024.
  • DisCoRD. DisCoRD: Discrete Tokens to Continuous Motion via Rectified Flow Decoding, Cho et al. ArXiv 2024.
  • MoTe. MoTe: Learning Motion-Text Diffusion Model for Multiple Generation Tasks, Wue et al. ArXiv 2024.
  • InfiniDreamer. InfiniDreamer: Arbitrarily Long Human Motion Generation via Segment Score Distillation, Zhuo et al. ArXiv 2024.
  • FTMoMamba. FTMoMamba: Motion Generation with Frequency and Text State Space Models, Li et al. ArXiv 2024.
  • MARDM. Rethinking Diffusion for Text-Driven Human Motion Generation, Meng et al. ArXiv 2024.
  • KinMo. KinMo: Kinematic-aware Human Motion Understanding and Generation, Zhang et al. ArXiv 2024.
  • LLaMo. Human Motion Instruction Tuning, Li et al. ArXiv 2024.
  • Morph. Morph: A Motion-free Physics Optimization Framework for Human Motion Generation, Li et al. ArXiv 2024.
  • KMM. KMM: Key Frame Mask Mamba for Extended Motion Generation, Zhang et al. ArXiv 2024.
  • MotionGPT-2. MotionGPT-2: A General-Purpose Motion-Language Model for Motion Generation and Understanding, Wang et al. ArXiv 2024.
  • Lodge++. Lodge++: High-quality and Long Dance Generation with Vivid Choreography Patterns, Li et al. ArXiv 2024.
  • MotionCLR. MotionCLR: Motion Generation and Training-free Editing via Understanding Attention Mechanisms, Chen et al. ArXiv 2024.
  • MotionGlot. MotionGlot: A Multi-Embodied Motion Generation Model, Harithas et al. ArXiv 2024.
  • LEAD. LEAD: Latent Realignment for Human Motion Diffusion, Andreou et al. ArXiv 2024.
  • Leite et al. Enhancing Motion Variation in Text-to-Motion Models via Pose and Video Conditioned Editing, Leite et al. ArXiv 2024.
  • MotionRL. MotionRL: Align Text-to-Motion Generation to Human Preferences with Multi-Reward Reinforcement Learning, Liu et al. ArXiv 2024.
  • UniMuMo. UniMuMo: Unified Text, Music and Motion Generation, Yang et al. ArXiv 2024.
  • MotionLLM. MotionLLM: Understanding Human Behaviors from Human Motions and Videos, Chen et al. ArXiv 2024.
  • DART. DART: A Diffusion-Based Autoregressive Motion Model for Real-Time Text-Driven Motion Control, Zhao et al. ArXiv 2024.
  • CLoSD. CLoSD: Closing the Loop between Simulation and Diffusion for multi-task character control, Tevet et al. ArXiv 2024.
  • Wang et al. Quo Vadis, Motion Generation? From Large Language Models to Large Motion Models, Wang et al. ArXiv 2024.
  • T2M-X. T2M-X: Learning Expressive Text-to-Motion Generation from Partially Annotated Data, Liu et al. ArXiv 2024.
  • MoRAG. MoRAG -- Multi-Fusion Retrieval Augmented Generation for Human Motion, Shashank et al. ArXiv 2024.
  • Mandelli et al. Generation of Complex 3D Human Motion by Temporal and Spatial Composition of Diffusion Models, Mandelli et al. ArXiv 2024.
  • BAD. BAD: Bidirectional Auto-regressive Diffusion for Text-to-Motion Generation, Hosseyni et al. ArXiv 2024.
  • synNsync. Synergy and Synchrony in Couple Dances, Manukele et al. ArXiv 2024.
  • Dong et al. Word-Conditioned 3D American Sign Language Motion Generation, Dong et al. EMNLP 2024.
  • Text to blind motion. Text to Blind Motion, Kim et al. NeurIPS D&B 2024.
  • UniMTS. UniMTS: Unified Pre-training for Motion Time Series, Zhang et al. NeurIPS 2024.
  • Christopher et al.. Constrained Synthesis with Projected Diffusion Models, Christopher et al. NeurIPS 2024.
  • MoMu-Diffusion. MoMu-Diffusion: On Learning Long-Term Motion-Music Synchronization and Correspondence, You et al. NeurIPS 2024.
  • MoGenTS. MoGenTS: Motion Generation based on Spatial-Temporal Joint Modeling, Yuan et al. NeurIPS 2024.
  • M3GPT. M3GPT: An Advanced Multimodal, Multitask Framework for Motion Comprehension and Generation, Luo et al. NeurIPS 2024.
  • Bikov et al. Fitness Aware Human Motion Generation with Fine-Tuning, Bikov et al. NeurIPS Workshop 2024.
  • FG-MDM. FG-MDM: Towards Zero-Shot Human Motion Generation via ChatGPT-Refined Descriptions, ICPR 2024.
  • SynTalker. Enabling Synergistic Full-Body Control in Prompt-Based Co-Speech Motion Generation, Chen et al. ACM MM 2024.
  • L3EM. Towards Emotion-enriched Text-to-Motion Generation via LLM-guided Limb-level Emotion Manipulating. Yu et al. ACM MM 2024.
  • StableMoFusion. StableMoFusion: Towards Robust and Efficient Diffusion-based Motion Generation Framework, Huang et al. ACM MM 2024.
  • SATO. SATO: Stable Text-to-Motion Framework, Chen et al. ACM MM 2024.
  • PIDM. PIDM: Personality-Aware Interaction Diffusion Model for Gesture Generation, Shibasaki et al. ICANN 2024.
  • Macwan et al. High-Fidelity Worker Motion Simulation With Generative AI, Macwan et al. HFES 2024.
  • Jin et al.. Local Action-Guided Motion Diffusion Model for Text-to-Motion Generation, Jin et al. ECCV 2024.
  • Motion Mamba. Motion Mamba: Efficient and Long Sequence Motion Generation, Zhong et al. ECCV 2024.
  • EMDM. EMDM: Efficient Motion Diffusion Model for Fast, High-Quality Human Motion Generation, Zhou et al. ECCV 2024.
  • CoMo. CoMo: Controllable Motion Generation through Language Guided Pose Code Editing, Huang et al. ECCV 2024.
  • CoMusion. CoMusion: Towards Consistent Stochastic Human Motion Prediction via Motion Diffusion, Sun et al. ECCV 2024.
  • Shan et al.. Towards Open Domain Text-Driven Synthesis of Multi-Person Motions, Shan et al. ECCV 2024.
  • ParCo. ParCo: Part-Coordinating Text-to-Motion Synthesis, Zou et al. ECCV 2024.
  • Sampieri et al.. Length-Aware Motion Synthesis via Latent Diffusion, Sampieri et al. ECCV 2024.
  • ChroAccRet. Chronologically Accurate Retrieval for Temporal Grounding of Motion-Language Models, Fujiwara et al. ECCV 2024.
  • MHC. Generating Physically Realistic and Directable Human Motions from Multi-Modal Inputs, Liu et al. ECCV 2024.
  • ProMotion. Plan, Posture and Go: Towards Open-vocabulary Text-to-Motion Generation, Liu et al. ECCV 2024.
  • FreeMotion. FreeMotion: MoCap-Free Human Motion Synthesis with Multimodal Large Language Models, Zhang et al. ECCV 2024.
  • Text Motion Translator. Text Motion Translator: A Bi-Directional Model for Enhanced 3D Human Motion Generation from Open-Vocabulary Descriptions, Qian et al. ECCV 2024.
  • FreeMotion. FreeMotion: A Unified Framework for Number-free Text-to-Motion Synthesis, Fan et al. ECCV 2024.
  • Kinematic Phrases. Bridging the Gap between Human Motion and Action Semantics via Kinematic Phrases, Liu et al. ECCV 2024.
  • MotionChain. MotionChain: Conversational Motion Controllers via Multimodal Prompts, Jiang et al. ECCV 2024.
  • SMooDi. SMooDi: Stylized Motion Diffusion Model, Zhong et al. ECCV 2024.
  • BAMM. BAMM: Bidirectional Autoregressive Motion Model, Pinyoanuntapong et al. ECCV 2024.
  • MotionLCM. MotionLCM: Real-time Controllable Motion Generation via Latent Consistency Model, Dai et al. ECCV 2024.
  • Ren et al.. Realistic Human Motion Generation with Cross-Diffusion Models, Ren et al. ECCV 2024.
  • M2D2M. M2D2M: Multi-Motion Generation from Text with Discrete Diffusion Models, Chi et al. ECCV 2024.
  • Large Motion Model. Large Motion Model for Unified Multi-Modal Motion Generation, Zhang et al. ECCV 2024.
  • TesMo. Generating Human Interaction Motions in Scenes with Text Control, Yi et al. ECCV 2024.
  • TLcontrol. TLcontrol: Trajectory and Language Control for Human Motion Synthesis, Wan et al. ECCV 2024.
  • ExpGest. ExpGest: Expressive Speaker Generation Using Diffusion Model and Hybrid Audio-Text Guidance, Cheng et al. ICME 2024.
  • Chen et al. Anatomically-Informed Vector Quantization Variational Auto-Encoder for Text-to-Motion Generation, Chen et al. ICME Workshop 2024.
  • HumanTOMATO. HumanTOMATO: Text-aligned Whole-body Motion Generation, Lu et al. ICML 2024.
  • GPHLVM. Bringing Motion Taxonomies to Continuous Domains via GPLVM on Hyperbolic Manifolds, Jaquier et al. ICML 2024.
  • CondMDI. Flexible Motion In-betweening with Diffusion Models, Cohan et al. SIGGRAPH 2024.
  • CAMDM. Taming Diffusion Probabilistic Models for Character Control, Chen et al. SIGGRAPH 2024.
  • LGTM. LGTM: Local-to-Global Text-Driven Human Motion Diffusion Models, Sun et al. SIGGRAPH 2024.
  • TEDi. TEDi: Temporally-Entangled Diffusion for Long-Term Motion Synthesis, Zhang et al. SIGGRAPH 2024.
  • A-MDM. Interactive Character Control with Auto-Regressive Motion Diffusion Models, Shi et al. SIGGRAPH 2024.
  • Starke et al.. Categorical Codebook Matching for Embodied Character Controllers, Starke et al. SIGGRAPH 2024.
  • SuperPADL. SuperPADL: Scaling Language-Directed Physics-Based Control with Progressive Supervised Distillation, Juravsky et al. SIGGRAPH 2024.
  • ProgMoGen. Programmable Motion Generation for Open-set Motion Control Tasks, Liu et al. CVPR 2024.
  • PACER+. PACER+: On-Demand Pedestrian Animation Controller in Driving Scenarios, Wang et al. CVPR 2024.
  • AMUSE. Emotional Speech-driven 3D Body Animation via Disentangled Latent Diffusion, Chhatre et al. CVPR 2024.
  • Liu et al.. Towards Variable and Coordinated Holistic Co-Speech Motion Generation, Liu et al. CVPR 2024.
  • MAS. MAS: Multi-view Ancestral Sampling for 3D motion generation using 2D diffusion, Kapon et al. CVPR 2024.
  • WANDR. WANDR: Intention-guided Human Motion Generation, Diomataris et al. CVPR 2024.
  • MoMask. MoMask: Generative Masked Modeling of 3D Human Motions, Guo et al. CVPR 2024.
  • ChapPose. ChatPose: Chatting about 3D Human Pose, Feng et al. CVPR 2024.
  • AvatarGPT. AvatarGPT: All-in-One Framework for Motion Understanding, Planning, Generation and Beyond, Zhou et al. CVPR 2024.
  • MMM. MMM: Generative Masked Motion Model, Pinyoanuntapong et al. CVPR 2024.
  • AAMDM. AAMDM: Accelerated Auto-regressive Motion Diffusion Model, Li et al. CVPR 2024.
  • OMG. OMG: Towards Open-vocabulary Motion Generation via Mixture of Controllers, Liang et al. CVPR 2024.
  • FlowMDM. FlowMDM: Seamless Human Motion Composition with Blended Positional Encodings, Barquero et al. CVPR 2024.
  • Digital Life Project. Digital Life Project: Autonomous 3D Characters with Social Intelligence, Cai et al. CVPR 2024.
  • EMAGE. EMAGE: Towards Unified Holistic Co-Speech Gesture Generation via Expressive Masked Audio Gesture Modeling, Liu et al. CVPR 2024.
  • STMC. Multi-Track Timeline Control for Text-Driven 3D Human Motion Generation, Petrovich et al. CVPR Workshop 2024.
  • InstructMotion. Exploring Text-to-Motion Generation with Human Preference, Sheng et al. CVPR Workshop 2024.
  • Single Motion Diffusion. Raab et al. ICLR 2024.
  • NeRM. NeRM: Learning Neural Representations for High-Framerate Human Motion Synthesis, Wei et al. ICLR 2024.
  • PriorMDM. PriorMDM: Human Motion Diffusion as a Generative Prior, Shafir et al. ICLR 2024.
  • OmniControl. OmniControl: Control Any Joint at Any Time for Human Motion Generation, Xie et al. ICLR 2024.
  • Adiya et al.. Bidirectional Temporal Diffusion Model for Temporally Consistent Human Animation, Adiya et al. ICLR 2024.
  • Duolando. Duolando: Follower GPT with Off-Policy Reinforcement Learning for Dance Accompaniment, Li et al. ICLR 2024.
  • HuTuDiffusion. HuTuMotion: Human-Tuned Navigation of Latent Motion Diffusion Models with Minimal Feedback, Han et al. AAAI 2024.
  • AMD. AMD: Anatomical Motion Diffusion with Interpretable Motion Decomposition and Fusion, Jing et al. AAAI 2024.
  • MotionMix. MotionMix: Weakly-Supervised Diffusion for Controllable Motion Generation, Hoang et al. AAAI 2024.
  • B2A-HDM. Towards Detailed Text-to-Motion Synthesis via Basic-to-Advanced Hierarchical Diffusion Model, Xie et al. AAAI 2024.
  • Everything2Motion. Everything2Motion: Synchronizing Diverse Inputs via a Unified Framework for Human Motion Synthesis, Fan et al. AAAI 2024.
  • MotionGPT. MotionGPT: Finetuned LLMs are General-Purpose Motion Generators, Zhang et al. AAAI 2024.
  • Dong et al. Enhanced Fine-grained Motion Diffusion for Text-driven Human Motion Synthesis, Dong et al. AAAI 2024.
  • UNIMASKM. A Unified Masked Autoencoder with Patchified Skeletons for Motion Synthesis, Mascaro et al. AAAI 2024.
  • B2A-HDM. Towards Detailed Text-to-Motion Synthesis via Basic-to-Advanced Hierarchical Diffusion Model, Xie et al. AAAI 2024.
  • GUESS. GUESS: GradUally Enriching SyntheSis for Text-Driven Human Motion Generation, Gao et al. TPAMI 2024.
  • Xie et al.. Sign Language Production with Latent Motion Transformer, Xie et al. WACV 2024.
  • GraphMotion. Act As You Wish: Fine-grained Control of Motion Diffusion Model with Hierarchical Semantic Graphs, Jin et al. NeurIPS 2023.
  • MotionGPT. MotionGPT: Human Motion as Foreign Language, Jiang et al. NeurIPS 2023.
  • FineMoGen. FineMoGen: Fine-Grained Spatio-Temporal Motion Generation and Editing, Zhang et al. NeurIPS 2023.
  • InsActor. InsActor: Instruction-driven Physics-based Characters, Ren et al. NeurIPS 2023.
  • AttT2M. AttT2M: Text-Driven Human Motion Generation with Multi-Perspective Attention Mechanism, Zhong et al. ICCV 2023.
  • TMR. TMR: Text-to-Motion Retrieval Using Contrastive 3D Human Motion Synthesis, Petrovich et al. ICCV 2023.
  • MAA. Make-An-Animation: Large-Scale Text-conditional 3D Human Motion Generation, Azadi et al. ICCV 2023.
  • PhysDiff. PhysDiff: Physics-Guided Human Motion Diffusion Model, Yuan et al. ICCV 2023.
  • ReMoDiffusion. ReMoDiffuse: Retrieval-Augmented Motion Diffusion Model, Zhang et al. ICCV 2023.
  • BelFusion. BeLFusion: Latent Diffusion for Behavior-Driven Human Motion Prediction, Barquero et al. ICCV 2023.
  • GMD. GMD: Guided Motion Diffusion for Controllable Human Motion Synthesis, Karunratanakul et al. ICCV 2023.
  • HMD-NeMo. HMD-NeMo: Online 3D Avatar Motion Generation From Sparse Observations, Aliakbarian et al. ICCV 2023.
  • SINC. SINC: Spatial Composition of 3D Human Motions for Simultaneous Action Generation, Athanasiou et al. ICCV 2023.
  • Kong et al.. Priority-Centric Human Motion Generation in Discrete Latent Space, Kong et al. ICCV 2023.
  • FgT2M. Fg-T2M: Fine-Grained Text-Driven Human Motion Generation via Diffusion Model, Wang et al. ICCV 2023.
  • EMS. Breaking The Limits of Text-conditioned 3D Motion Synthesis with Elaborative Descriptions, Qian et al. ICCV 2023.
  • GenMM. Example-based Motion Synthesis via Generative Motion Matching, Li et al. SIGGRAPH 2023.
  • GestureDiffuCLIP. GestureDiffuCLIP: Gesture Diffusion Model with CLIP Latents, Ao et al. SIGGRAPH 2023.
  • BodyFormer. BodyFormer: Semantics-guided 3D Body Gesture Synthesis with Transformer, Pang et al. SIGGRAPH 2023.
  • Alexanderson et al.. Listen, denoise, action! Audio-driven motion synthesis with diffusion models, Alexanderson et al. SIGGRAPH 2023.
  • AGroL. Avatars Grow Legs: Generating Smooth Human Motion from Sparse Tracking Inputs with Diffusion Model, Du et al. CVPR 2023.
  • TALKSHOW. Generating Holistic 3D Human Motion from Speech, Yi et al. CVPR 2023.
  • T2M-GPT. T2M-GPT: Generating Human Motion from Textual Descriptions with Discrete Representations, Zhang et al. CVPR 2023.
  • UDE. UDE: A Unified Driving Engine for Human Motion Generation, Zhou et al. CVPR 2023.
  • OOHMG. Being Comes from Not-being: Open-vocabulary Text-to-Motion Generation with Wordless Training, Lin et al. CVPR 2023.
  • EDGE. EDGE: Editable Dance Generation From Music, Tseng et al. CVPR 2023.
  • MLD. Executing your Commands via Motion Diffusion in Latent Space, Chen et al. CVPR 2023.
  • MoDi. MoDi: Unconditional Motion Synthesis from Diverse Data, Raab et al. CVPR 2023.
  • MoFusion. MoFusion: A Framework for Denoising-Diffusion-based Motion Synthesis, Dabral et al. CVPR 2023.
  • Mo et al.. Continuous Intermediate Token Learning with Implicit Motion Manifold for Keyframe Based Motion Interpolation, Mo et al. CVPR 2023.
  • HMDM. MDM: Human Motion Diffusion Model, Tevet et al. ICLR 2023.
  • MotionDiffuse. MotionDiffuse: Text-Driven Human Motion Generation with Diffusion Model, Zhang et al. TPAMI 2023.
  • Bailando++. Bailando++: 3D Dance GPT with Choreographic Memory, Li et al. TPAMI 2023.
  • UDE-2. A Unified Framework for Multimodal, Multi-Part Human Motion Synthesis, Zhou et al. ArXiv 2023.
  • Motion Script. MotionScript: Natural Language Descriptions for Expressive 3D Human Motions, Yazdian et al. ArXiv 2023.
  • NeMF. NeMF: Neural Motion Fields for Kinematic Animation, He et al. NeurIPS 2022.
  • PADL. PADL: Language-Directed Physics-Based Character, Juravsky et al. SIGGRAPH Asia 2022.
  • Rhythmic Gesticulator. Rhythmic Gesticulator: Rhythm-Aware Co-Speech Gesture Synthesis with Hierarchical Neural Embeddings, Ao et al. SIGGRAPH Asia 2022.
  • TEACH. TEACH: Temporal Action Composition for 3D Human, Athanasiou et al. 3DV 2022.
  • Implicit Motion. Implicit Neural Representations for Variable Length Human Motion Generation, Cervantes et al. ECCV 2022.
  • Zhong et al.. Learning Uncoupled-Modulation CVAE for 3D Action-Conditioned Human Motion Synthesis, Zhong et al. ECCV 2022.
  • MotionCLIP. MotionCLIP: Exposing Human Motion Generation to CLIP Space, Tevet et al. ECCV 2022.
  • PoseGPT. PoseGPT: Quantizing human motion for large scale generative modeling, Lucas et al. ECCV 2022.
  • TEMOS. TEMOS: Generating diverse human motions from textual descriptions, Petrovich et al. ECCV 2022.
  • TM2T. TM2T: Stochastic and Tokenized Modeling for the Reciprocal Generation of 3D Human Motions and Texts, Guo et al. ECCV 2022.
  • AvatarCLIP. AvatarCLIP: Zero-Shot Text-Driven Generation and Animation of 3D Avatars, Hong et al. SIGGRAPH 2022.
  • DeepPhase. Deepphase: Periodic autoencoders for learning motion phase manifolds, Starke et al. SIGGRAPH 2022.
  • Guo et al.. Generating Diverse and Natural 3D Human Motions from Text, Guo et al. CVPR 2022
  • Bailando. Bailando: 3D Dance Generation by Actor-Critic GPT with Choreographic Memory, Li et al. CVPR 2022.
  • ACTOR. Action-Conditioned 3D Human Motion Synthesis with Transformer VAE, Petrovich et al. ICCV 2021.
  • AIST++. AI Choreographer: Music Conditioned 3D Dance Generation with AIST++, Li et al. ICCV 2021.
  • Starke et al.. Neural animation layering for synthesizing martial arts movements, Starke et al. SIGGRAPH 2021.
  • MOJO. We are More than Our Joints: Predicting how 3D Bodies Move, Zhang et al. CVPR 2021.
  • DLow. DLow: Diversifying Latent Flows for Diverse Human Motion Prediction, Yuan et al. ECCV 2020.
  • Starke et al.. Local motion phases for learning multi-contact character movements, Starke et al. SIGGRAPH 2020.

Motion Editing

  • MotionFix. MotionFix: Text-Driven 3D Human Motion Editing, Athanasiou et al. SIGGRAPH Asia 2024.
  • CigTime. CigTime: Corrective Instruction Generation Through Inverse Motion Editing, Fang et al. NeurIPS 2024.
  • Iterative Motion Editing. Iterative Motion Editing with Natural Language, Goel et al. SIGGRAPH 2024.
  • DNO. DNO: Optimizing Diffusion Noise Can Serve As Universal Motion Priors, Karunratanakul et al. CVPR 2024.

Motion Stylization

  • MulSMo. MulSMo: Multimodal Stylized Motion Generation by Bidirectional Control Flow, Li et al. ArXiv 2024.
  • D-LORD. D-LORD for Motion Stylization, Gupta et al. TSMC 2024.
  • HUMOS. HUMOS: Human Motion Model Conditioned on Body Shape, Tripathi et al. ECCV 2024.
  • SMEAR. SMEAR: Stylized Motion Exaggeration with ARt-direction, Basset et al. SIGGRAPH 2024.
  • MCM-LDM. Arbitrary Motion Style Transfer with Multi-condition Motion Latent Diffusion Model, Song et al. CVPR 2024.
  • MoST. MoST: Motion Style Transformer between Diverse Action Contents, Kim et al. CVPR 2024.
  • GenMoStyle. Generative Human Motion Stylization in Latent Space, Guo et al. ICLR 2024.

Human-Object Interaction

  • DiffGrasp. Diffgrasp: Whole-Body Grasping Synthesis Guided by Object Motion Using a Diffusion Model, Zhang et al. AAAI 2025.
  • InterTrack. InterTrack: Tracking Human Object Interaction without Object Templates, Xie et al. 3DV 2025.
  • FORCE. FORCE: Dataset and Method for Intuitive Physics Guided Human-object Interaction, Zhang et al. 3DV 2025.
  • SyncDiff. SyncDiff: Synchronized Motion Diffusion for Multi-Body Human-Object Interaction Synthesis, He et al. ArXiv 2024.
  • CHOICE. CHOICE: Coordinated Human-Object Interaction in Cluttered Environments for Pick-and-Place Actions, Lu et al. ArXiv 2024.
  • TriDi. TriDi: Trilateral Diffusion of 3D Humans, Objects and Interactions, Petrov et al. ArXiv 2024.
  • OOD-HOI. OOD-HOI: Text-Driven 3D Whole-Body Human-Object Interactions Generation Beyond Training Domains, Zhang et al. ArXiv 2024.
  • COLLAGE. COLLAGE: Collaborative Human-Agent Interaction Generation using Hierarchical Latent Diffusion and Language Models, Daiya et al. ArXiv 2024.
  • SMGDiff. SMGDiff: Soccer Motion Generation using diffusion probabilistic models, Yang et al. ArXiv 2024.
  • SkillMimic. SkillMimic: Learning Reusable Basketball Skills from Demonstrations, Wang et al. ArXiv 2024.
  • CORE4D. CORE4D: A 4D Human-Object-Human Interaction Dataset for Collaborative Object REarrangement, Zhang et al. ArXiv 2024.
  • Wu et al. Human-Object Interaction from Human-Level Instructions, Wu et al. ArXiv 2024.
  • HumanVLA. HumanVLA: Towards Vision-Language Directed Object Rearrangement by Physical Humanoid, Xu et al. NeurIPS 2024.
  • OmniGrasp. Grasping Diverse Objects with Simulated Humanoids, Luo et al. NeurIPS 2024.
  • EgoChoir. EgoChoir: Capturing 3D Human-Object Interaction Regions from Egocentric Views, Yang et al. NeurIPS 2024.
  • CooHOI. CooHOI: Learning Cooperative Human-Object Interaction with Manipulated Object Dynamics, Gao et al. NeurIPS 2024.
  • InterDreamer. InterDreamer: Zero-Shot Text to 3D Dynamic Human-Object Interaction, Xu et al. NeurIPS 2024.
  • PiMForce. Posture-Informed Muscular Force Learning for Robust Hand Pressure Estimation, Seo et al. NeurIPS 2024.
  • InterFusion. InterFusion: Text-Driven Generation of 3D Human-Object Interaction, Dai et al. ECCV 2024.
  • CHOIS. Controllable Human-Object Interaction Synthesis, Li et al. ECCV 2024.
  • F-HOI. F-HOI: Toward Fine-grained Semantic-Aligned 3D Human-Object Interactions, Yang et al. ECCV 2024.
  • HIMO. HIMO: A New Benchmark for Full-Body Human Interacting with Multiple Objects, Lv et al. ECCV 2024.
  • PhysicsPingPong. Strategy and Skill Learning for Physics-based Table Tennis Animation, Wang et al. SIGGRAPH 2024.
  • NIFTY. NIFTY: Neural Object Interaction Fields for Guided Human Motion Synthesis, Kulkarni et al. CVPR 2024.
  • HOI Animator. HOIAnimator: Generating Text-prompt Human-object Animations using Novel Perceptive Diffusion Models, Son et al. CVPR 2024.
  • CG-HOI. CG-HOI: Contact-Guided 3D Human-Object Interaction Generation, Diller et al. CVPR 2024.
  • InterCap. InterCap: Joint Markerless 3D Tracking of Humans and Objects in Interaction, Huang et al. IJCV 2024.
  • Phys-Fullbody-Grasp. Physically Plausible Full-Body Hand-Object Interaction Synthesis, Braun et al. 3DV 2024.
  • GRIP. GRIP: Generating Interaction Poses Using Spatial Cues and Latent Consistency, Taheri et al. 3DV 2024.
  • FAVOR. Favor: Full-Body AR-driven Virtual Object Rearrangement Guided by Instruction Text, Li et al. AAAI 2024.
  • OMOMO. Object Motion Guided Human Motion Synthesis, Li et al. SIGGRAPH Asia 2023.
  • CHAIRS. Full-Body Articulated Human-Object Interaction, Jiang et al. ICCV 2023.
  • HGHOI. Hierarchical Generation of Human-Object Interactions with Diffusion Probabilistic Models, Pi et al. ICCV 2023.
  • InterDiff. InterDiff: Generating 3D Human-Object Interactions with Physics-Informed Diffusion, Xu et al. ICCV 2023.
  • Object Pop Up. Object pop-up: Can we infer 3D objects and their poses from human interactions alone? Petrov et al. CVPR 2023.
  • ARCTIC. A Dataset for Dexterous Bimanual Hand-Object Manipulation, Fan et al. CVPR 2023.
  • TOCH. TOCH: Spatio-Temporal Object-to-Hand Correspondence for Motion Refinement, Zhou et al. ECCV 2022.
  • COUCH. COUCH: Towards Controllable Human-Chair Interactions, Zhang et al. ECCV 2022.
  • SAGA. SAGA: Stochastic Whole-Body Grasping with Contact, Wu et al. ECCV 2022.
  • GOAL. GOAL: Generating 4D Whole-Body Motion for Hand-Object Grasping, Taheri et al. CVPR 2022.
  • BEHAVE. BEHAVE: Dataset and Method for Tracking Human Object Interactions, Bhatnagar et al. CVPR 2022.
  • GRAB. GRAB: A Dataset of Whole-Body Human Grasping of Objects, Taheri et al. ECCV 2020.

Human-Scene Interaction

  • ZeroHSI. ZeroHSI: Zero-Shot 4D Human-Scene Interaction by Video Generation, Li et al. ArXiv 2024.
  • Mimicking-Bench. Mimicking-Bench: A Benchmark for Generalizable Humanoid-Scene Interaction Learning via Human Mimicking, Liu et al. ArXiv 2024.
  • SCENIC. SCENIC: Scene-aware Semantic Navigation with Instruction-guided Control, Zhang et al. ArXiv 2024.
  • Sitcom-Crafter. Sitcom-Crafter: A Plot-Driven Human Motion Generation System in 3D Scenes, Chen et al. ArXiv 2024.
  • Paschalidis et al. 3D Whole-body Grasp Synthesis with Directional Controllability, Paschalidis et al. ArXiv 2024.
  • EnvPoser. EnvPoser: Environment-aware Realistic Human Motion Estimation from Sparse Observations with Uncertainty Modeling. Xia et al. ArXiv 2024.
  • Diffusion Implicit Policy, Diffusion Implicit Policy for Unpaired Scene-aware Motion synthesis, Gong et al. ArXiv 2024.
  • SIMS. SIMS: Simulating Human-Scene Interactions with Real World Script Planning, Wang et al. ArXiv 2024.
  • LaserHuman. LaserHuman: Language-guided Scene-aware Human Motion Generation in Free Environment, Cong et al. ArXiv 2024.
  • LINGO. Autonomous Character-Scene Interaction Synthesis from Text Instruction, Jiang et al. SIGGRAPH Asia 2024.
  • DiMoP3D. Harmonizing Stochasticity and Determinism: Scene-responsive Diverse Human Motion Prediction, Lou et al. NeurIPS 2024.
  • Liu et al.. Revisit Human-Scene Interaction via Space Occupancy, Liu et al. ECCV 2024.
  • TesMo. Generating Human Interaction Motions in Scenes with Text Control, Yi et al. ECCV 2024.
  • SAST. Massively Multi-Person 3D Human Motion Forecasting with Scene Context, Mueller et al. ECCV 2024 Workshop.
  • Kang et al. Learning Climbing Controllers for Physics-Based Characters, Kang et al. Eurographics 2024.
  • Afford-Motion. Move as You Say, Interact as You Can: Language-guided Human Motion Generation with Scene Affordance, Wang et al. CVPR 2024.
  • GenZI. GenZI: Zero-Shot 3D Human-Scene Interaction Generation, Li et al. CVPR 2024.
  • Cen et al.. Generating Human Motion in 3D Scenes from Text Descriptions, Cen et al. CVPR 2024.
  • TRUMANS. Scaling Up Dynamic Human-Scene Interaction Modeling, Jiang et al. CVPR 2024.
  • UniHSI. UniHSI: Unified Human-Scene Interaction via Prompted Chain-of-Contacts, Xiao et al. ICLR 2024.
  • Purposer. Purposer: Putting Human Motion Generation in Context, Ugrinovic et al. 3DV 2024.
  • InterScene. Synthesizing Physically Plausible Human Motions in 3D Scenes, Pan et al. 3DV 2024.
  • Mir et al. Generating Continual Human Motion in Diverse 3D Scenes, Mir et al. 3DV 2024.
  • DIMOS. DIMOS: Synthesizing Diverse Human Motions in 3D Indoor Scenes, Zhao et al. ICCV 2023.
  • LAMA. Locomotion-Action-Manipulation: Synthesizing Human-Scene Interactions in Complex 3D Environments, Lee et al. ICCV 2023.
  • Narrator. Narrator: Towards Natural Control of Human-Scene Interaction Generation via Relationship Reasoning, Xuan et al. ICCV 2023.
  • CIMI4D. CIMI4D: A Large Multimodal Climbing Motion Dataset under Human-scene Interactions, Yan et al. CVPR 2023.
  • Scene-Ego. Scene-aware Egocentric 3D Human Pose Estimation, Wang et al. CVPR 2023.
  • SLOPER4D. SLOPER4D: A Scene-Aware Dataset for Global 4D Human Pose Estimation in Urban Environments, Dai et al. CVPR 2023.
  • CIRCLE. CIRCLE: Capture in Rich Contextual Environments, Araujo et al. CVPR 2023.
  • SceneDiffuser. Diffusion-based Generation, Optimization, and Planning in 3D Scenes, Huang et al. CVPR 2023.
  • PMP. PMP: Learning to Physically Interact with Environments using Part-wise Motion Priors, Bae et al. SIGGRAPH 2023.
  • QuestEnvSim. QuestEnvSim: Environment-Aware Simulated Motion Tracking from Sparse Sensors, Lee et al. SIGGRAPH 2023.
  • Hassan et al. Synthesizing Physical Character-Scene Interactions, Hassan et al. SIGGRAPH 2023.
  • Mao et al.. Contact-aware Human Motion Forecasting, Mao et al. NeurIPS 2022.
  • HUMANISE. HUMANISE: Language-conditioned Human Motion Generation in 3D Scenes, Wang et al. NeurIPS 2022.
  • EmbodiedPose. Embodied Scene-aware Human Pose Estimation, Luo et al. NeurIPS 2022.
  • GIMO. GIMO: Gaze-Informed Human Motion Prediction in Context, Zheng et al. ECCV 2022.
  • COINS. COINS: Compositional Human-Scene Interaction Synthesis with Semantic Control, Zhao et al. ECCV 2022.
  • Wang et al.. Towards Diverse and Natural Scene-aware 3D Human Motion Synthesis, Wang et al. CVPR 2022.
  • GAMMA. The Wanderings of Odysseus in 3D Scenes, Zhang et al. CVPR 2022.
  • SAMP. Stochastic Scene-Aware Motion Prediction, Hassan et al, ICCV 2021.
  • LEMO. Learning Motion Priors for 4D Human Body Capture in 3D Scenes, Zhang et al. ICCV 2021.
  • PLACE. PLACE: Proximity Learning of Articulation and Contact in 3D Environments, Zhang et al. 3DV 2020.
  • Starke et al.. Local motion phases for learning multi-contact character movements, Starke et al. SIGGRAPH 2020.
  • PSI. Generating 3D People in Scenes without People, Zhang et al. CVPR 2020.
  • NSM. Neural State Machine for Character-Scene Interactions, Starke et al. SIGGRAPH Asia 2019.
  • PROX. Resolving 3D Human Pose Ambiguities with 3D Scene Constraints, Hassan et al. ICCV 2019

Human-Human Interaction

  • Two in One. Two-in-One: Unified Multi-Person Interactive Motion Generation by Latent Diffusion Transformer, Li et al. ArXiv 2024.
  • It Takes Two. It Takes Two: Real-time Co-Speech Two-person’s Interaction Generation via Reactive Auto-regressive Diffusion Model, Shi et al. ArXiv 2024.
  • InterMask. InterMask: 3D Human Interaction Generation via Collaborative Masked Modelling, Javed et al. ArXiv 2024.
  • COLLAGE. COLLAGE: Collaborative Human-Agent Interaction Generation using Hierarchical Latent Diffusion and Language Models, Daiya et al. ArXiv 2024.
  • Harmony4D. Harmony4D: A Video Dataset for In-The-Wild Close Human Interactions, Khirodkar et al. NeurIPS 2024.
  • InterControl. InterControl: Generate Human Motion Interactions by Controlling Every Joint, Wang et al. NeurIPS 2024.
  • PhysReaction. PhysReaction: Physically Plausible Real-Time Humanoid Reaction Synthesis via Forward Dynamics Guided 4D Imitation, Liu et al. ACM MM 2024.
  • Shan et al.. Towards Open Domain Text-Driven Synthesis of Multi-Person Motions, Shan et al. ECCV 2024.
  • ReMoS. ReMoS: 3D Motion-Conditioned Reaction Synthesis for Two-Person Interactions, Ghosh et al. ECCV 2024.
  • Inter-X. Inter-X: Towards Versatile Human-Human Interaction Analysis, Xu et al. CVPR 2024.
  • ReGenNet. ReGenNet: Towards Human Action-Reaction Synthesis, Xu et al. CVPR 2024.
  • Fang et al.. Capturing Closely Interacted Two-Person Motions with Reaction Priors, Fan et al. CVPR 2024.
  • in2IN. in2IN: Leveraging Individual Information to Generate Human INteractions, Ruiz-Ponce et al. CVPR Workshop 2024.
  • InterGen. InterGen: Diffusion-based Multi-human Motion Generation under Complex Interactions, Liang et al. IJCV 2024.
  • ActFormer. ActFormer: A GAN-based Transformer towards General Action-Conditioned 3D Human Motion Generation, Xu et al. ICCV 2023.
  • Tanaka et al.. Role-aware Interaction Generation from Textual Description, Tanaka et al. ICCV 2023.
  • Hi4D. Hi4D: 4D Instance Segmentation of Close Human Interaction, Yin et al. CVPR 2023.
  • ExPI. Multi-Person Extreme Motion Prediction, Guo et al. CVPR 2022.
  • CHI3D. Three-Dimensional Reconstruction of Human Interactions, Fieraru et al. CVPR 2020.

Datasets & Benchmarks

  • Mimicking-Bench. Mimicking-Bench: A Benchmark for Generalizable Humanoid-Scene Interaction Learning via Human Mimicking, Liu et al. ArXiv 2024.
  • SCENIC. SCENIC: Scene-aware Semantic Navigation with Instruction-guided Control, Zhang et al. ArXiv 2024.
  • CMP & CMR. AnimationGPT:An AIGC tool for generating game combat motion assets, Liao et al.
  • LINGO. Autonomous Character-Scene Interaction Synthesis from Text Instruction, Jiang et al. SIGGRAPH Asia 2024.
  • LaserHuman. LaserHuman: Language-guided Scene-aware Human Motion Generation in Free Environment, Cong et al. ArXiv 2024.
  • AtoM. AToM: Aligning Text-to-Motion Model at Event-Level with GPT-4Vision Reward, Han et al. ArXiv 2024.
  • Evans et al. Synchronized Video, Motion Capture and Force Plate Dataset for Validating Markerless Human Movement Analysis, Evans et al. Scientific Data 2024.
  • MotionCritic. Aligning Human Motion Generation with Human Perceptions, Wang et al. ArXiv 2024.
  • EMHI. EMHI: A Multimodal Egocentric Human Motion Dataset with HMD and Body-Worn IMUs, Fan et al. ArXiv 2024.
  • Harmony4D. Harmony4D: A Video Dataset for In-The-Wild Close Human Interactions, Khirodkar et al. NeurIPS 2024.
  • EgoSim. EgoSim: An Egocentric Multi-view Simulator for Body-worn Cameras during Human Motion, Hollidt et al. NeurIPS D&B 2024.
  • synNsync. Synergy and Synchrony in Couple Dances, Manukele et al. ArXiv 2024.
  • Muscles in Time. Muscles in Time: Learning to Understand Human Motion by Simulating Muscle Activations, Schneider et al. NeurIPS D&B 2024.
  • Text to blind motion. Text to blind motion, Kim et al. NeurIPS D&B 2024.
  • MotionBank. MotionBank: A Large-scale Video Motion Benchmark with Disentangled Rule-based Annotations, Xu et al. ArXiv 2024.
  • CORE4D. CORE4D: A 4D Human-Object-Human Interaction Dataset for Collaborative Object REarrangement, Zhang et al. ArXiv 2024.
  • CLaM. CLaM: An Open-Source Library for Performance Evaluation of Text-driven Human Motion Generation, Chen et al. ACM MM 2024.
  • AddBiomechanics. AddBiomechanics Dataset: Capturing the Physics of Human Motion at Scale, Werling et al. ECCV 2024.
  • LiveHPS++. LiveHPS++: Robust and Coherent Motion Capture in Dynamic Free Environment, Ren et al. ECCV 2024.
  • SignAvatars. SignAvatars: A Large-scale 3D Sign Language Holistic Motion Dataset and Benchmark, Yu et al. ECCV 2024.
  • Nymeria. Nymeria: A massive collection of multimodal egocentric daily motion in the wild, Ma et al. ECCV 2024.
  • Human3.6M+. Using musculoskeletal models to generate physically-consistent data for 3D human pose, kinematic, dynamic, and muscle estimation, Nasr et al. Multibody System Dynamics 2024.
  • Inter-X. Inter-X: Towards Versatile Human-Human Interaction Analysis, Xu et al. CVPR 2024.
  • HardMo. HardMo: ALarge-Scale Hardcase Dataset for Motion Capture, Liao et al. CVPR 2024.
  • MMVP. MMVP: A Multimodal MoCap Dataset with Vision and Pressure Sensors, Zhang et al. CVPR 2024.
  • RELI11D. RELI11D: A Comprehensive Multimodal Human Motion Dataset and Method, Yan et al. CVPR 2024.
  • GroundLink. GroundLink: A Dataset Unifying Human Body Movement and Ground Reaction Dynamics, Han et al. SIGGRAPH Asia 2023.
  • HOH. HOH: Markerless Multimodal Human-Object-Human Handover Dataset with Large Object Count, Wiederhold et al. NeurIPS D&B 2023.
  • Motion-X. Motion-X: A Large-scale 3D Expressive Whole-body Human Motion Dataset, Lin et al. NeurIPS D&B 2023.
  • Humans in Kitchens. Humans in Kitchens: A Dataset for Multi-Person Human Motion Forecasting with Scene Context, Tanke et al. NeurIPS D&B 2023.
  • CHAIRS. Full-Body Articulated Human-Object Interaction, Jiang et al. ICCV 2023.
  • CIMI4D. CIMI4D: A Large Multimodal Climbing Motion Dataset under Human-scene Interactions, Yan et al. CVPR 2023.
  • FLAG3D. FLAG3D: A 3D Fitness Activity Dataset with Language Instruction, Tang et al. CVPR 2023.
  • Hi4D. Hi4D: 4D Instance Segmentation of Close Human Interaction, Yin et al. CVPR 2023.
  • CIRCLE. CIRCLE: Capture in Rich Contextual Environments, Araujo et al. CVPR 2023.
  • MoCapAct. MoCapAct: A Multi-Task Dataset for Simulated Humanoid Control, Wagener et al. NeurIPS 2022.
  • ForcePose. Learning to Estimate External Forces of Human Motion in Video, Louis et al. ACM MM 2022.
  • BEAT. BEAT: A Large-Scale Semantic and Emotional Multi-Modal Dataset for Conversational Gestures Synthesis, Liu et al. ECCV 2022.
  • BRACE. BRACE: The Breakdancing Competition Dataset for Dance Motion Synthesis, Moltisanti et al. ECCV 2022.
  • EgoBody. Egobody: Human body shape and motion of interacting people from head-mounted devices, Zhang et al. ECCV 2022.
  • GIMO. GIMO: Gaze-Informed Human Motion Prediction in Context, Zheng et al. ECCV 2022.
  • HuMMan. HuMMan: Multi-Modal 4D Human Dataset for Versatile Sensing and Modeling, Cai et al. ECCV 2022.
  • ExPI. Multi-Person Extreme Motion Prediction, Guo et al. CVPR 2022.
  • HumanML3D. Generating Diverse and Natural 3D Human Motions from Text, Guo et al. CVPR 2022.
  • BEHAVE. BEHAVE: Dataset and Method for Tracking Human Object Interactions, Bhatnagar et al. CVPR 2022.
  • AIST++. AI Choreographer: Music Conditioned 3D Dance Generation with AIST++, Li et al. ICCV 2021.
  • Fit3D. AIFit: Automatic 3D Human-Interpretable Feedback Models for Fitness Training, Fieraru et al. CVPR 2021.
  • BABEL. BABEL: Bodies, Action and Behavior with English Labels, Punnakkal et al. CVPR 2021
  • HumanSC3D. Learning complex 3d human self-contact, Fieraru et al. AAAI 2021.
  • CHI3D. Three-Dimensional Reconstruction of Human Interactions, Fieraru et al. CVPR 2020.
  • PROX. Resolving 3D Human Pose Ambiguities with 3D Scene Constraints, Hassan et al. ICCV 2019
  • AMASS. AMASS: Archive of Motion Capture As Surface Shapes, Mahmood et al. ICCV 2019

Humanoid, Simulated or Real

  • UH-1. Learning from Massive Human Videos for Universal Humanoid Pose Control, Mao et al. ArXiv 2024.
  • Mimicking-Bench. Mimicking-Bench: A Benchmark for Generalizable Humanoid-Scene Interaction Learning via Human Mimicking, Liu et al. ArXiv 2024.
  • Liu et al. Human-Humanoid Robots Cross-Embodiment Behavior-Skill Transfer Using Decomposed Adversarial Learning from Demonstration, Liu et al. RA-L 2024.
  • UH-1. Learning from Massive Human Videos for Universal Humanoid Pose Control, Mao et al. ArXiv 2024.
  • Exbody2. Exbody2: Advanced Expressive Humanoid Whole-Body Control, Ji et al. ArXiv 2024.
  • SIMS. SIMS: Simulating Human-Scene Interactions with Real World Script Planning, Wang et al. ArXiv 2024.
  • HOVER. HOVER: Versatile Neural Whole-Body Controller for Humanoid Robots, He et al. ArXiv 2024.
  • CLoSD. CLoSD: Closing the Loop between Simulation and Diffusion for multi-task character control, Tevet et al. ArXiv 2024.
  • Humanoidlympics. Humanoidlympics: Sports Environments for Physically Simulated Humanoids, Luo et al. ArXiv 2024.
  • SkillMimic. SkillMimic: Learning Reusable Basketball Skills from Demonstrations, Wang et al. ArXiv 2024.
  • PDP. PDP: Physics-Based Character Animation via Diffusion Policy, Truong et al. SIGGRAPH Asia 2024.
  • MaskedMimic. MaskedMimic: Unified Physics-Based Character Control Through Masked Motion Inpainting, Tessler et al, SIGGRAPH Asia 2024.
  • HumanVLA. HumanVLA: Towards Vision-Language Directed Object Rearrangement by Physical Humanoid, Xu et al. NeurIPS 2024.
  • OmniGrasp. Grasping Diverse Objects with Simulated Humanoids, Luo et al. NeurIPS 2024.
  • InterControl. InterControl: Generate Human Motion Interactions by Controlling Every Joint, Wang et al. NeurIPS 2024.
  • CooHOI. CooHOI: Learning Cooperative Human-Object Interaction with Manipulated Object Dynamics, Gao et al. NeurIPS 2024.
  • Radosavovic et al.. Humanoid Locomotion as Next Token Prediction, Radosavovic et al. NeurIPS 2024.
  • HARMON. Harmon: Whole-Body Motion Generation of Humanoid Robots from Language Descriptions, Jiang et al. CoRL 2024.
  • OKAMI. OKAMI: Teaching Humanoid Robots Manipulation Skills through Single Video Imitation, Li et al. CoRL 2024.
  • HumanPlus. HumanPlus: Humanoid Shadowing and Imitation from Humans, Fu et al. CoRL 2024.
  • OmniH2O. OmniH2O: Universal and Dexterous Human-to-Humanoid Whole-Body Teleoperation and Learning, He et al. CoRL 2024.
  • Self-Aware. Know your limits! Optimize the behavior of bipedal robots through self-awareness, Mascaro et al. Humanoids 2024.
  • PhysReaction. PhysReaction: Physically Plausible Real-Time Humanoid Reaction Synthesis via Forward Dynamics Guided 4D Imitation, Liu et al. ACM MM 2024.
  • H2O. Learning Human-to-Humanoid Real-Time Whole-Body Teleoperation, He et al. IROS 2024.
  • MHC. Generating Physically Realistic and Directable Human Motions from Multi-Modal Inputs, Shrestha et al. ECCV 2024.
  • DIDI. DIDI: Diffusion-Guided Diversity for Offline Behavioral Generation, Liu et al. ICML 2024.
  • MoConVQ. MoConVQ: Unified Physics-Based Motion Control via Scalable Discrete Representations, Yao et al. SIGGRAPH 2024.
  • PhysicsPingPong. Strategy and Skill Learning for Physics-based Table Tennis Animation, Wang et al. SIGGRAPH 2024.
  • SuperPADL. SuperPADL: Scaling Language-Directed Physics-Based Control with Progressive Supervised Distillation, Juravsky et al. SIGGRAPH 2024.
  • SimXR. Real-Time Simulated Avatar from Head-Mounted Sensors, Luo et al. CVPR 2024.
  • AnySkill. AnySkill: Learning Open-Vocabulary Physical Skill for Interactive Agents, Cui et al. CVPR 2024.
  • PULSE. Universal Humanoid Motion Representations for Physics-Based Control, Luo et al. ICLR 2024.
  • H-GAP. H-GAP: Humanoid Control with a Generalist Planner, Jiang et al. ICLR 2024.
  • UniHSI. UniHSI: Unified Human-Scene Interaction via Prompted Chain-of-Contacts, Xiao et al. ICLR 2024.
  • Phys-Fullbody-Grasp. Physically Plausible Full-Body Hand-Object Interaction Synthesis, Braun et al. 3DV 2024.
  • Expressive Humanoid. Expressive Whole-Body Control for Humanoid Robots, Cheng et al. RSS 2024.
  • PhySHOI. PhysHOI: Physics-Based Imitation of Dynamic Human-Object Interaction, Wang et al. ArXiv 2024.
  • CASE. C·ASE: Learning Conditional Adversarial Skill Embeddings for Physics-based Characters, Dou et al. SIGGRAPH Asia 2023.
  • AdaptNet. AdaptNet: Policy Adaptation for Physics-Based Character Control, Xu et al. SIGGRAPH Asia 2023.
  • NCP. Neural Categorical Priors for Physics-Based Character Control, Zhu et al. SIGGRAPH Asia 2023.
  • DROP. DROP: Dynamics Responses from Human Motion Prior and Projective Dynamics, Jiang et al. SIGGRAPH Asia 2023.
  • InsActor. InsActor: Instruction-driven Physics-based Characters, Ren et al. NeurIPS 2023.
  • Humanoid4Parkour. Humanoid Parkour Learning, Zhuang et al. CoRL 2023.
  • PHC. Perpetual Humanoid Control for Real-time Simulated Avatars, Luo et al. ICCV 2023.
  • DiffMimic. DiffMimic: Efficient Motion Mimicking with Differentiable Physics, Ren et al. ICLR 2023.
  • Vid2Player3D. DiffMimic: Efficient Motion Mimicking with Differentiable Physics, Zhang et al. SIGGRAPH 2023.
  • QuestEnvSim. QuestEnvSim: Environment-Aware Simulated Motion Tracking from Sparse Sensors, Lee et al. SIGGRAPH 2023.
  • Hassan et al. Synthesizing Physical Character-Scene Interactions, Hassan et al. SIGGRAPH 2023.
  • CALM. CALM: Conditional Adversarial Latent Models for Directable Virtual Characters, Tessler et al.
  • Composite Motion. Composite Motion Learning with Task Control, Xu et al. SIGGRAPH 2023.
  • Trace and Pace. Trace and Pace: Controllable Pedestrian Animation via Guided Trajectory Diffusion, Rempe et al. CVPR 2023.
  • EmbodiedPose. Embodied Scene-aware Human Pose Estimation, Luo et al. NeurIPS 2022.
  • MoCapAct. MoCapAct: A Multi-Task Dataset for Simulated Humanoid Control, Wagener et al. NeurIPS 2022.
  • Gopinath et al.. Motion In-betweening for Physically Simulated Characters, Gopinath et al. SIGGRAPH Asia 2022.
  • AIP. AIP: Adversarial Interaction Priors for Multi-Agent Physics-based Character Control, Younes et al. SIGGRAPH Asia 2022.
  • ControlVAE. ControlVAE: Model-Based Learning of Generative Controllers for Physics-Based Characters, Yao et al. SIGGRAPH Asia 2022.
  • QuestSim. QuestSim: Human Motion Tracking from Sparse Sensors with Simulated Avatars, Winkler et al. SIGGRAPH Asia 2022.
  • PADL. PADL: Language-Directed Physics-Based Character, Juravsky et al. SIGGRAPH Asia 2022.
  • Wang et al. Differentiable Simulation of Inertial Musculotendons, Wang et al. SIGGRAPH Asia 2022.
  • ASE. ASE: Large-Scale Reusable Adversarial Skill Embeddings for Physically Simulated Characters, Peng et al.
  • Learn to Move. Deep Reinforcement Learning for Modeling Human Locomotion Control in Neuromechanical Simulation, Peng et al. Journal of Neuro-Engineering and Rehabilitation 2021
  • KinPoly. Dynamics-Regulated Kinematic Policy for Egocentric Pose Estimation, Luo et al. NeurIPS 2021.
  • AMP. AMP: Adversarial Motion Priors for Stylized Physics-Based Character Control, SIGGRAPH 2021.
  • SimPoE. SimPoE: Simulated Character Control for 3D Human Pose Estimation, Yuan et al. CVPR 2021.
  • RFC. Residual Force Control for Agile Human Behavior Imitation and Extended Motion Synthesis, Yuan et al. NeurIPS 2020.
  • Yuan et al.. Diverse Trajectory Forecasting with Determinantal Point Processes, Yuan et al. ICLR 2020.
  • Ego-Pose. Ego-Pose Estimation and Forecasting as Real-Time PD Control, Yuan et al. ICCV 2019.
  • DeepMimic. DeepMimic: Example-Guided Deep Reinforcement Learning of Physics-Based Character Skills, SIGGRAPH 2018.

Bio-stuff: Human Anatomy, Biomechanics, Physiology

  • OpenCapBench. A Benchmark to Bridge Pose Estimation and Biomechanics, Gozlan et al. WACV 2025.
  • Shahriari et al. Enhancing Robustness in Manipulability Assessment: The Pseudo-Ellipsoid Approach, Shahriari et al. IROS 2024.
  • Lai et al. Mapping Grip Force to Muscular Activity Towards Understanding Upper Limb Musculoskeletal Intent using a Novel Grip Strength Model, Lai et al. BiorXiv 2024.
  • HDyS. Homogeneous Dynamics Space for Heterogeneous Humans, Liu et al. ArXiv 2024.
  • BioDesign. Motion-Driven Neural Optimizer for Prophylactic Braces Made by Distributed Microstructures, Han et al. SIGGRAPH Asia 2024.
  • Evans et al. Synchronized Video, Motion Capture and Force Plate Dataset for Validating Markerless Human Movement Analysis, Evans et al. Scientific Data 2024.
  • Muscles in Time. Muscles in Time: Learning to Understand Human Motion by Simulating Muscle Activations, Schneider et al. NeurIPS D&B 2024.
  • Wei et al. Safe Bayesian Optimization for the Control of High-Dimensional Embodied Systems, Wei et al. CoRL 2024.
  • ImDy. ImDy: Human Inverse Dynamics from Imitated Observations, Liu et al. ArXiv 2024.
  • Macwan et al. High-Fidelity Worker Motion Simulation With Generative AI, Macwan et al. HFES 2024.
  • AddBiomechanics. AddBiomechanics Dataset: Capturing the Physics of Human Motion at Scale, Werling et al. ECCV 2024.
  • MANIKIN. MANIKIN: Biomechanically Accurate Neural Inverse Kinematics for Human Motion Estimation, Jiang et al. ECCV 2024.
  • Human3.6M+. Using musculoskeletal models to generate physically-consistent data for 3D human pose, kinematic, dynamic, and muscle estimation, Nasr et al. Multibody System Dynamics 2024.
  • HIT. HIT: Estimating Internal Human Implicit Tissues from the Body Surface, Keller et al. CVPR 2024.
  • Dai et al. Full-body pose reconstruction and correction in virtual reality for rehabilitation training, Dai et al. Frontiers in Neuroscience 2024.
  • DynSyn. DynSyn: Dynamical Synergistic Representation for Efficient Learning and Control in Overactuated Embodied Systems, He et al. ICML 2024.
  • He et al.. Self Model for Embodied Intelligence: Modeling Full-Body Human Musculoskeletal System and Locomotion Control with Hierarchical Low-Dimensional Representation, He et al. ICRA 2024.
  • SKEL. From skin to skeleton: Towards biomechanically accurate 3d digital humans, Keller et al. SIGGRAPH Asia 2023.
  • MuscleVAE. MuscleVAE: Model-Based Controllers of Muscle-Actuated Characters, Feng et al. SIGGRAPH Asia 2023.
  • Bidirectional GaitNet Bidirectional GaitNet, Park et al. SIGGRAPH 2023.
  • Lee et al.. Anatomically Detailed Simulation of Human Torso, Lee et al. SIGGRAPH 2023.
  • MiA. Muscles in Action, Chiquer et al. ICCV 2023.
  • OSSO. OSSO: Obtaining Skeletal Shape from Outside, Keller et al. CVPR 2022.
  • Xing et al. Functional movement screen dataset collected with two Azure Kinect depth sensors, Xing et al. Scientific Data 2022.
  • LRLE. Synthesis of biologically realistic human motion using joint torque actuation, Jiang et al. SIGGRAPH 2019.
  • HuGaDb. HuGaDB: Human Gait Database for Activity Recognition from Wearable Inertial Sensor Networks, Chereshnev et al. AIST 2017.