An aggregation of human motion understanding research.
Motion Generation Motion Editing Motion Stylization
Human-Object Interaction Human-Scene Interaction Human-Human Interaction
- Zhao et al. Motion Generation Review: Exploring Deep Learning for Lifelike Animation with Manifold, Zhao et al. ArXiv 2024.
- Zhu et al. Human Motion Generation: A Survey, Zhu et al. T-PAMI 2023.
- LS-GAN. LS-GAN: HumanMotion Synthesis with Latent-space GANs, Amballa et al. ArXiv 2025.
- SemTalk. SemTalk: Holistic Co-speech Motion Generation with Frame-level Semantic Emphasis, Zhang et al. ArXiv 2024.
- InterDance. InterDance: Reactive 3D Dance Generation with Realistic Duet Interactions, Li et al. ArXiv 2024.
- Unimotion. Unimotion: Unifying 3D Human Motion Synthesis and Understanding, Li et al. 3DV 2025.
- MotionCraft. MotionCraft: Crafting Whole-Body Motion with Plug-and-Play Multimodal Controls, Bian et al. AAAI 2025.
- ReinDiffuse. ReinDiffuse: Crafting Physically Plausible Motions with Reinforced Diffusion Model, Han et al. WACV 2025.
- ScaMo. ScaMo: Exploring the Scaling Law in Autoregressive Motion Generation Model, Lu et al. ArXiv 2024.
- EnergyMoGen. EnergyMoGen: Compositional Human Motion Generation with Energy-Based Diffusion Model in Latent Space, Zhang et al. ArXiv 2024.
- Move in 2D. Move-in-2D: 2D-Conditioned Human Motion Generation, Huang et al. ArXiv 2024.
- Motion-2-to-3. Motion-2-to-3: Leveraging 2D Motion Data to Boost 3D Motion Generation, Pi et al. ArXiv 2024.
- Light-T2M. Light-T2M: A Lightweight and Fast Model for Text-to-motion Generation, Zeng et al. AAAI 2025.
- Languate of Motion. The Language of Motion: Unifying Verbal and Non-verbal Language of 3D Human Motion, Chen et al. ArXiv 2024.
- Mogo. Mogo: RQ Hierarchical Causal Transformer for High-Quality 3D Human Motion Generation, Fu et al. ArXiv 2024.
- CoMA. CoMA: Compositional Human Motion Generation with Multi-modal Agents, Sun et al. ArXiv 2024.
- SoPo. SoPo: Text-to-Motion Generation Using Semi-Online Preference Optimization, Tan et al. ArXiv 2024.
- RMD. RMD: A Simple Baseline for More General Human Motion Generation via Training-free Retrieval-Augmented Motion Diffuse, Liao et al. ArXiv 2024.
- BiPO. BiPO: Bidirectional Partial Occlusion Network for Text-to-Motion Synthesis, Hong et al. ArXiv 2024.
- AtoM. AToM: Aligning Text-to-Motion Model at Event-Level with GPT-4Vision Reward, Han et al. ArXiv 2024.
- MVLift. Lifting Motion to the 3D World via 2D Diffusion, Li et al. ArXiv 2024.
- DisCoRD. DisCoRD: Discrete Tokens to Continuous Motion via Rectified Flow Decoding, Cho et al. ArXiv 2024.
- MoTe. MoTe: Learning Motion-Text Diffusion Model for Multiple Generation Tasks, Wue et al. ArXiv 2024.
- InfiniDreamer. InfiniDreamer: Arbitrarily Long Human Motion Generation via Segment Score Distillation, Zhuo et al. ArXiv 2024.
- FTMoMamba. FTMoMamba: Motion Generation with Frequency and Text State Space Models, Li et al. ArXiv 2024.
- MARDM. Rethinking Diffusion for Text-Driven Human Motion Generation, Meng et al. ArXiv 2024.
- KinMo. KinMo: Kinematic-aware Human Motion Understanding and Generation, Zhang et al. ArXiv 2024.
- LLaMo. Human Motion Instruction Tuning, Li et al. ArXiv 2024.
- Morph. Morph: A Motion-free Physics Optimization Framework for Human Motion Generation, Li et al. ArXiv 2024.
- KMM. KMM: Key Frame Mask Mamba for Extended Motion Generation, Zhang et al. ArXiv 2024.
- MotionGPT-2. MotionGPT-2: A General-Purpose Motion-Language Model for Motion Generation and Understanding, Wang et al. ArXiv 2024.
- Lodge++. Lodge++: High-quality and Long Dance Generation with Vivid Choreography Patterns, Li et al. ArXiv 2024.
- MotionCLR. MotionCLR: Motion Generation and Training-free Editing via Understanding Attention Mechanisms, Chen et al. ArXiv 2024.
- MotionGlot. MotionGlot: A Multi-Embodied Motion Generation Model, Harithas et al. ArXiv 2024.
- LEAD. LEAD: Latent Realignment for Human Motion Diffusion, Andreou et al. ArXiv 2024.
- Leite et al. Enhancing Motion Variation in Text-to-Motion Models via Pose and Video Conditioned Editing, Leite et al. ArXiv 2024.
- MotionRL. MotionRL: Align Text-to-Motion Generation to Human Preferences with Multi-Reward Reinforcement Learning, Liu et al. ArXiv 2024.
- UniMuMo. UniMuMo: Unified Text, Music and Motion Generation, Yang et al. ArXiv 2024.
- MotionLLM. MotionLLM: Understanding Human Behaviors from Human Motions and Videos, Chen et al. ArXiv 2024.
- DART. DART: A Diffusion-Based Autoregressive Motion Model for Real-Time Text-Driven Motion Control, Zhao et al. ArXiv 2024.
- CLoSD. CLoSD: Closing the Loop between Simulation and Diffusion for multi-task character control, Tevet et al. ArXiv 2024.
- Wang et al. Quo Vadis, Motion Generation? From Large Language Models to Large Motion Models, Wang et al. ArXiv 2024.
- T2M-X. T2M-X: Learning Expressive Text-to-Motion Generation from Partially Annotated Data, Liu et al. ArXiv 2024.
- MoRAG. MoRAG -- Multi-Fusion Retrieval Augmented Generation for Human Motion, Shashank et al. ArXiv 2024.
- Mandelli et al. Generation of Complex 3D Human Motion by Temporal and Spatial Composition of Diffusion Models, Mandelli et al. ArXiv 2024.
- BAD. BAD: Bidirectional Auto-regressive Diffusion for Text-to-Motion Generation, Hosseyni et al. ArXiv 2024.
- synNsync. Synergy and Synchrony in Couple Dances, Manukele et al. ArXiv 2024.
- Dong et al. Word-Conditioned 3D American Sign Language Motion Generation, Dong et al. EMNLP 2024.
- Text to blind motion. Text to Blind Motion, Kim et al. NeurIPS D&B 2024.
- UniMTS. UniMTS: Unified Pre-training for Motion Time Series, Zhang et al. NeurIPS 2024.
- Christopher et al.. Constrained Synthesis with Projected Diffusion Models, Christopher et al. NeurIPS 2024.
- MoMu-Diffusion. MoMu-Diffusion: On Learning Long-Term Motion-Music Synchronization and Correspondence, You et al. NeurIPS 2024.
- MoGenTS. MoGenTS: Motion Generation based on Spatial-Temporal Joint Modeling, Yuan et al. NeurIPS 2024.
- M3GPT. M3GPT: An Advanced Multimodal, Multitask Framework for Motion Comprehension and Generation, Luo et al. NeurIPS 2024.
- Bikov et al. Fitness Aware Human Motion Generation with Fine-Tuning, Bikov et al. NeurIPS Workshop 2024.
- FG-MDM. FG-MDM: Towards Zero-Shot Human Motion Generation via ChatGPT-Refined Descriptions, ICPR 2024.
- SynTalker. Enabling Synergistic Full-Body Control in Prompt-Based Co-Speech Motion Generation, Chen et al. ACM MM 2024.
- L3EM. Towards Emotion-enriched Text-to-Motion Generation via LLM-guided Limb-level Emotion Manipulating. Yu et al. ACM MM 2024.
- StableMoFusion. StableMoFusion: Towards Robust and Efficient Diffusion-based Motion Generation Framework, Huang et al. ACM MM 2024.
- SATO. SATO: Stable Text-to-Motion Framework, Chen et al. ACM MM 2024.
- PIDM. PIDM: Personality-Aware Interaction Diffusion Model for Gesture Generation, Shibasaki et al. ICANN 2024.
- Macwan et al. High-Fidelity Worker Motion Simulation With Generative AI, Macwan et al. HFES 2024.
- Jin et al.. Local Action-Guided Motion Diffusion Model for Text-to-Motion Generation, Jin et al. ECCV 2024.
- Motion Mamba. Motion Mamba: Efficient and Long Sequence Motion Generation, Zhong et al. ECCV 2024.
- EMDM. EMDM: Efficient Motion Diffusion Model for Fast, High-Quality Human Motion Generation, Zhou et al. ECCV 2024.
- CoMo. CoMo: Controllable Motion Generation through Language Guided Pose Code Editing, Huang et al. ECCV 2024.
- CoMusion. CoMusion: Towards Consistent Stochastic Human Motion Prediction via Motion Diffusion, Sun et al. ECCV 2024.
- Shan et al.. Towards Open Domain Text-Driven Synthesis of Multi-Person Motions, Shan et al. ECCV 2024.
- ParCo. ParCo: Part-Coordinating Text-to-Motion Synthesis, Zou et al. ECCV 2024.
- Sampieri et al.. Length-Aware Motion Synthesis via Latent Diffusion, Sampieri et al. ECCV 2024.
- ChroAccRet. Chronologically Accurate Retrieval for Temporal Grounding of Motion-Language Models, Fujiwara et al. ECCV 2024.
- MHC. Generating Physically Realistic and Directable Human Motions from Multi-Modal Inputs, Liu et al. ECCV 2024.
- ProMotion. Plan, Posture and Go: Towards Open-vocabulary Text-to-Motion Generation, Liu et al. ECCV 2024.
- FreeMotion. FreeMotion: MoCap-Free Human Motion Synthesis with Multimodal Large Language Models, Zhang et al. ECCV 2024.
- Text Motion Translator. Text Motion Translator: A Bi-Directional Model for Enhanced 3D Human Motion Generation from Open-Vocabulary Descriptions, Qian et al. ECCV 2024.
- FreeMotion. FreeMotion: A Unified Framework for Number-free Text-to-Motion Synthesis, Fan et al. ECCV 2024.
- Kinematic Phrases. Bridging the Gap between Human Motion and Action Semantics via Kinematic Phrases, Liu et al. ECCV 2024.
- MotionChain. MotionChain: Conversational Motion Controllers via Multimodal Prompts, Jiang et al. ECCV 2024.
- SMooDi. SMooDi: Stylized Motion Diffusion Model, Zhong et al. ECCV 2024.
- BAMM. BAMM: Bidirectional Autoregressive Motion Model, Pinyoanuntapong et al. ECCV 2024.
- MotionLCM. MotionLCM: Real-time Controllable Motion Generation via Latent Consistency Model, Dai et al. ECCV 2024.
- Ren et al.. Realistic Human Motion Generation with Cross-Diffusion Models, Ren et al. ECCV 2024.
- M2D2M. M2D2M: Multi-Motion Generation from Text with Discrete Diffusion Models, Chi et al. ECCV 2024.
- Large Motion Model. Large Motion Model for Unified Multi-Modal Motion Generation, Zhang et al. ECCV 2024.
- TesMo. Generating Human Interaction Motions in Scenes with Text Control, Yi et al. ECCV 2024.
- TLcontrol. TLcontrol: Trajectory and Language Control for Human Motion Synthesis, Wan et al. ECCV 2024.
- ExpGest. ExpGest: Expressive Speaker Generation Using Diffusion Model and Hybrid Audio-Text Guidance, Cheng et al. ICME 2024.
- Chen et al. Anatomically-Informed Vector Quantization Variational Auto-Encoder for Text-to-Motion Generation, Chen et al. ICME Workshop 2024.
- HumanTOMATO. HumanTOMATO: Text-aligned Whole-body Motion Generation, Lu et al. ICML 2024.
- GPHLVM. Bringing Motion Taxonomies to Continuous Domains via GPLVM on Hyperbolic Manifolds, Jaquier et al. ICML 2024.
- CondMDI. Flexible Motion In-betweening with Diffusion Models, Cohan et al. SIGGRAPH 2024.
- CAMDM. Taming Diffusion Probabilistic Models for Character Control, Chen et al. SIGGRAPH 2024.
- LGTM. LGTM: Local-to-Global Text-Driven Human Motion Diffusion Models, Sun et al. SIGGRAPH 2024.
- TEDi. TEDi: Temporally-Entangled Diffusion for Long-Term Motion Synthesis, Zhang et al. SIGGRAPH 2024.
- A-MDM. Interactive Character Control with Auto-Regressive Motion Diffusion Models, Shi et al. SIGGRAPH 2024.
- Starke et al.. Categorical Codebook Matching for Embodied Character Controllers, Starke et al. SIGGRAPH 2024.
- SuperPADL. SuperPADL: Scaling Language-Directed Physics-Based Control with Progressive Supervised Distillation, Juravsky et al. SIGGRAPH 2024.
- ProgMoGen. Programmable Motion Generation for Open-set Motion Control Tasks, Liu et al. CVPR 2024.
- PACER+. PACER+: On-Demand Pedestrian Animation Controller in Driving Scenarios, Wang et al. CVPR 2024.
- AMUSE. Emotional Speech-driven 3D Body Animation via Disentangled Latent Diffusion, Chhatre et al. CVPR 2024.
- Liu et al.. Towards Variable and Coordinated Holistic Co-Speech Motion Generation, Liu et al. CVPR 2024.
- MAS. MAS: Multi-view Ancestral Sampling for 3D motion generation using 2D diffusion, Kapon et al. CVPR 2024.
- WANDR. WANDR: Intention-guided Human Motion Generation, Diomataris et al. CVPR 2024.
- MoMask. MoMask: Generative Masked Modeling of 3D Human Motions, Guo et al. CVPR 2024.
- ChapPose. ChatPose: Chatting about 3D Human Pose, Feng et al. CVPR 2024.
- AvatarGPT. AvatarGPT: All-in-One Framework for Motion Understanding, Planning, Generation and Beyond, Zhou et al. CVPR 2024.
- MMM. MMM: Generative Masked Motion Model, Pinyoanuntapong et al. CVPR 2024.
- AAMDM. AAMDM: Accelerated Auto-regressive Motion Diffusion Model, Li et al. CVPR 2024.
- OMG. OMG: Towards Open-vocabulary Motion Generation via Mixture of Controllers, Liang et al. CVPR 2024.
- FlowMDM. FlowMDM: Seamless Human Motion Composition with Blended Positional Encodings, Barquero et al. CVPR 2024.
- Digital Life Project. Digital Life Project: Autonomous 3D Characters with Social Intelligence, Cai et al. CVPR 2024.
- EMAGE. EMAGE: Towards Unified Holistic Co-Speech Gesture Generation via Expressive Masked Audio Gesture Modeling, Liu et al. CVPR 2024.
- STMC. Multi-Track Timeline Control for Text-Driven 3D Human Motion Generation, Petrovich et al. CVPR Workshop 2024.
- InstructMotion. Exploring Text-to-Motion Generation with Human Preference, Sheng et al. CVPR Workshop 2024.
- Single Motion Diffusion. Raab et al. ICLR 2024.
- NeRM. NeRM: Learning Neural Representations for High-Framerate Human Motion Synthesis, Wei et al. ICLR 2024.
- PriorMDM. PriorMDM: Human Motion Diffusion as a Generative Prior, Shafir et al. ICLR 2024.
- OmniControl. OmniControl: Control Any Joint at Any Time for Human Motion Generation, Xie et al. ICLR 2024.
- Adiya et al.. Bidirectional Temporal Diffusion Model for Temporally Consistent Human Animation, Adiya et al. ICLR 2024.
- Duolando. Duolando: Follower GPT with Off-Policy Reinforcement Learning for Dance Accompaniment, Li et al. ICLR 2024.
- HuTuDiffusion. HuTuMotion: Human-Tuned Navigation of Latent Motion Diffusion Models with Minimal Feedback, Han et al. AAAI 2024.
- AMD. AMD: Anatomical Motion Diffusion with Interpretable Motion Decomposition and Fusion, Jing et al. AAAI 2024.
- MotionMix. MotionMix: Weakly-Supervised Diffusion for Controllable Motion Generation, Hoang et al. AAAI 2024.
- B2A-HDM. Towards Detailed Text-to-Motion Synthesis via Basic-to-Advanced Hierarchical Diffusion Model, Xie et al. AAAI 2024.
- Everything2Motion. Everything2Motion: Synchronizing Diverse Inputs via a Unified Framework for Human Motion Synthesis, Fan et al. AAAI 2024.
- MotionGPT. MotionGPT: Finetuned LLMs are General-Purpose Motion Generators, Zhang et al. AAAI 2024.
- Dong et al. Enhanced Fine-grained Motion Diffusion for Text-driven Human Motion Synthesis, Dong et al. AAAI 2024.
- UNIMASKM. A Unified Masked Autoencoder with Patchified Skeletons for Motion Synthesis, Mascaro et al. AAAI 2024.
- B2A-HDM. Towards Detailed Text-to-Motion Synthesis via Basic-to-Advanced Hierarchical Diffusion Model, Xie et al. AAAI 2024.
- GUESS. GUESS: GradUally Enriching SyntheSis for Text-Driven Human Motion Generation, Gao et al. TPAMI 2024.
- Xie et al.. Sign Language Production with Latent Motion Transformer, Xie et al. WACV 2024.
- GraphMotion. Act As You Wish: Fine-grained Control of Motion Diffusion Model with Hierarchical Semantic Graphs, Jin et al. NeurIPS 2023.
- MotionGPT. MotionGPT: Human Motion as Foreign Language, Jiang et al. NeurIPS 2023.
- FineMoGen. FineMoGen: Fine-Grained Spatio-Temporal Motion Generation and Editing, Zhang et al. NeurIPS 2023.
- InsActor. InsActor: Instruction-driven Physics-based Characters, Ren et al. NeurIPS 2023.
- AttT2M. AttT2M: Text-Driven Human Motion Generation with Multi-Perspective Attention Mechanism, Zhong et al. ICCV 2023.
- TMR. TMR: Text-to-Motion Retrieval Using Contrastive 3D Human Motion Synthesis, Petrovich et al. ICCV 2023.
- MAA. Make-An-Animation: Large-Scale Text-conditional 3D Human Motion Generation, Azadi et al. ICCV 2023.
- PhysDiff. PhysDiff: Physics-Guided Human Motion Diffusion Model, Yuan et al. ICCV 2023.
- ReMoDiffusion. ReMoDiffuse: Retrieval-Augmented Motion Diffusion Model, Zhang et al. ICCV 2023.
- BelFusion. BeLFusion: Latent Diffusion for Behavior-Driven Human Motion Prediction, Barquero et al. ICCV 2023.
- GMD. GMD: Guided Motion Diffusion for Controllable Human Motion Synthesis, Karunratanakul et al. ICCV 2023.
- HMD-NeMo. HMD-NeMo: Online 3D Avatar Motion Generation From Sparse Observations, Aliakbarian et al. ICCV 2023.
- SINC. SINC: Spatial Composition of 3D Human Motions for Simultaneous Action Generation, Athanasiou et al. ICCV 2023.
- Kong et al.. Priority-Centric Human Motion Generation in Discrete Latent Space, Kong et al. ICCV 2023.
- FgT2M. Fg-T2M: Fine-Grained Text-Driven Human Motion Generation via Diffusion Model, Wang et al. ICCV 2023.
- EMS. Breaking The Limits of Text-conditioned 3D Motion Synthesis with Elaborative Descriptions, Qian et al. ICCV 2023.
- GenMM. Example-based Motion Synthesis via Generative Motion Matching, Li et al. SIGGRAPH 2023.
- GestureDiffuCLIP. GestureDiffuCLIP: Gesture Diffusion Model with CLIP Latents, Ao et al. SIGGRAPH 2023.
- BodyFormer. BodyFormer: Semantics-guided 3D Body Gesture Synthesis with Transformer, Pang et al. SIGGRAPH 2023.
- Alexanderson et al.. Listen, denoise, action! Audio-driven motion synthesis with diffusion models, Alexanderson et al. SIGGRAPH 2023.
- AGroL. Avatars Grow Legs: Generating Smooth Human Motion from Sparse Tracking Inputs with Diffusion Model, Du et al. CVPR 2023.
- TALKSHOW. Generating Holistic 3D Human Motion from Speech, Yi et al. CVPR 2023.
- T2M-GPT. T2M-GPT: Generating Human Motion from Textual Descriptions with Discrete Representations, Zhang et al. CVPR 2023.
- UDE. UDE: A Unified Driving Engine for Human Motion Generation, Zhou et al. CVPR 2023.
- OOHMG. Being Comes from Not-being: Open-vocabulary Text-to-Motion Generation with Wordless Training, Lin et al. CVPR 2023.
- EDGE. EDGE: Editable Dance Generation From Music, Tseng et al. CVPR 2023.
- MLD. Executing your Commands via Motion Diffusion in Latent Space, Chen et al. CVPR 2023.
- MoDi. MoDi: Unconditional Motion Synthesis from Diverse Data, Raab et al. CVPR 2023.
- MoFusion. MoFusion: A Framework for Denoising-Diffusion-based Motion Synthesis, Dabral et al. CVPR 2023.
- Mo et al.. Continuous Intermediate Token Learning with Implicit Motion Manifold for Keyframe Based Motion Interpolation, Mo et al. CVPR 2023.
- HMDM. MDM: Human Motion Diffusion Model, Tevet et al. ICLR 2023.
- MotionDiffuse. MotionDiffuse: Text-Driven Human Motion Generation with Diffusion Model, Zhang et al. TPAMI 2023.
- Bailando++. Bailando++: 3D Dance GPT with Choreographic Memory, Li et al. TPAMI 2023.
- UDE-2. A Unified Framework for Multimodal, Multi-Part Human Motion Synthesis, Zhou et al. ArXiv 2023.
- Motion Script. MotionScript: Natural Language Descriptions for Expressive 3D Human Motions, Yazdian et al. ArXiv 2023.
- NeMF. NeMF: Neural Motion Fields for Kinematic Animation, He et al. NeurIPS 2022.
- PADL. PADL: Language-Directed Physics-Based Character, Juravsky et al. SIGGRAPH Asia 2022.
- Rhythmic Gesticulator. Rhythmic Gesticulator: Rhythm-Aware Co-Speech Gesture Synthesis with Hierarchical Neural Embeddings, Ao et al. SIGGRAPH Asia 2022.
- TEACH. TEACH: Temporal Action Composition for 3D Human, Athanasiou et al. 3DV 2022.
- Implicit Motion. Implicit Neural Representations for Variable Length Human Motion Generation, Cervantes et al. ECCV 2022.
- Zhong et al.. Learning Uncoupled-Modulation CVAE for 3D Action-Conditioned Human Motion Synthesis, Zhong et al. ECCV 2022.
- MotionCLIP. MotionCLIP: Exposing Human Motion Generation to CLIP Space, Tevet et al. ECCV 2022.
- PoseGPT. PoseGPT: Quantizing human motion for large scale generative modeling, Lucas et al. ECCV 2022.
- TEMOS. TEMOS: Generating diverse human motions from textual descriptions, Petrovich et al. ECCV 2022.
- TM2T. TM2T: Stochastic and Tokenized Modeling for the Reciprocal Generation of 3D Human Motions and Texts, Guo et al. ECCV 2022.
- AvatarCLIP. AvatarCLIP: Zero-Shot Text-Driven Generation and Animation of 3D Avatars, Hong et al. SIGGRAPH 2022.
- DeepPhase. Deepphase: Periodic autoencoders for learning motion phase manifolds, Starke et al. SIGGRAPH 2022.
- Guo et al.. Generating Diverse and Natural 3D Human Motions from Text, Guo et al. CVPR 2022
- Bailando. Bailando: 3D Dance Generation by Actor-Critic GPT with Choreographic Memory, Li et al. CVPR 2022.
- ACTOR. Action-Conditioned 3D Human Motion Synthesis with Transformer VAE, Petrovich et al. ICCV 2021.
- AIST++. AI Choreographer: Music Conditioned 3D Dance Generation with AIST++, Li et al. ICCV 2021.
- Starke et al.. Neural animation layering for synthesizing martial arts movements, Starke et al. SIGGRAPH 2021.
- MOJO. We are More than Our Joints: Predicting how 3D Bodies Move, Zhang et al. CVPR 2021.
- DLow. DLow: Diversifying Latent Flows for Diverse Human Motion Prediction, Yuan et al. ECCV 2020.
- Starke et al.. Local motion phases for learning multi-contact character movements, Starke et al. SIGGRAPH 2020.
- MotionFix. MotionFix: Text-Driven 3D Human Motion Editing, Athanasiou et al. SIGGRAPH Asia 2024.
- CigTime. CigTime: Corrective Instruction Generation Through Inverse Motion Editing, Fang et al. NeurIPS 2024.
- Iterative Motion Editing. Iterative Motion Editing with Natural Language, Goel et al. SIGGRAPH 2024.
- DNO. DNO: Optimizing Diffusion Noise Can Serve As Universal Motion Priors, Karunratanakul et al. CVPR 2024.
- MulSMo. MulSMo: Multimodal Stylized Motion Generation by Bidirectional Control Flow, Li et al. ArXiv 2024.
- D-LORD. D-LORD for Motion Stylization, Gupta et al. TSMC 2024.
- HUMOS. HUMOS: Human Motion Model Conditioned on Body Shape, Tripathi et al. ECCV 2024.
- SMEAR. SMEAR: Stylized Motion Exaggeration with ARt-direction, Basset et al. SIGGRAPH 2024.
- MCM-LDM. Arbitrary Motion Style Transfer with Multi-condition Motion Latent Diffusion Model, Song et al. CVPR 2024.
- MoST. MoST: Motion Style Transformer between Diverse Action Contents, Kim et al. CVPR 2024.
- GenMoStyle. Generative Human Motion Stylization in Latent Space, Guo et al. ICLR 2024.
- DiffGrasp. Diffgrasp: Whole-Body Grasping Synthesis Guided by Object Motion Using a Diffusion Model, Zhang et al. AAAI 2025.
- InterTrack. InterTrack: Tracking Human Object Interaction without Object Templates, Xie et al. 3DV 2025.
- FORCE. FORCE: Dataset and Method for Intuitive Physics Guided Human-object Interaction, Zhang et al. 3DV 2025.
- SyncDiff. SyncDiff: Synchronized Motion Diffusion for Multi-Body Human-Object Interaction Synthesis, He et al. ArXiv 2024.
- CHOICE. CHOICE: Coordinated Human-Object Interaction in Cluttered Environments for Pick-and-Place Actions, Lu et al. ArXiv 2024.
- TriDi. TriDi: Trilateral Diffusion of 3D Humans, Objects and Interactions, Petrov et al. ArXiv 2024.
- OOD-HOI. OOD-HOI: Text-Driven 3D Whole-Body Human-Object Interactions Generation Beyond Training Domains, Zhang et al. ArXiv 2024.
- COLLAGE. COLLAGE: Collaborative Human-Agent Interaction Generation using Hierarchical Latent Diffusion and Language Models, Daiya et al. ArXiv 2024.
- SMGDiff. SMGDiff: Soccer Motion Generation using diffusion probabilistic models, Yang et al. ArXiv 2024.
- SkillMimic. SkillMimic: Learning Reusable Basketball Skills from Demonstrations, Wang et al. ArXiv 2024.
- CORE4D. CORE4D: A 4D Human-Object-Human Interaction Dataset for Collaborative Object REarrangement, Zhang et al. ArXiv 2024.
- Wu et al. Human-Object Interaction from Human-Level Instructions, Wu et al. ArXiv 2024.
- HumanVLA. HumanVLA: Towards Vision-Language Directed Object Rearrangement by Physical Humanoid, Xu et al. NeurIPS 2024.
- OmniGrasp. Grasping Diverse Objects with Simulated Humanoids, Luo et al. NeurIPS 2024.
- EgoChoir. EgoChoir: Capturing 3D Human-Object Interaction Regions from Egocentric Views, Yang et al. NeurIPS 2024.
- CooHOI. CooHOI: Learning Cooperative Human-Object Interaction with Manipulated Object Dynamics, Gao et al. NeurIPS 2024.
- InterDreamer. InterDreamer: Zero-Shot Text to 3D Dynamic Human-Object Interaction, Xu et al. NeurIPS 2024.
- PiMForce. Posture-Informed Muscular Force Learning for Robust Hand Pressure Estimation, Seo et al. NeurIPS 2024.
- InterFusion. InterFusion: Text-Driven Generation of 3D Human-Object Interaction, Dai et al. ECCV 2024.
- CHOIS. Controllable Human-Object Interaction Synthesis, Li et al. ECCV 2024.
- F-HOI. F-HOI: Toward Fine-grained Semantic-Aligned 3D Human-Object Interactions, Yang et al. ECCV 2024.
- HIMO. HIMO: A New Benchmark for Full-Body Human Interacting with Multiple Objects, Lv et al. ECCV 2024.
- PhysicsPingPong. Strategy and Skill Learning for Physics-based Table Tennis Animation, Wang et al. SIGGRAPH 2024.
- NIFTY. NIFTY: Neural Object Interaction Fields for Guided Human Motion Synthesis, Kulkarni et al. CVPR 2024.
- HOI Animator. HOIAnimator: Generating Text-prompt Human-object Animations using Novel Perceptive Diffusion Models, Son et al. CVPR 2024.
- CG-HOI. CG-HOI: Contact-Guided 3D Human-Object Interaction Generation, Diller et al. CVPR 2024.
- InterCap. InterCap: Joint Markerless 3D Tracking of Humans and Objects in Interaction, Huang et al. IJCV 2024.
- Phys-Fullbody-Grasp. Physically Plausible Full-Body Hand-Object Interaction Synthesis, Braun et al. 3DV 2024.
- GRIP. GRIP: Generating Interaction Poses Using Spatial Cues and Latent Consistency, Taheri et al. 3DV 2024.
- FAVOR. Favor: Full-Body AR-driven Virtual Object Rearrangement Guided by Instruction Text, Li et al. AAAI 2024.
- OMOMO. Object Motion Guided Human Motion Synthesis, Li et al. SIGGRAPH Asia 2023.
- CHAIRS. Full-Body Articulated Human-Object Interaction, Jiang et al. ICCV 2023.
- HGHOI. Hierarchical Generation of Human-Object Interactions with Diffusion Probabilistic Models, Pi et al. ICCV 2023.
- InterDiff. InterDiff: Generating 3D Human-Object Interactions with Physics-Informed Diffusion, Xu et al. ICCV 2023.
- Object Pop Up. Object pop-up: Can we infer 3D objects and their poses from human interactions alone? Petrov et al. CVPR 2023.
- ARCTIC. A Dataset for Dexterous Bimanual Hand-Object Manipulation, Fan et al. CVPR 2023.
- TOCH. TOCH: Spatio-Temporal Object-to-Hand Correspondence for Motion Refinement, Zhou et al. ECCV 2022.
- COUCH. COUCH: Towards Controllable Human-Chair Interactions, Zhang et al. ECCV 2022.
- SAGA. SAGA: Stochastic Whole-Body Grasping with Contact, Wu et al. ECCV 2022.
- GOAL. GOAL: Generating 4D Whole-Body Motion for Hand-Object Grasping, Taheri et al. CVPR 2022.
- BEHAVE. BEHAVE: Dataset and Method for Tracking Human Object Interactions, Bhatnagar et al. CVPR 2022.
- GRAB. GRAB: A Dataset of Whole-Body Human Grasping of Objects, Taheri et al. ECCV 2020.
- ZeroHSI. ZeroHSI: Zero-Shot 4D Human-Scene Interaction by Video Generation, Li et al. ArXiv 2024.
- Mimicking-Bench. Mimicking-Bench: A Benchmark for Generalizable Humanoid-Scene Interaction Learning via Human Mimicking, Liu et al. ArXiv 2024.
- SCENIC. SCENIC: Scene-aware Semantic Navigation with Instruction-guided Control, Zhang et al. ArXiv 2024.
- Sitcom-Crafter. Sitcom-Crafter: A Plot-Driven Human Motion Generation System in 3D Scenes, Chen et al. ArXiv 2024.
- Paschalidis et al. 3D Whole-body Grasp Synthesis with Directional Controllability, Paschalidis et al. ArXiv 2024.
- EnvPoser. EnvPoser: Environment-aware Realistic Human Motion Estimation from Sparse Observations with Uncertainty Modeling. Xia et al. ArXiv 2024.
- Diffusion Implicit Policy, Diffusion Implicit Policy for Unpaired Scene-aware Motion synthesis, Gong et al. ArXiv 2024.
- SIMS. SIMS: Simulating Human-Scene Interactions with Real World Script Planning, Wang et al. ArXiv 2024.
- LaserHuman. LaserHuman: Language-guided Scene-aware Human Motion Generation in Free Environment, Cong et al. ArXiv 2024.
- LINGO. Autonomous Character-Scene Interaction Synthesis from Text Instruction, Jiang et al. SIGGRAPH Asia 2024.
- DiMoP3D. Harmonizing Stochasticity and Determinism: Scene-responsive Diverse Human Motion Prediction, Lou et al. NeurIPS 2024.
- Liu et al.. Revisit Human-Scene Interaction via Space Occupancy, Liu et al. ECCV 2024.
- TesMo. Generating Human Interaction Motions in Scenes with Text Control, Yi et al. ECCV 2024.
- SAST. Massively Multi-Person 3D Human Motion Forecasting with Scene Context, Mueller et al. ECCV 2024 Workshop.
- Kang et al. Learning Climbing Controllers for Physics-Based Characters, Kang et al. Eurographics 2024.
- Afford-Motion. Move as You Say, Interact as You Can: Language-guided Human Motion Generation with Scene Affordance, Wang et al. CVPR 2024.
- GenZI. GenZI: Zero-Shot 3D Human-Scene Interaction Generation, Li et al. CVPR 2024.
- Cen et al.. Generating Human Motion in 3D Scenes from Text Descriptions, Cen et al. CVPR 2024.
- TRUMANS. Scaling Up Dynamic Human-Scene Interaction Modeling, Jiang et al. CVPR 2024.
- UniHSI. UniHSI: Unified Human-Scene Interaction via Prompted Chain-of-Contacts, Xiao et al. ICLR 2024.
- Purposer. Purposer: Putting Human Motion Generation in Context, Ugrinovic et al. 3DV 2024.
- InterScene. Synthesizing Physically Plausible Human Motions in 3D Scenes, Pan et al. 3DV 2024.
- Mir et al. Generating Continual Human Motion in Diverse 3D Scenes, Mir et al. 3DV 2024.
- DIMOS. DIMOS: Synthesizing Diverse Human Motions in 3D Indoor Scenes, Zhao et al. ICCV 2023.
- LAMA. Locomotion-Action-Manipulation: Synthesizing Human-Scene Interactions in Complex 3D Environments, Lee et al. ICCV 2023.
- Narrator. Narrator: Towards Natural Control of Human-Scene Interaction Generation via Relationship Reasoning, Xuan et al. ICCV 2023.
- CIMI4D. CIMI4D: A Large Multimodal Climbing Motion Dataset under Human-scene Interactions, Yan et al. CVPR 2023.
- Scene-Ego. Scene-aware Egocentric 3D Human Pose Estimation, Wang et al. CVPR 2023.
- SLOPER4D. SLOPER4D: A Scene-Aware Dataset for Global 4D Human Pose Estimation in Urban Environments, Dai et al. CVPR 2023.
- CIRCLE. CIRCLE: Capture in Rich Contextual Environments, Araujo et al. CVPR 2023.
- SceneDiffuser. Diffusion-based Generation, Optimization, and Planning in 3D Scenes, Huang et al. CVPR 2023.
- PMP. PMP: Learning to Physically Interact with Environments using Part-wise Motion Priors, Bae et al. SIGGRAPH 2023.
- QuestEnvSim. QuestEnvSim: Environment-Aware Simulated Motion Tracking from Sparse Sensors, Lee et al. SIGGRAPH 2023.
- Hassan et al. Synthesizing Physical Character-Scene Interactions, Hassan et al. SIGGRAPH 2023.
- Mao et al.. Contact-aware Human Motion Forecasting, Mao et al. NeurIPS 2022.
- HUMANISE. HUMANISE: Language-conditioned Human Motion Generation in 3D Scenes, Wang et al. NeurIPS 2022.
- EmbodiedPose. Embodied Scene-aware Human Pose Estimation, Luo et al. NeurIPS 2022.
- GIMO. GIMO: Gaze-Informed Human Motion Prediction in Context, Zheng et al. ECCV 2022.
- COINS. COINS: Compositional Human-Scene Interaction Synthesis with Semantic Control, Zhao et al. ECCV 2022.
- Wang et al.. Towards Diverse and Natural Scene-aware 3D Human Motion Synthesis, Wang et al. CVPR 2022.
- GAMMA. The Wanderings of Odysseus in 3D Scenes, Zhang et al. CVPR 2022.
- SAMP. Stochastic Scene-Aware Motion Prediction, Hassan et al, ICCV 2021.
- LEMO. Learning Motion Priors for 4D Human Body Capture in 3D Scenes, Zhang et al. ICCV 2021.
- PLACE. PLACE: Proximity Learning of Articulation and Contact in 3D Environments, Zhang et al. 3DV 2020.
- Starke et al.. Local motion phases for learning multi-contact character movements, Starke et al. SIGGRAPH 2020.
- PSI. Generating 3D People in Scenes without People, Zhang et al. CVPR 2020.
- NSM. Neural State Machine for Character-Scene Interactions, Starke et al. SIGGRAPH Asia 2019.
- PROX. Resolving 3D Human Pose Ambiguities with 3D Scene Constraints, Hassan et al. ICCV 2019
- Two in One. Two-in-One: Unified Multi-Person Interactive Motion Generation by Latent Diffusion Transformer, Li et al. ArXiv 2024.
- It Takes Two. It Takes Two: Real-time Co-Speech Two-person’s Interaction Generation via Reactive Auto-regressive Diffusion Model, Shi et al. ArXiv 2024.
- InterMask. InterMask: 3D Human Interaction Generation via Collaborative Masked Modelling, Javed et al. ArXiv 2024.
- COLLAGE. COLLAGE: Collaborative Human-Agent Interaction Generation using Hierarchical Latent Diffusion and Language Models, Daiya et al. ArXiv 2024.
- Harmony4D. Harmony4D: A Video Dataset for In-The-Wild Close Human Interactions, Khirodkar et al. NeurIPS 2024.
- InterControl. InterControl: Generate Human Motion Interactions by Controlling Every Joint, Wang et al. NeurIPS 2024.
- PhysReaction. PhysReaction: Physically Plausible Real-Time Humanoid Reaction Synthesis via Forward Dynamics Guided 4D Imitation, Liu et al. ACM MM 2024.
- Shan et al.. Towards Open Domain Text-Driven Synthesis of Multi-Person Motions, Shan et al. ECCV 2024.
- ReMoS. ReMoS: 3D Motion-Conditioned Reaction Synthesis for Two-Person Interactions, Ghosh et al. ECCV 2024.
- Inter-X. Inter-X: Towards Versatile Human-Human Interaction Analysis, Xu et al. CVPR 2024.
- ReGenNet. ReGenNet: Towards Human Action-Reaction Synthesis, Xu et al. CVPR 2024.
- Fang et al.. Capturing Closely Interacted Two-Person Motions with Reaction Priors, Fan et al. CVPR 2024.
- in2IN. in2IN: Leveraging Individual Information to Generate Human INteractions, Ruiz-Ponce et al. CVPR Workshop 2024.
- InterGen. InterGen: Diffusion-based Multi-human Motion Generation under Complex Interactions, Liang et al. IJCV 2024.
- ActFormer. ActFormer: A GAN-based Transformer towards General Action-Conditioned 3D Human Motion Generation, Xu et al. ICCV 2023.
- Tanaka et al.. Role-aware Interaction Generation from Textual Description, Tanaka et al. ICCV 2023.
- Hi4D. Hi4D: 4D Instance Segmentation of Close Human Interaction, Yin et al. CVPR 2023.
- ExPI. Multi-Person Extreme Motion Prediction, Guo et al. CVPR 2022.
- CHI3D. Three-Dimensional Reconstruction of Human Interactions, Fieraru et al. CVPR 2020.
- Mimicking-Bench. Mimicking-Bench: A Benchmark for Generalizable Humanoid-Scene Interaction Learning via Human Mimicking, Liu et al. ArXiv 2024.
- SCENIC. SCENIC: Scene-aware Semantic Navigation with Instruction-guided Control, Zhang et al. ArXiv 2024.
- CMP & CMR. AnimationGPT:An AIGC tool for generating game combat motion assets, Liao et al.
- LINGO. Autonomous Character-Scene Interaction Synthesis from Text Instruction, Jiang et al. SIGGRAPH Asia 2024.
- LaserHuman. LaserHuman: Language-guided Scene-aware Human Motion Generation in Free Environment, Cong et al. ArXiv 2024.
- AtoM. AToM: Aligning Text-to-Motion Model at Event-Level with GPT-4Vision Reward, Han et al. ArXiv 2024.
- Evans et al. Synchronized Video, Motion Capture and Force Plate Dataset for Validating Markerless Human Movement Analysis, Evans et al. Scientific Data 2024.
- MotionCritic. Aligning Human Motion Generation with Human Perceptions, Wang et al. ArXiv 2024.
- EMHI. EMHI: A Multimodal Egocentric Human Motion Dataset with HMD and Body-Worn IMUs, Fan et al. ArXiv 2024.
- Harmony4D. Harmony4D: A Video Dataset for In-The-Wild Close Human Interactions, Khirodkar et al. NeurIPS 2024.
- EgoSim. EgoSim: An Egocentric Multi-view Simulator for Body-worn Cameras during Human Motion, Hollidt et al. NeurIPS D&B 2024.
- synNsync. Synergy and Synchrony in Couple Dances, Manukele et al. ArXiv 2024.
- Muscles in Time. Muscles in Time: Learning to Understand Human Motion by Simulating Muscle Activations, Schneider et al. NeurIPS D&B 2024.
- Text to blind motion. Text to blind motion, Kim et al. NeurIPS D&B 2024.
- MotionBank. MotionBank: A Large-scale Video Motion Benchmark with Disentangled Rule-based Annotations, Xu et al. ArXiv 2024.
- CORE4D. CORE4D: A 4D Human-Object-Human Interaction Dataset for Collaborative Object REarrangement, Zhang et al. ArXiv 2024.
- CLaM. CLaM: An Open-Source Library for Performance Evaluation of Text-driven Human Motion Generation, Chen et al. ACM MM 2024.
- AddBiomechanics. AddBiomechanics Dataset: Capturing the Physics of Human Motion at Scale, Werling et al. ECCV 2024.
- LiveHPS++. LiveHPS++: Robust and Coherent Motion Capture in Dynamic Free Environment, Ren et al. ECCV 2024.
- SignAvatars. SignAvatars: A Large-scale 3D Sign Language Holistic Motion Dataset and Benchmark, Yu et al. ECCV 2024.
- Nymeria. Nymeria: A massive collection of multimodal egocentric daily motion in the wild, Ma et al. ECCV 2024.
- Human3.6M+. Using musculoskeletal models to generate physically-consistent data for 3D human pose, kinematic, dynamic, and muscle estimation, Nasr et al. Multibody System Dynamics 2024.
- Inter-X. Inter-X: Towards Versatile Human-Human Interaction Analysis, Xu et al. CVPR 2024.
- HardMo. HardMo: ALarge-Scale Hardcase Dataset for Motion Capture, Liao et al. CVPR 2024.
- MMVP. MMVP: A Multimodal MoCap Dataset with Vision and Pressure Sensors, Zhang et al. CVPR 2024.
- RELI11D. RELI11D: A Comprehensive Multimodal Human Motion Dataset and Method, Yan et al. CVPR 2024.
- GroundLink. GroundLink: A Dataset Unifying Human Body Movement and Ground Reaction Dynamics, Han et al. SIGGRAPH Asia 2023.
- HOH. HOH: Markerless Multimodal Human-Object-Human Handover Dataset with Large Object Count, Wiederhold et al. NeurIPS D&B 2023.
- Motion-X. Motion-X: A Large-scale 3D Expressive Whole-body Human Motion Dataset, Lin et al. NeurIPS D&B 2023.
- Humans in Kitchens. Humans in Kitchens: A Dataset for Multi-Person Human Motion Forecasting with Scene Context, Tanke et al. NeurIPS D&B 2023.
- CHAIRS. Full-Body Articulated Human-Object Interaction, Jiang et al. ICCV 2023.
- CIMI4D. CIMI4D: A Large Multimodal Climbing Motion Dataset under Human-scene Interactions, Yan et al. CVPR 2023.
- FLAG3D. FLAG3D: A 3D Fitness Activity Dataset with Language Instruction, Tang et al. CVPR 2023.
- Hi4D. Hi4D: 4D Instance Segmentation of Close Human Interaction, Yin et al. CVPR 2023.
- CIRCLE. CIRCLE: Capture in Rich Contextual Environments, Araujo et al. CVPR 2023.
- MoCapAct. MoCapAct: A Multi-Task Dataset for Simulated Humanoid Control, Wagener et al. NeurIPS 2022.
- ForcePose. Learning to Estimate External Forces of Human Motion in Video, Louis et al. ACM MM 2022.
- BEAT. BEAT: A Large-Scale Semantic and Emotional Multi-Modal Dataset for Conversational Gestures Synthesis, Liu et al. ECCV 2022.
- BRACE. BRACE: The Breakdancing Competition Dataset for Dance Motion Synthesis, Moltisanti et al. ECCV 2022.
- EgoBody. Egobody: Human body shape and motion of interacting people from head-mounted devices, Zhang et al. ECCV 2022.
- GIMO. GIMO: Gaze-Informed Human Motion Prediction in Context, Zheng et al. ECCV 2022.
- HuMMan. HuMMan: Multi-Modal 4D Human Dataset for Versatile Sensing and Modeling, Cai et al. ECCV 2022.
- ExPI. Multi-Person Extreme Motion Prediction, Guo et al. CVPR 2022.
- HumanML3D. Generating Diverse and Natural 3D Human Motions from Text, Guo et al. CVPR 2022.
- BEHAVE. BEHAVE: Dataset and Method for Tracking Human Object Interactions, Bhatnagar et al. CVPR 2022.
- AIST++. AI Choreographer: Music Conditioned 3D Dance Generation with AIST++, Li et al. ICCV 2021.
- Fit3D. AIFit: Automatic 3D Human-Interpretable Feedback Models for Fitness Training, Fieraru et al. CVPR 2021.
- BABEL. BABEL: Bodies, Action and Behavior with English Labels, Punnakkal et al. CVPR 2021
- HumanSC3D. Learning complex 3d human self-contact, Fieraru et al. AAAI 2021.
- CHI3D. Three-Dimensional Reconstruction of Human Interactions, Fieraru et al. CVPR 2020.
- PROX. Resolving 3D Human Pose Ambiguities with 3D Scene Constraints, Hassan et al. ICCV 2019
- AMASS. AMASS: Archive of Motion Capture As Surface Shapes, Mahmood et al. ICCV 2019
- UH-1. Learning from Massive Human Videos for Universal Humanoid Pose Control, Mao et al. ArXiv 2024.
- Mimicking-Bench. Mimicking-Bench: A Benchmark for Generalizable Humanoid-Scene Interaction Learning via Human Mimicking, Liu et al. ArXiv 2024.
- Liu et al. Human-Humanoid Robots Cross-Embodiment Behavior-Skill Transfer Using Decomposed Adversarial Learning from Demonstration, Liu et al. RA-L 2024.
- UH-1. Learning from Massive Human Videos for Universal Humanoid Pose Control, Mao et al. ArXiv 2024.
- Exbody2. Exbody2: Advanced Expressive Humanoid Whole-Body Control, Ji et al. ArXiv 2024.
- SIMS. SIMS: Simulating Human-Scene Interactions with Real World Script Planning, Wang et al. ArXiv 2024.
- HOVER. HOVER: Versatile Neural Whole-Body Controller for Humanoid Robots, He et al. ArXiv 2024.
- CLoSD. CLoSD: Closing the Loop between Simulation and Diffusion for multi-task character control, Tevet et al. ArXiv 2024.
- Humanoidlympics. Humanoidlympics: Sports Environments for Physically Simulated Humanoids, Luo et al. ArXiv 2024.
- SkillMimic. SkillMimic: Learning Reusable Basketball Skills from Demonstrations, Wang et al. ArXiv 2024.
- PDP. PDP: Physics-Based Character Animation via Diffusion Policy, Truong et al. SIGGRAPH Asia 2024.
- MaskedMimic. MaskedMimic: Unified Physics-Based Character Control Through Masked Motion Inpainting, Tessler et al, SIGGRAPH Asia 2024.
- HumanVLA. HumanVLA: Towards Vision-Language Directed Object Rearrangement by Physical Humanoid, Xu et al. NeurIPS 2024.
- OmniGrasp. Grasping Diverse Objects with Simulated Humanoids, Luo et al. NeurIPS 2024.
- InterControl. InterControl: Generate Human Motion Interactions by Controlling Every Joint, Wang et al. NeurIPS 2024.
- CooHOI. CooHOI: Learning Cooperative Human-Object Interaction with Manipulated Object Dynamics, Gao et al. NeurIPS 2024.
- Radosavovic et al.. Humanoid Locomotion as Next Token Prediction, Radosavovic et al. NeurIPS 2024.
- HARMON. Harmon: Whole-Body Motion Generation of Humanoid Robots from Language Descriptions, Jiang et al. CoRL 2024.
- OKAMI. OKAMI: Teaching Humanoid Robots Manipulation Skills through Single Video Imitation, Li et al. CoRL 2024.
- HumanPlus. HumanPlus: Humanoid Shadowing and Imitation from Humans, Fu et al. CoRL 2024.
- OmniH2O. OmniH2O: Universal and Dexterous Human-to-Humanoid Whole-Body Teleoperation and Learning, He et al. CoRL 2024.
- Self-Aware. Know your limits! Optimize the behavior of bipedal robots through self-awareness, Mascaro et al. Humanoids 2024.
- PhysReaction. PhysReaction: Physically Plausible Real-Time Humanoid Reaction Synthesis via Forward Dynamics Guided 4D Imitation, Liu et al. ACM MM 2024.
- H2O. Learning Human-to-Humanoid Real-Time Whole-Body Teleoperation, He et al. IROS 2024.
- MHC. Generating Physically Realistic and Directable Human Motions from Multi-Modal Inputs, Shrestha et al. ECCV 2024.
- DIDI. DIDI: Diffusion-Guided Diversity for Offline Behavioral Generation, Liu et al. ICML 2024.
- MoConVQ. MoConVQ: Unified Physics-Based Motion Control via Scalable Discrete Representations, Yao et al. SIGGRAPH 2024.
- PhysicsPingPong. Strategy and Skill Learning for Physics-based Table Tennis Animation, Wang et al. SIGGRAPH 2024.
- SuperPADL. SuperPADL: Scaling Language-Directed Physics-Based Control with Progressive Supervised Distillation, Juravsky et al. SIGGRAPH 2024.
- SimXR. Real-Time Simulated Avatar from Head-Mounted Sensors, Luo et al. CVPR 2024.
- AnySkill. AnySkill: Learning Open-Vocabulary Physical Skill for Interactive Agents, Cui et al. CVPR 2024.
- PULSE. Universal Humanoid Motion Representations for Physics-Based Control, Luo et al. ICLR 2024.
- H-GAP. H-GAP: Humanoid Control with a Generalist Planner, Jiang et al. ICLR 2024.
- UniHSI. UniHSI: Unified Human-Scene Interaction via Prompted Chain-of-Contacts, Xiao et al. ICLR 2024.
- Phys-Fullbody-Grasp. Physically Plausible Full-Body Hand-Object Interaction Synthesis, Braun et al. 3DV 2024.
- Expressive Humanoid. Expressive Whole-Body Control for Humanoid Robots, Cheng et al. RSS 2024.
- PhySHOI. PhysHOI: Physics-Based Imitation of Dynamic Human-Object Interaction, Wang et al. ArXiv 2024.
- CASE. C·ASE: Learning Conditional Adversarial Skill Embeddings for Physics-based Characters, Dou et al. SIGGRAPH Asia 2023.
- AdaptNet. AdaptNet: Policy Adaptation for Physics-Based Character Control, Xu et al. SIGGRAPH Asia 2023.
- NCP. Neural Categorical Priors for Physics-Based Character Control, Zhu et al. SIGGRAPH Asia 2023.
- DROP. DROP: Dynamics Responses from Human Motion Prior and Projective Dynamics, Jiang et al. SIGGRAPH Asia 2023.
- InsActor. InsActor: Instruction-driven Physics-based Characters, Ren et al. NeurIPS 2023.
- Humanoid4Parkour. Humanoid Parkour Learning, Zhuang et al. CoRL 2023.
- PHC. Perpetual Humanoid Control for Real-time Simulated Avatars, Luo et al. ICCV 2023.
- DiffMimic. DiffMimic: Efficient Motion Mimicking with Differentiable Physics, Ren et al. ICLR 2023.
- Vid2Player3D. DiffMimic: Efficient Motion Mimicking with Differentiable Physics, Zhang et al. SIGGRAPH 2023.
- QuestEnvSim. QuestEnvSim: Environment-Aware Simulated Motion Tracking from Sparse Sensors, Lee et al. SIGGRAPH 2023.
- Hassan et al. Synthesizing Physical Character-Scene Interactions, Hassan et al. SIGGRAPH 2023.
- CALM. CALM: Conditional Adversarial Latent Models for Directable Virtual Characters, Tessler et al.
- Composite Motion. Composite Motion Learning with Task Control, Xu et al. SIGGRAPH 2023.
- Trace and Pace. Trace and Pace: Controllable Pedestrian Animation via Guided Trajectory Diffusion, Rempe et al. CVPR 2023.
- EmbodiedPose. Embodied Scene-aware Human Pose Estimation, Luo et al. NeurIPS 2022.
- MoCapAct. MoCapAct: A Multi-Task Dataset for Simulated Humanoid Control, Wagener et al. NeurIPS 2022.
- Gopinath et al.. Motion In-betweening for Physically Simulated Characters, Gopinath et al. SIGGRAPH Asia 2022.
- AIP. AIP: Adversarial Interaction Priors for Multi-Agent Physics-based Character Control, Younes et al. SIGGRAPH Asia 2022.
- ControlVAE. ControlVAE: Model-Based Learning of Generative Controllers for Physics-Based Characters, Yao et al. SIGGRAPH Asia 2022.
- QuestSim. QuestSim: Human Motion Tracking from Sparse Sensors with Simulated Avatars, Winkler et al. SIGGRAPH Asia 2022.
- PADL. PADL: Language-Directed Physics-Based Character, Juravsky et al. SIGGRAPH Asia 2022.
- Wang et al. Differentiable Simulation of Inertial Musculotendons, Wang et al. SIGGRAPH Asia 2022.
- ASE. ASE: Large-Scale Reusable Adversarial Skill Embeddings for Physically Simulated Characters, Peng et al.
- Learn to Move. Deep Reinforcement Learning for Modeling Human Locomotion Control in Neuromechanical Simulation, Peng et al. Journal of Neuro-Engineering and Rehabilitation 2021
- KinPoly. Dynamics-Regulated Kinematic Policy for Egocentric Pose Estimation, Luo et al. NeurIPS 2021.
- AMP. AMP: Adversarial Motion Priors for Stylized Physics-Based Character Control, SIGGRAPH 2021.
- SimPoE. SimPoE: Simulated Character Control for 3D Human Pose Estimation, Yuan et al. CVPR 2021.
- RFC. Residual Force Control for Agile Human Behavior Imitation and Extended Motion Synthesis, Yuan et al. NeurIPS 2020.
- Yuan et al.. Diverse Trajectory Forecasting with Determinantal Point Processes, Yuan et al. ICLR 2020.
- Ego-Pose. Ego-Pose Estimation and Forecasting as Real-Time PD Control, Yuan et al. ICCV 2019.
- DeepMimic. DeepMimic: Example-Guided Deep Reinforcement Learning of Physics-Based Character Skills, SIGGRAPH 2018.
- OpenCapBench. A Benchmark to Bridge Pose Estimation and Biomechanics, Gozlan et al. WACV 2025.
- Shahriari et al. Enhancing Robustness in Manipulability Assessment: The Pseudo-Ellipsoid Approach, Shahriari et al. IROS 2024.
- Lai et al. Mapping Grip Force to Muscular Activity Towards Understanding Upper Limb Musculoskeletal Intent using a Novel Grip Strength Model, Lai et al. BiorXiv 2024.
- HDyS. Homogeneous Dynamics Space for Heterogeneous Humans, Liu et al. ArXiv 2024.
- BioDesign. Motion-Driven Neural Optimizer for Prophylactic Braces Made by Distributed Microstructures, Han et al. SIGGRAPH Asia 2024.
- Evans et al. Synchronized Video, Motion Capture and Force Plate Dataset for Validating Markerless Human Movement Analysis, Evans et al. Scientific Data 2024.
- Muscles in Time. Muscles in Time: Learning to Understand Human Motion by Simulating Muscle Activations, Schneider et al. NeurIPS D&B 2024.
- Wei et al. Safe Bayesian Optimization for the Control of High-Dimensional Embodied Systems, Wei et al. CoRL 2024.
- ImDy. ImDy: Human Inverse Dynamics from Imitated Observations, Liu et al. ArXiv 2024.
- Macwan et al. High-Fidelity Worker Motion Simulation With Generative AI, Macwan et al. HFES 2024.
- AddBiomechanics. AddBiomechanics Dataset: Capturing the Physics of Human Motion at Scale, Werling et al. ECCV 2024.
- MANIKIN. MANIKIN: Biomechanically Accurate Neural Inverse Kinematics for Human Motion Estimation, Jiang et al. ECCV 2024.
- Human3.6M+. Using musculoskeletal models to generate physically-consistent data for 3D human pose, kinematic, dynamic, and muscle estimation, Nasr et al. Multibody System Dynamics 2024.
- HIT. HIT: Estimating Internal Human Implicit Tissues from the Body Surface, Keller et al. CVPR 2024.
- Dai et al. Full-body pose reconstruction and correction in virtual reality for rehabilitation training, Dai et al. Frontiers in Neuroscience 2024.
- DynSyn. DynSyn: Dynamical Synergistic Representation for Efficient Learning and Control in Overactuated Embodied Systems, He et al. ICML 2024.
- He et al.. Self Model for Embodied Intelligence: Modeling Full-Body Human Musculoskeletal System and Locomotion Control with Hierarchical Low-Dimensional Representation, He et al. ICRA 2024.
- SKEL. From skin to skeleton: Towards biomechanically accurate 3d digital humans, Keller et al. SIGGRAPH Asia 2023.
- MuscleVAE. MuscleVAE: Model-Based Controllers of Muscle-Actuated Characters, Feng et al. SIGGRAPH Asia 2023.
- Bidirectional GaitNet Bidirectional GaitNet, Park et al. SIGGRAPH 2023.
- Lee et al.. Anatomically Detailed Simulation of Human Torso, Lee et al. SIGGRAPH 2023.
- MiA. Muscles in Action, Chiquer et al. ICCV 2023.
- OSSO. OSSO: Obtaining Skeletal Shape from Outside, Keller et al. CVPR 2022.
- Xing et al. Functional movement screen dataset collected with two Azure Kinect depth sensors, Xing et al. Scientific Data 2022.
- LRLE. Synthesis of biologically realistic human motion using joint torque actuation, Jiang et al. SIGGRAPH 2019.
- HuGaDb. HuGaDB: Human Gait Database for Activity Recognition from Wearable Inertial Sensor Networks, Chereshnev et al. AIST 2017.