ReadPapers
1.
Introduction
2.
Locomotion 技术洞察
3.
AMP: Adversarial Motion Priors for Stylized Physics-Based Character Control
4.
ASE: Large-Scale Reusable Adversarial Skill Embeddings for Physically Simulated Characters
5.
Feature-Based Locomotion Controllers
6.
DeepMimic: Example-Guided Deep Reinforcement Learning of Physics-Based Character Skills
7.
ControlVAE: Model-Based Learning of Generative Controllers for Physics-Based Characters
8.
Trace and Pace: Controllable Pedestrian Animation via Guided Trajectory Diffusion
9.
UniPhys Unified Planner and Controller with Diffusion for Flexible
10.
Diffuse-CLoC: Guided Diffusion for Physics-based Character Look-ahead
11.
PDP: Physics-Based Character Animation via Diffusion Policy
12.
DiffuseLoco: Real-Time Legged Locomotion Control with Diffusion from Offline Datasets
13.
Perpetual Humanoid Control for Real-time Simulated Avatars
14.
Calm: Conditional Adversarial Latent Models for Directable Virtual Characters
15.
Universal humanoid motion representations for physics-based control
16.
DReCon: data-driven responsive control of physics-based characters
17.
PARC: Physics-based Augmentation with Reinforcement Learning for Character Controllers
18.
TaoAvatar: Real-Time Lifelike Full-Body Talking Avatars for Augmented Reality via 3D Gaussian Splatting
19.
SAM 3: Segment Anything with Concepts
20.
CLOSD: CLOSING THE LOOP BETWEEN SIMULATION AND DIFFUSION FOR MULTI-TASK CHARACTER CONTROL
21.
MotionPersona: Characteristics-aware Locomotion Control
22.
Diffuse-CLoC Guided Diffusion for Physics-based Character Look-ahead
23.
Gait-Conditioned Reinforcement Learning with Multi-Phase Curriculum for Humanoid Locomotion
24.
UniPhys: Unified Planner and Controller with Diffusion for Flexible
25.
Maskedmimic: Unified physics-based character control through masked motion
26.
Regional Time Stepping for SPH
27.
FreeGave: 3D Physics Learning from Dynamic Videos by Gaussian Velocity
28.
Physics-informed neural networks: A deep learning framework for solving forward and inverse problems involving nonlinear partial differential equations
29.
ParticleGS: Particle-Based Dynamics Modeling of 3D Gaussians for Prior-free Motion Extrapolation
30.
Animate3d: Animating any 3d model with multi-view video diffusion
31.
Particle-Grid Neural Dynamics for Learning Deformable Object Models from RGB-D Videos
32.
HAIF-GS: Hierarchical and Induced Flow-Guided Gaussian Splatting for Dynamic Scene
33.
PIG: Physically-based Multi-Material Interaction with 3D Gaussians
34.
EnliveningGS: Active Locomotion of 3DGS
35.
SplineGS: Learning Smooth Trajectories in Gaussian Splatting for Dynamic Scene Reconstruction
36.
PAMD: Plausibility-Aware Motion Diffusion Model for Long Dance Generation
37.
PMG: Progressive Motion Generation via Sparse Anchor Postures Curriculum Learning
38.
LengthAware Motion Synthesis via Latent Diffusion
39.
IKMo: Image-Keyframed Motion Generation with Trajectory-Pose Conditioned Motion Diffusion Model
40.
UniMoGen: Universal Motion Generation
41.
AMD: Anatomical Motion Diffusion with Interpretable Motion Decomposition and Fusion
42.
Flame: Free-form language-based motion synthesis & editing
43.
Human Motion Diffusion as a Generative Prior
44.
Text-driven Human Motion Generation with Motion Masked Diffusion Model
45.
ReMoDiffuse: RetrievalAugmented Motion Diffusion Model
46.
MotionLCM: Real-time Controllable Motion Generation via Latent Consistency Model
47.
ReAlign: Bilingual Text-to-Motion Generation via Step-Aware Reward-Guided Alignment
48.
Absolute Coordinates Make Motion Generation Easy
49.
Seamless Human Motion Composition with Blended Positional Encodings
50.
FineMoGen: Fine-Grained Spatio-Temporal Motion Generation and Editing
51.
Fg-T2M: Fine-Grained Text-Driven Human Motion Generation via Diffusion Model
52.
Make-An-Animation: Large-Scale Text-conditional 3D Human Motion Generation
53.
StableMoFusion: Towards Robust and Efficient Diffusion-based Motion Generation Framework
54.
EMDM: Efficient Motion Diffusion Model for Fast and High-Quality Motion Generation
55.
Motion Mamba: Efficient and Long Sequence Motion Generation
56.
M2D2M: Multi-Motion Generation from Text with Discrete Diffusion Models
57.
T2LM: Long-Term 3D Human Motion Generation from Multiple Sentences
58.
AttT2M:Text-Driven Human Motion Generation with Multi-Perspective Attention Mechanism
59.
BAD: Bidirectional Auto-Regressive Diffusion for Text-to-Motion Generation
60.
MMM: Generative Masked Motion Model
61.
Priority-Centric Human Motion Generation in Discrete Latent Space
62.
AvatarGPT: All-in-One Framework for Motion Understanding, Planning, Generation and Beyond
63.
MotionGPT: Human Motion as a Foreign Language
64.
Action-GPT: Leveraging Large-scale Language Models for Improved and Generalized Action Generation
65.
PoseGPT: Quantization-based 3D Human Motion Generation and Forecasting
66.
Incorporating Physics Principles for Precise Human Motion Prediction
67.
PIMNet: Physics-infused Neural Network for Human Motion Prediction
68.
PhysDiff: Physics-Guided Human Motion Diffusion Model
69.
NRDF: Neural Riemannian Distance Fields for Learning Articulated Pose Priors
70.
Riemannian Motion Generation: A Unified Framework for Human Motion Representation and Generation via Riemannian Flow Matching
71.
GaussiAnimate: Rethinking Gaussian Splatting for Articulated Models via Skeleton-Aware Representation
72.
SIM1: Physics-Aligned Simulator as Zero-Shot Data Scaler in Deformable Worlds
73.
FIT: A Large-Scale Dataset for Fit-Aware Virtual Try-On
74.
Geometric Neural Distance Fields for Learning Human Motion Priors
75.
Character Controllers Using Motion VAEs
76.
Improving Human Motion Plausibility with Body Momentum
77.
MoGlow: Probabilistic and controllable motion synthesis using normalising flows
78.
Modi: Unconditional motion synthesis from diverse data
79.
MotionDiffuse: Text-Driven Human Motion Generation with Diffusion Model
80.
A deep learning framework for character motion synthesis and editing
81.
Multi-Object Sketch Animation with Grouping and Motion Trajectory Priors
82.
TRACE: Learning 3D Gaussian Physical Dynamics from Multi-view Videos
83.
X-MoGen: Unified Motion Generation across Humans and Animals
84.
Gaussian Variation Field Diffusion for High-fidelity Video-to-4D Synthesis
85.
MotionShot: Adaptive Motion Transfer across Arbitrary Objects for Text-to-Video Generation
86.
Drop: Dynamics responses from human motion prior and projective dynamics
87.
POMP: Physics-constrainable Motion Generative Model through Phase Manifolds
88.
Dreamgaussian4d: Generative 4d gaussian splatting
89.
Drive Any Mesh: 4D Latent Diffusion for Mesh Deformation from Video
90.
AnimateAnyMesh: A Feed-Forward 4D Foundation Model for Text-Driven Universal Mesh Animation
91.
ReVision: High-Quality, Low-Cost Video Generation with Explicit 3D Physics Modeling for Complex Motion and Interaction
92.
Text2Video-Zero: Text-to-Image Diffusion Models are Zero-Shot Video Generators
93.
Force Prompting: Video Generation Models Can Learn and Generalize Physics-based Control Signals
94.
Think Before You Diffuse: LLMs-Guided Physics-Aware Video Generation
95.
Generating time-consistent dynamics with discriminator-guided image diffusion models
96.
GENMO:AGENeralist Model for Human MOtion
97.
HGM3: HIERARCHICAL GENERATIVE MASKED MOTION MODELING WITH HARD TOKEN MINING
98.
Towards Robust and Controllable Text-to-Motion via Masked Autoregressive Diffusion
99.
MoCLIP: Motion-Aware Fine-Tuning and Distillation of CLIP for Human Motion Generation
100.
FinePhys: Fine-grained Human Action Generation by Explicitly Incorporating Physical Laws for Effective Skeletal Guidance
101.
VideoSwap: Customized Video Subject Swapping with Interactive Semantic Point Correspondence
102.
DragAnything: Motion Control for Anything using Entity Representation
103.
PhysAnimator: Physics-Guided Generative Cartoon Animation
104.
SOAP: Style-Omniscient Animatable Portraits
105.
Neural Discrete Representation Learning
106.
TSTMotion: Training-free Scene-aware Text-to-motion Generation
107.
Deterministic-to-Stochastic Diverse Latent Feature Mapping for Human Motion Synthesis
108.
A lip sync expert is all you need for speech to lip generation in the wild
109.
MUSETALK: REAL-TIME HIGH QUALITY LIP SYN-CHRONIZATION WITH LATENT SPACE INPAINTING
110.
LatentSync: Audio Conditioned Latent Diffusion Models for Lip Sync
111.
T2m-gpt: Generating human motion from textual descriptions with discrete representations
112.
Motiongpt: Finetuned llms are general-purpose motion generators
113.
Guided Motion Diffusion for Controllable Human Motion Synthesis
114.
OmniControl: Control Any Joint at Any Time for Human Motion Generation
115.
Learning Long-form Video Prior via Generative Pre-Training
116.
Instant Neural Graphics Primitives with a Multiresolution Hash Encoding
117.
Magic3D: High-Resolution Text-to-3D Content Creation
118.
CogVideo: Large-scale Pretraining for Text-to-Video Generation via Transformers
119.
One-Minute Video Generation with Test-Time Training
120.
Key-Locked Rank One Editing for Text-to-Image Personalization
121.
MARCHING CUBES: A HIGH RESOLUTION 3D SURFACE CONSTRUCTION ALGORITHM
122.
Plug-and-Play Diffusion Features for Text-Driven Image-to-Image Translation
123.
NULL-text Inversion for Editing Real Images Using Guided Diffusion Models
124.
simple diffusion: End-to-end diffusion for high resolution images
125.
One Transformer Fits All Distributions in Multi-Modal Diffusion at Scale
126.
Scalable Diffusion Models with Transformers
127.
All are Worth Words: a ViT Backbone for Score-based Diffusion Models
128.
An image is worth 16x16 words: Transformers for image recognition at scale
129.
eDiff-I: Text-to-Image Diffusion Models with an Ensemble of Expert Denoisers
130.
Photorealistic text-to-image diffusion models with deep language understanding||Imagen
131.
DreamFusion: Text-to-3D using 2D Diffusion
132.
GLIGEN: Open-Set Grounded Text-to-Image Generation
133.
Adding Conditional Control to Text-to-Image Diffusion Models
134.
T2I-Adapter: Learning Adapters to Dig out More Controllable Ability for Text-to-Image Diffusion Models
135.
Multi-Concept Customization of Text-to-Image Diffusion
136.
An Image is Worth One Word: Personalizing Text-to-Image Generation using Textual Inversion
137.
DreamBooth: Fine Tuning Text-to-Image Diffusion Models for Subject-Driven Generation
138.
VisorGPT: Learning Visual Prior via Generative Pre-Training
139.
NUWA-XL: Diffusion over Diffusion for eXtremely Long Video Generation
140.
AnimateDiff: Animate Your Personalized Text-to-Image Diffusion Models without Specific Tuning
141.
ModelScope Text-to-Video Technical Report
142.
Show-1: Marrying Pixel and Latent Diffusion Models for Text-to-Video Generation
143.
Make-A-Video: Text-to-Video Generation without Text-Video Data
144.
Video Diffusion Models
145.
Learning Transferable Visual Models From Natural Language Supervision
146.
Implicit Warping for Animation with Image Sets
147.
Mix-of-Show: Decentralized Low-Rank Adaptation for Multi-Concept Customization of Diffusion Models
148.
Motion-Conditioned Diffusion Model for Controllable Video Synthesis
149.
Stable Video Diffusion: Scaling Latent Video Diffusion Models to Large Datasets
150.
UniAnimate: Taming Unified Video Diffusion Models for Consistent Human Image Animation
151.
Align your Latents: High-Resolution Video Synthesis with Latent Diffusion Models
152.
Puppet-Master: Scaling Interactive Video Generation as a Motion Prior for Part-Level Dynamics
153.
A Recipe for Scaling up Text-to-Video Generation
154.
High-Resolution Image Synthesis with Latent Diffusion Models
155.
Motion-I2V: Consistent and Controllable Image-to-Video Generation with Explicit Motion Modeling
156.
数据集:HumanVid
157.
HumanVid: Demystifying Training Data for Camera-controllable Human Image Animation
158.
StoryDiffusion: Consistent Self-Attention for Long-Range Image and Video Generation
159.
数据集:Zoo-300K
160.
Motion Avatar: Generate Human and Animal Avatars with Arbitrary Motion
161.
LORA: LOW-RANK ADAPTATION OF LARGE LAN-GUAGE MODELS
162.
TCAN: Animating Human Images with Temporally Consistent Pose Guidance using Diffusion Models
163.
GaussianAvatar: Towards Realistic Human Avatar Modeling from a Single Video via Animatable 3D Gaussians
164.
MagicPony: Learning Articulated 3D Animals in the Wild
165.
Splatter a Video: Video Gaussian Representation for Versatile Processing
166.
数据集:Dynamic Furry Animal Dataset
167.
Artemis: Articulated Neural Pets with Appearance and Motion Synthesis
168.
SMPLer: Taming Transformers for Monocular 3D Human Shape and Pose Estimation
169.
CAT3D: Create Anything in 3D with Multi-View Diffusion Models
170.
PACER+: On-Demand Pedestrian Animation Controller in Driving Scenarios
171.
Humans in 4D: Reconstructing and Tracking Humans with Transformers
172.
Learning Human Motion from Monocular Videos via Cross-Modal Manifold Alignment
173.
PhysPT: Physics-aware Pretrained Transformer for Estimating Human Dynamics from Monocular Videos
174.
Imagic: Text-Based Real Image Editing with Diffusion Models
175.
DiffEdit: Diffusion-based semantic image editing with mask guidance
176.
Dual diffusion implicit bridges for image-to-image translation
177.
SDEdit: Guided Image Synthesis and Editing with Stochastic Differential Equations
178.
Prompt-to-Prompt Image Editing with Cross-Attention Control
179.
WANDR: Intention-guided Human Motion Generation
180.
TRAM: Global Trajectory and Motion of 3D Humans from in-the-wild Videos
181.
3D Gaussian Splatting for Real-Time Radiance Field Rendering
182.
Decoupling Human and Camera Motion from Videos in the Wild
183.
HMP: Hand Motion Priors for Pose and Shape Estimation from Video
184.
HuMoR: 3D Human Motion Model for Robust Pose Estimation
185.
Co-Evolution of Pose and Mesh for 3D Human Body Estimation from Video
186.
Global-to-Local Modeling for Video-based 3D Human Pose and Shape Estimation
187.
WHAM: Reconstructing World-grounded Humans with Accurate 3D Motion
188.
Tackling the Generative Learning Trilemma with Denoising Diffusion GANs
189.
Elucidating the Design Space of Diffusion-Based Generative Models
190.
SCORE-BASED GENERATIVE MODELING THROUGHSTOCHASTIC DIFFERENTIAL EQUATIONS
191.
Consistency Models
192.
Classifier-Free Diffusion Guidance
193.
Cascaded Diffusion Models for High Fidelity Image Generation
194.
LEARNING ENERGY-BASED MODELS BY DIFFUSIONRECOVERY LIKELIHOOD
195.
On Distillation of Guided Diffusion Models
196.
Denoising Diffusion Implicit Models
197.
PROGRESSIVE DISTILLATION FOR FAST SAMPLING OF DIFFUSION MODELS
198.
Instruct-NeRF2NeRF: Editing 3D Scenes with Instructions
199.
ControlVideo: Training-free Controllable Text-to-Video Generation
200.
Pix2Video: Video Editing using Image Diffusion
201.
Structure and Content-Guided Video Synthesis with Diffusion Models
202.
MagicAnimate: Temporally Consistent Human Image Animation using Diffusion Model
203.
MotionDirector: Motion Customization of Text-to-Video Diffusion Models
204.
Dreamix: Video Diffusion Models are General Video Editors
205.
Tune-A-Video: One-Shot Tuning of Image Diffusion Models for Text-to-Video Generation
206.
TokenFlow: Consistent Diffusion Features for Consistent Video Editing
207.
DynVideo-E: Harnessing Dynamic NeRF for Large-Scale Motion- and View-Change Human-Centric Video Editing
208.
Content Deformation Fields for Temporally Consistent Video Processing
209.
PFNN: Phase-Functioned Neural Networks
210.
Recurrent Transition Networks for Character Locomotion
211.
Real-Time Style Modelling of Human Locomotion
212.
Motion In-Betweening with Phase Manifolds
213.
Mode-Adaptive Neural Networks for Quadruped Motion Control
214.
Few-shot Learning of Homogeneous Human Locomotion Styles
215.
Learning predict-and-simulate policies from unorganized human motion data
216.
Local Motion Phases for Learning Multi-Contact Character Movements
217.
Interactive Control of Diverse Complex Characters with Neural Networks
218.
Accelerated Auto-regressive Motion Diffusion Model
219.
DARTControl: A Diffusion-based Autoregressive Motion Model for Real-time Text-driven Motion Control
220.
Interactive Character Control with Auto-Regressive Motion Diffusion Models
221.
Taming Diffusion Probabilistic Models for Character Control
222.
Learned Motion Matching
223.
MOCHA: Real-Time Motion Characterization via Context Matching
224.
DeepLoco: Dynamic Locomotion Skills Using Hierarchical Deep Reinforcement Learning
225.
Benchmarking Deep Reinforcement Learning for Continuous Control
226.
SIMBICON: Simple Biped Locomotion Control
227.
RLOC: Terrain-Aware Legged Locomotion using Reinforcement Learning and Optimal Control
228.
Efficient Self-Supervised Data Collection for Offline Robot Learning
229.
Learning Robust Autonomous Navigation and Locomotion for Wheeled-Legged Robots
230.
Dataset Distillation for Offline Reinforcement Learning
231.
mimic-one: A Scalable Model Recipe for General Purpose Robot Dexterity
Light (default)
Rust
Coal
Navy
Ayu
ReadPapers
数据集:Zoo-300K
该数据集包含约 300,000 对文本描述和跨越 65 个不同动物类别的相应动物运动。
原始数据
Truebones Zoo [2] 数据集
合成数据
对原始数据的动作进行增强
人工标注
用表示动物和运动类别的文本标签进行注释
生成标注
reference
论文:
link