ReadPapers
1.
Introduction
2.
DreamFusion: Text-to-3D using 2D Diffusion
3.
GLIGEN: Open-Set Grounded Text-to-Image Generation
4.
Adding Conditional Control to Text-to-Image Diffusion Models
5.
T2I-Adapter: Learning Adapters to Dig out More Controllable Ability for Text-to-Image Diffusion Models
6.
Multi-Concept Customization of Text-to-Image Diffusion
7.
An Image is Worth One Word: Personalizing Text-to-Image Generation using Textual Inversion
8.
DreamBooth: Fine Tuning Text-to-Image Diffusion Models for Subject-Driven Generation
9.
VisorGPT: Learning Visual Prior via Generative Pre-Training
10.
NUWA-XL: Diffusion over Diffusion for eXtremely Long Video Generation
11.
AnimateDiff: Animate Your Personalized Text-to-Image Diffusion Models without Specific Tuning
12.
ModelScope Text-to-Video Technical Report
13.
Show-1: Marrying Pixel and Latent Diffusion Models for Text-to-Video Generation
14.
Make-A-Video: Text-to-Video Generation without Text-Video Data
15.
Video Diffusion Models
16.
Learning Transferable Visual Models From Natural Language Supervision
17.
Implicit Warping for Animation with Image Sets
18.
Mix-of-Show: Decentralized Low-Rank Adaptation for Multi-Concept Customization of Diffusion Models
19.
Motion-Conditioned Diffusion Model for Controllable Video Synthesis
20.
Stable Video Diffusion: Scaling Latent Video Diffusion Models to Large Datasets
21.
UniAnimate: Taming Unified Video Diffusion Models for Consistent Human Image Animation
22.
Align your Latents: High-Resolution Video Synthesis with Latent Diffusion Models
23.
Puppet-Master: Scaling Interactive Video Generation as a Motion Prior for Part-Level Dynamics
24.
A Recipe for Scaling up Text-to-Video Generation
25.
High-Resolution Image Synthesis with Latent Diffusion Models
26.
Motion-I2V: Consistent and Controllable Image-to-Video Generation with Explicit Motion Modeling
27.
数据集:HumanVid
28.
HumanVid: Demystifying Training Data for Camera-controllable Human Image Animation
29.
StoryDiffusion: Consistent Self-Attention for Long-Range Image and Video Generation
30.
数据集:Zoo-300K
31.
Motion Avatar: Generate Human and Animal Avatars with Arbitrary Motion
32.
LORA: LOW-RANK ADAPTATION OF LARGE LAN-GUAGE MODELS
33.
TCAN: Animating Human Images with Temporally Consistent Pose Guidance using Diffusion Models
34.
GaussianAvatar: Towards Realistic Human Avatar Modeling from a Single Video via Animatable 3D Gaussians
35.
MagicPony: Learning Articulated 3D Animals in the Wild
36.
Splatter a Video: Video Gaussian Representation for Versatile Processing
37.
数据集:Dynamic Furry Animal Dataset
38.
Artemis: Articulated Neural Pets with Appearance and Motion Synthesis
39.
SMPLer: Taming Transformers for Monocular 3D Human Shape and Pose Estimation
40.
CAT3D: Create Anything in 3D with Multi-View Diffusion Models
41.
PACER+: On-Demand Pedestrian Animation Controller in Driving Scenarios
42.
Humans in 4D: Reconstructing and Tracking Humans with Transformers
43.
Learning Human Motion from Monocular Videos via Cross-Modal Manifold Alignment
44.
PhysPT: Physics-aware Pretrained Transformer for Estimating Human Dynamics from Monocular Videos
45.
Imagic: Text-Based Real Image Editing with Diffusion Models
46.
DiffEdit: Diffusion-based semantic image editing with mask guidance
47.
Dual diffusion implicit bridges for image-to-image translation
48.
SDEdit: Guided Image Synthesis and Editing with Stochastic Differential Equations
49.
Prompt-to-Prompt Image Editing with Cross-Attention Control
50.
WANDR: Intention-guided Human Motion Generation
51.
TRAM: Global Trajectory and Motion of 3D Humans from in-the-wild Videos
52.
3D Gaussian Splatting for Real-Time Radiance Field Rendering
53.
Decoupling Human and Camera Motion from Videos in the Wild
54.
HMP: Hand Motion Priors for Pose and Shape Estimation from Video
55.
HuMoR: 3D Human Motion Model for Robust Pose Estimation
56.
Co-Evolution of Pose and Mesh for 3D Human Body Estimation from Video
57.
Global-to-Local Modeling for Video-based 3D Human Pose and Shape Estimation
58.
WHAM: Reconstructing World-grounded Humans with Accurate 3D Motion
59.
Tackling the Generative Learning Trilemma with Denoising Diffusion GANs
60.
Elucidating the Design Space of Diffusion-Based Generative Models
61.
SCORE-BASED GENERATIVE MODELING THROUGHSTOCHASTIC DIFFERENTIAL EQUATIONS
62.
Consistency Models
63.
Classifier-Free Diffusion Guidance
64.
Cascaded Diffusion Models for High Fidelity Image Generation
65.
LEARNING ENERGY-BASED MODELS BY DIFFUSIONRECOVERY LIKELIHOOD
66.
On Distillation of Guided Diffusion Models
67.
Denoising Diffusion Implicit Models
68.
PROGRESSIVE DISTILLATION FOR FAST SAMPLING OF DIFFUSION MODELS
Light (default)
Rust
Coal
Navy
Ayu
ReadPapers
数据集:Zoo-300K
原始数据
合成数据
人工标注
生成标注
reference
数据集:Zoo-300K
该数据集包含约 300,000 对文本描述和跨越 65 个不同动物类别的相应动物运动。
原始数据
Truebones Zoo [2] 数据集
合成数据
对原始数据的动作进行增强
人工标注
用表示动物和运动类别的文本标签进行注释
生成标注
reference
论文:
link