ReadPapers
1.
Introduction
2.
GLIGEN: Open-Set Grounded Text-to-Image Generation
3.
Adding Conditional Control to Text-to-Image Diffusion Models
4.
T2I-Adapter: Learning Adapters to Dig out More Controllable Ability for Text-to-Image Diffusion Models
5.
Multi-Concept Customization of Text-to-Image Diffusion
6.
An Image is Worth One Word: Personalizing Text-to-Image Generation using Textual Inversion
7.
DreamBooth: Fine Tuning Text-to-Image Diffusion Models for Subject-Driven Generation
8.
VisorGPT: Learning Visual Prior via Generative Pre-Training
9.
NUWA-XL: Diffusion over Diffusion for eXtremely Long Video Generation
10.
AnimateDiff: Animate Your Personalized Text-to-Image Diffusion Models without Specific Tuning
11.
ModelScope Text-to-Video Technical Report
12.
Show-1: Marrying Pixel and Latent Diffusion Models for Text-to-Video Generation
13.
Make-A-Video: Text-to-Video Generation without Text-Video Data
14.
Video Diffusion Models
15.
Learning Transferable Visual Models From Natural Language Supervision
16.
Implicit Warping for Animation with Image Sets
17.
Mix-of-Show: Decentralized Low-Rank Adaptation for Multi-Concept Customization of Diffusion Models
18.
Motion-Conditioned Diffusion Model for Controllable Video Synthesis
19.
Stable Video Diffusion: Scaling Latent Video Diffusion Models to Large Datasets
20.
UniAnimate: Taming Unified Video Diffusion Models for Consistent Human Image Animation
21.
Align your Latents: High-Resolution Video Synthesis with Latent Diffusion Models
22.
Puppet-Master: Scaling Interactive Video Generation as a Motion Prior for Part-Level Dynamics
23.
A Recipe for Scaling up Text-to-Video Generation
24.
High-Resolution Image Synthesis with Latent Diffusion Models
25.
Motion-I2V: Consistent and Controllable Image-to-Video Generation with Explicit Motion Modeling
26.
数据集:HumanVid
27.
HumanVid: Demystifying Training Data for Camera-controllable Human Image Animation
28.
StoryDiffusion: Consistent Self-Attention for Long-Range Image and Video Generation
29.
数据集:Zoo-300K
30.
Motion Avatar: Generate Human and Animal Avatars with Arbitrary Motion
31.
LORA: LOW-RANK ADAPTATION OF LARGE LAN-GUAGE MODELS
32.
TCAN: Animating Human Images with Temporally Consistent Pose Guidance using Diffusion Models
33.
GaussianAvatar: Towards Realistic Human Avatar Modeling from a Single Video via Animatable 3D Gaussians
34.
MagicPony: Learning Articulated 3D Animals in the Wild
35.
Splatter a Video: Video Gaussian Representation for Versatile Processing
36.
数据集:Dynamic Furry Animal Dataset
37.
Artemis: Articulated Neural Pets with Appearance and Motion Synthesis
38.
SMPLer: Taming Transformers for Monocular 3D Human Shape and Pose Estimation
39.
CAT3D: Create Anything in 3D with Multi-View Diffusion Models
40.
PACER+: On-Demand Pedestrian Animation Controller in Driving Scenarios
41.
Humans in 4D: Reconstructing and Tracking Humans with Transformers
42.
Learning Human Motion from Monocular Videos via Cross-Modal Manifold Alignment
43.
PhysPT: Physics-aware Pretrained Transformer for Estimating Human Dynamics from Monocular Videos
44.
Imagic: Text-Based Real Image Editing with Diffusion Models
45.
DiffEdit: Diffusion-based semantic image editing with mask guidance
46.
Dual diffusion implicit bridges for image-to-image translation
47.
SDEdit: Guided Image Synthesis and Editing with Stochastic Differential Equations
48.
Prompt-to-Prompt Image Editing with Cross-Attention Control
49.
WANDR: Intention-guided Human Motion Generation
50.
TRAM: Global Trajectory and Motion of 3D Humans from in-the-wild Videos
51.
3D Gaussian Splatting for Real-Time Radiance Field Rendering
52.
Decoupling Human and Camera Motion from Videos in the Wild
53.
HMP: Hand Motion Priors for Pose and Shape Estimation from Video
54.
HuMoR: 3D Human Motion Model for Robust Pose Estimation
55.
Co-Evolution of Pose and Mesh for 3D Human Body Estimation from Video
56.
Global-to-Local Modeling for Video-based 3D Human Pose and Shape Estimation
57.
WHAM: Reconstructing World-grounded Humans with Accurate 3D Motion
58.
Tackling the Generative Learning Trilemma with Denoising Diffusion GANs
59.
Elucidating the Design Space of Diffusion-Based Generative Models
60.
SCORE-BASED GENERATIVE MODELING THROUGHSTOCHASTIC DIFFERENTIAL EQUATIONS
61.
Consistency Models
62.
Classifier-Free Diffusion Guidance
63.
Cascaded Diffusion Models for High Fidelity Image Generation
64.
LEARNING ENERGY-BASED MODELS BY DIFFUSIONRECOVERY LIKELIHOOD
65.
On Distillation of Guided Diffusion Models
66.
Denoising Diffusion Implicit Models
67.
PROGRESSIVE DISTILLATION FOR FAST SAMPLING OF DIFFUSION MODELS
Light (default)
Rust
Coal
Navy
Ayu
ReadPapers
数据集:Zoo-300K
该数据集包含约 300,000 对文本描述和跨越 65 个不同动物类别的相应动物运动。
原始数据
Truebones Zoo [2] 数据集
合成数据
对原始数据的动作进行增强
人工标注
用表示动物和运动类别的文本标签进行注释
生成标注
reference
论文:
link