OmniInsert: Mask-Free Video Insertion of Any Reference via Diffusion Transformer Models Paper • 2509.17627 • Published Sep 22 • 66
HuMo: Human-Centric Video Generation via Collaborative Multi-Modal Conditioning Paper • 2509.08519 • Published Sep 10 • 128
UMO: Scaling Multi-Identity Consistency for Image Customization via Matching Reward Paper • 2509.06818 • Published Sep 8 • 29
USO: Unified Style and Subject-Driven Generation via Disentangled and Reward Learning Paper • 2508.18966 • Published Aug 26 • 56
Phantom-Data : Towards a General Subject-Consistent Video Generation Dataset Paper • 2506.18851 • Published Jun 23 • 30
MagicVideo-V2: Multi-Stage High-Aesthetic Video Generation Paper • 2401.04468 • Published Jan 9, 2024 • 49
DEADiff: An Efficient Stylization Diffusion Model with Disentangled Representations Paper • 2403.06951 • Published Mar 11, 2024 • 2
I2VControl: Disentangled and Unified Video Motion Synthesis Control Paper • 2411.17765 • Published Nov 26, 2024
Phantom: Subject-consistent video generation via cross-modal alignment Paper • 2502.11079 • Published Feb 16 • 59
I2VControl-Camera: Precise Video Camera Control with Adjustable Motion Strength Paper • 2411.06525 • Published Nov 10, 2024
Mask$^2$DiT: Dual Mask-based Diffusion Transformer for Multi-Scene Long Video Generation Paper • 2503.19881 • Published Mar 25 • 6
AR-Diffusion: Asynchronous Video Generation with Auto-Regressive Diffusion Paper • 2503.07418 • Published Mar 10
Seedance 1.0: Exploring the Boundaries of Video Generation Models Paper • 2506.09113 • Published Jun 10 • 104
Autoregressive Adversarial Post-Training for Real-Time Interactive Video Generation Paper • 2506.09350 • Published Jun 11 • 48
Seaweed-7B: Cost-Effective Training of Video Generation Foundation Model Paper • 2504.08685 • Published Apr 11 • 130
Less-to-More Generalization: Unlocking More Controllability by In-Context Generation Paper • 2504.02160 • Published Apr 2 • 37