MUST FOLLOWS
updated
Explorative Inbetweening of Time and Space
Paper
• 2403.14611
• Published
• 13
MovieLLM: Enhancing Long Video Understanding with AI-Generated Movies
Paper
• 2403.01422
• Published
• 30
DiLightNet: Fine-grained Lighting Control for Diffusion-based Image
Generation
Paper
• 2402.11929
• Published
• 11
StreamingT2V: Consistent, Dynamic, and Extendable Long Video Generation
from Text
Paper
• 2403.14773
• Published
• 11
SDXS: Real-Time One-Step Latent Diffusion Models with Image Conditions
Paper
• 2403.16627
• Published
• 22
CameraCtrl: Enabling Camera Control for Text-to-Video Generation
Paper
• 2404.02101
• Published
• 24
PointInfinity: Resolution-Invariant Point Diffusion Models
Paper
• 2404.03566
• Published
• 16
MotionLCM: Real-time Controllable Motion Generation via Latent
Consistency Model
Paper
• 2404.19759
• Published
• 27
InstantFamily: Masked Attention for Zero-shot Multi-ID Image Generation
Paper
• 2404.19427
• Published
• 74
StoryDiffusion: Consistent Self-Attention for Long-Range Image and Video
Generation
Paper
• 2405.01434
• Published
• 56
LogoMotion: Visually Grounded Code Generation for Content-Aware
Animation
Paper
• 2405.07065
• Published
• 18
Naturalistic Music Decoding from EEG Data via Latent Diffusion Models
Paper
• 2405.09062
• Published
• 13
Visual Echoes: A Simple Unified Transformer for Audio-Visual Generation
Paper
• 2405.14598
• Published
• 13
Searching Priors Makes Text-to-Video Synthesis Better
Paper
• 2406.03215
• Published
• 13