In recent years, diffusion models have rapidly overtaken previous methods to become the dominant approach in visual generative modeling, with widespread applications in generating images, videos, 3D objects, and more. However, these models also come with notable limitations, such as slow generation speeds, limited human intervention during the generation process, and challenges in modeling complex distributions like long videos.
This year, our Visual Generative Modeling workshop at CVPR aims to explore what lies beyond diffusion models in visual generative modeling. We will discuss novel insights, alternative approaches, and new possibilities in modeling and generating visual data. Join us for a full-day event featuring keynote talks from top universities and industrial labs -- all designed to ignite innovative ideas and novel research in visual generative modeling.
Time | Session Details | Speaker | Affiliation |
---|---|---|---|
9:00 – 9:15 | Welcome & Opening Remarks | — | — |
Academic Session | |||
9:15 – 9:45 | Keynote Talk | Bill Freeman | MIT |
9:45 – 10:15 | Keynote Talk | Phillip Isola | MIT |
10:15 – 10:45 | Keynote Talk | Jun-Yan Zhu | CMU |
10:45 – 11:15 | Keynote Talk | Jiajun Wu | Stanford University |
11:15 – 11:45 | Keynote Talk | Kaiming He | MIT |
11:45 – 12:15 | Keynote Talk | Robin Rombach | Black Forest Labs |
12:15 – 13:30 | Lunch Break | ||
Industrial Session | |||
13:30 – 15:00 | Spotlight Talks (3 × 20 min + 5 min Q&A) | TBD | |
15:00 – 15:30 | Keynote Talk | Varun Jampani | Stability AI |
15:30 – 16:00 | Keynote Talk | Arash Vahdat | NVIDIA |
16:00 – 16:30 | Keynote Talk | Ricky T. Q. Chen | FAIR |
16:30 – 17:00 | Keynote Talk | Liang-Chieh Chen | TikTok |
17:00 – 17:30 | Panel: What's Next in Visual Generative Modeling in Industry | Industry Speakers |
Additional details and spotlight speakers will be updated later.