
Bio
Anyi Rao is an Assistant Professor at the Hong Kong University of Science and Technology (HKUST). He leads the Multimedia Creativity Lab (MMLab@HKUST). He is the Associate Director of HKUST Media Intelligence Research Center. He studies human AI and agentic AI, focusing on the creation, editing and understanding of art, media and film, aiming to build human-AI collaborative intelligence and unleash human creativity and productivity. His works include ControlNet, AnimateDiff, MovieNet, Virtual Studio, and IC-Light, with a Marr Prize (ICCV best paper award). These works have been widely used in industry, including Amazon Prime Video, Netflix, Tencent, and more.
He was a Postdoctoral Scholar at Stanford with Maneesh Agrawala. He received the Ph.D. at MMLab, Chinese University of Hong Kong with Dahua Lin and Bolei Zhou. He has research experiences at Meta Reality Lab, Vector Institute, University of Toronto, Hong Kong University. He organized the SIGGRAPH/CVPR/ICCV/ECCV Creative Visual Content Workshop and the SIGGRAPH Generative Models Course, curated 2025 Hong Kong HKUST AI Film Festival and 2023 Paris ShortFest AI Film Festival. He also serves as a co-chair of MMSys26, UIST25, VINCI25, CVM25, UIST24 and area chair/TPC of CVPR26, ICLR26, SIGGRAPH Asia26, UIST26, SIGGRAPH Asia25.
He has hosted the Brown Media Innovation Research Fund, Amazon Video Research Fund, been featured in Forbes 30 Under 30 Asia 2025 List, and won the Rising Star Award at the World Artificial Intelligence Conference 2024. He gave keynote at the Golden Rooster Film Festival, the Shanghai Television Magnolia Festival, was featured by Shanghai TV Financial Channel, Hong Kong Cable Television.
Actively looking for highly motivated students to join the group. See openings for more details. Please fill out this 2026 form for 2026 intake and send an email to me if you are interested in.
News
More
Selected Publication [Full List]
| CineVision: An Interactive Pre-visualization Storyboard System for Director–Cinematographer Collaboration |
| VideoRepainter: Keyframe-Guided Creative Video Inpainting |
| IC-Light: Scaling In-the-Wild Training for Diffusion-based Illumination Harmonization and Editing by Imposing Consistent Light Transport |
| ScriptViz: A Visualization Tool to Aid Scriptwriting based on a Large Movie Database |
| Cinematic Behavior Transfer via NeRF-based Differentiable Filming |
| AnimateDiff: Animate Your Personalized Text-to-Image Diffusion Models without Specific Tuning |
| SparseCtrl: Adding Sparse Controls to Text-to-Video Diffusion Models |
| ControlNet: Adding Conditional Control to Text-to-Image Diffusion Models |
| Automated Conversion of Music Videos into Lyric Videos |
| Dynamic Storyboard Generation in an Engine-based Virtual Environments for Video Production |
| Shoot360: Normal View Video Creation from City Panorama Footage |
| A Coarse-to-Fine Framework for Automatic Video Unscreen |
| BungeeNeRF: Progressive Neural Radiance Field for Extreme Multi-scale Scene Rendering |
| A Local-to-Global Approach to Multi-modal Movie Scene Segmentation |
| A Unified Framework for Shot Type Classification Based on Subject Centric Lens |
| MovieNet: A Holistic Dataset for Movie Understanding |
Selected Awards and Grants
| 2025 | |
| 2025 | |
| 2025 | |
| 2025 | |
| 2024 | |
| 2024 | |
| 2024 | |
| 2023 | |
| 2023 | |
| 2023 |
More
| 2025 | |
| 2024 | |
| 2023 | |
| 2022 | |
| 2021 | |
| 2021 | |
| 2020 | |
| 2018 | |
| 2017 | |
| 2017 | |
| 2015 | |
| 2013 | |
| 2015 | |
| 2016 | |
| 2015 | |
| 2016 | |
| 2017 | |
| 2016 | |
| 2017 | |
| 2016 | |
| 2016 | |
| 2014 |
Talks
|
|
|
|
|
|
|
|
Press Coverage
Professional Activities
Research Experiences
Teaching Experiences
Patents















