menu
close

MagicTime: AI Model Learns Physics to Create Realistic Metamorphic Videos

Computer scientists have developed MagicTime, a groundbreaking AI text-to-video model that learns real-world physics knowledge from time-lapse data. Released on May 5, 2025, this collaborative effort from researchers at the University of Rochester, Peking University, UC Santa Cruz, and National University of Singapore represents a significant advancement in generating metamorphic videos that accurately simulate physical transformations. The technology could revolutionize scientific visualization, content creation, and educational tools by enabling more realistic video generation from simple text descriptions.
MagicTime: AI Model Learns Physics to Create Realistic Metamorphic Videos

While text-to-video AI models like OpenAI's Sora have made impressive strides in video generation, they've struggled with creating realistic metamorphic videos - those showing gradual transformations like flowers blooming or buildings under construction. These processes are particularly challenging for AI to simulate because they require deep understanding of real-world physics and can vary widely in appearance.

The newly developed MagicTime model addresses this limitation by learning physical knowledge directly from time-lapse videos. Led by PhD student Jinfa Huang and Professor Jiebo Luo from Rochester's Department of Computer Science, the international research team trained their model on a high-quality dataset of over 2,000 meticulously captioned time-lapse videos to capture the nuances of physical transformations.

The current open-source version generates two-second clips at 512×512 pixel resolution, while an accompanying diffusion-transformer architecture extends this to ten-second videos. MagicTime can simulate various metamorphic processes including biological growth, construction projects, and even culinary transformations like bread baking.

"MagicTime is a step toward AI that can better simulate the physical, chemical, biological, or social properties of the world around us," explains Huang. The researchers envision significant scientific applications beyond entertainment, suggesting that "biologists could use generative video to speed up preliminary exploration of ideas" while reducing the need for physical experiments.

The technology's implications extend across multiple fields. In education, it could create dynamic visualizations of complex processes that are difficult to observe in real-time. For content creators and the entertainment industry, it offers new tools for special effects and storytelling. Scientists might use it to model and predict physical phenomena, potentially accelerating research in fields ranging from biology to materials science.

As AI continues integrating more deeply with physical modeling, MagicTime exemplifies how embedding domain-specific knowledge into generative models can produce outcomes that are not only visually compelling but scientifically meaningful. The research was published in IEEE Transactions on Pattern Analysis and Machine Intelligence.

Source: Sciencedaily

Latest News