Google Researchers Propose Unified Latents Framework
- •Google introduces Unified Latents (UL) for more efficient latent representation learning.
- •Framework achieves 1.4 FID on ImageNet-512 with significantly reduced training compute.
- •New state-of-the-art 1.3 FVD score set on the Kinetics-600 video dataset.
Most current generative models rely on pre-trained latent spaces, but Google researchers are now rethinking this foundational pipeline. The new Unified Latents (UL) framework shifts how we train latent representations by integrating diffusion priors directly into the regularization process. This approach moves away from static, off-the-shelf encoders toward a more integrated system where the latent space is optimized specifically for the generative task at hand.
The core technical innovation lies in linking the encoder's output noise directly to the diffusion prior's minimum noise level. This creates a tight upper bound on the latent bitrate—essentially a mathematical measure of how much information is packed into the compressed representation. By decoding with a diffusion model, the system maintains high reconstruction quality (PSNR) while remaining computationally leaner than predecessors, effectively doing more with less data processing power.
On the standard ImageNet-512 benchmark, UL achieved a competitive Frechet Inception Distance (FID) of 1.4. Perhaps more impressively, it required fewer training FLOPs than models built on top of the ubiquitous Stable Diffusion latents. It also dominated video benchmarks, setting a new state-of-the-art for video quality (FVD) on the Kinetics-600 dataset.
This research suggests that co-optimizing encoders with diffusion-based priors isn't just a theoretical curiosity; it's a practical path toward more efficient and higher-fidelity generative AI. By tightening the link between compression and generation, UL paves the way for faster training of high-resolution image and video models.