TurboDiffusion: Tsinghua Framework Delivers 200x Video Generation Speedup on RTX 5090

Open-source project utilizes SageAttention and rCM distillation to slash inference times on 14B parameter models.

· 3 min read · PSEEDR Editorial

Researchers from Tsinghua University and ShengShu Technology have open-sourced TurboDiffusion, a new acceleration framework that reduces inference latency for large-scale video diffusion models by up to 200 times. By leveraging the NVIDIA RTX 5090 and novel distillation techniques, the project demonstrates the capability to compress hour-long generation tasks into less than a minute.

As the generative AI sector moves toward high-fidelity video, inference latency has remained a primary bottleneck for deployment. On December 23, 2025, Tsinghua University's TSAIL Lab, in collaboration with ShengShu Technology, released TurboDiffusion, a framework designed to address this computational overhead. Official benchmarks released alongside the code indicate that the system achieves "100 to 200 times end-to-end generation acceleration" on a single NVIDIA RTX 5090 graphics card.

Performance Benchmarks

The performance gains are most visible in high-parameter models, which traditionally require substantial compute time. According to the project's technical documentation, generating a 720P video using the Wan-2.2-I2V-14B model (a 14-billion parameter image-to-video model) originally required 4,549 seconds-approximately 75 minutes. With TurboDiffusion, this duration is reduced to 38 seconds, representing a speedup factor of approximately 120x.

For lighter workloads, the framework enables near-real-time generation. The documentation states that for the Wan-2.1-T2V-1.3B model generating 480P content, the original method took 184 seconds, whereas "TurboDiffusion completes it in just 1.9 seconds". This reduction to under two seconds for a 1.3-billion parameter model suggests that interactive video generation applications may now be viable on consumer flagship hardware.

Technical Architecture: SageAttention and rCM

The acceleration is achieved through a combination of architectural optimizations and distillation techniques rather than simple hardware scaling. TurboDiffusion reportedly utilizes "SageAttention," a specialized attention mechanism designed to optimize the computational flow of diffusion transformers. Additionally, the framework employs rCM (temporal step distillation), a method that reduces the number of sampling steps required to produce a coherent video output without collapsing the model's visual fidelity.

While the speed gains are significant, the reliance on aggressive distillation techniques like rCM raises questions regarding potential trade-offs in fine-detail retention or motion consistency, though the project claims to maintain video quality. Independent verification of Visual Benchmark (VBench) scores will be necessary to confirm if the 200x speedup preserves the full integrity of the original 14B model's output.

Hardware Synergy with RTX 5090

The benchmarks explicitly cite the NVIDIA RTX 5090, released in January 2025, as the testbed for these results. The RTX 5090 remains the current flagship consumer GPU, and its memory bandwidth and tensor core architecture likely play a critical role in supporting the throughput required by TurboDiffusion. It remains unclear how the framework performs on previous-generation hardware, such as the RTX 4090, or if the 100-200x acceleration curve holds linearly across older architectures with lower VRAM capacities.

Market Implications

The release of TurboDiffusion as an open-source project places immediate pressure on proprietary inference engines. By reducing the barrier to entry for running 14B+ parameter video models locally, Tsinghua and ShengShu Technology are effectively making high-end video synthesis accessible to a wider audience. This development may force competitors offering closed-source acceleration-such as commercial APIs wrapping standard diffusion pipelines-to optimize their stacks further to compete with the raw throughput available to local users on flagship hardware.

Key Takeaways

Sources