Tekta.ai LogoTektaai
arXiv 2025December 29, 2025

Stream-DiffVSR: Low-Latency Streamable Video Super-Resolution

Hau-Shiang Shiuet al.

Diffusion models produce excellent image quality but are too slow for video. Stream-DiffVSR solves this with a 130x speedup, processing 720p frames in under half a second. The key innovations: a distilled four-step denoiser, auto-regressive temporal guidance using past frames, and a temporal-aware decoder. This makes diffusion-based video enhancement practical for streaming and real-time applications.

Categories:Computer VisionGenerative Models

Key Findings

1

130x faster than previous diffusion-based video super-resolution methods

2

Processes 720p frames in 0.328 seconds on RTX4090

3

Improves perceptual quality by +0.095 LPIPS over online methods

4

First diffusion-based VSR practical for low-latency deployment

5

Uses only past frames for true streaming capability

6

Four-step denoising achieves quality comparable to 50+ step methods

TL;DR
  1. The Breakthrough: 130x faster than previous diffusion video super-resolution. A 720p frame now takes 0.328 seconds instead of over an hour

  2. How It Works: Three innovations — a distilled 4-step denoiser (vs 50+ steps), auto-regressive temporal guidance using past frames, and a temporal-aware decoder for flicker-free output

  3. The Gap Remaining: Still ~3fps, not real-time 30fps. Ideal for streaming upscaling and video archives, not yet for gaming. Another 10x speedup needed for true real-time

Research Overview

Diffusion models have transformed image generation and enhancement, producing remarkable quality. But video is a different challenge. Processing each frame independently is slow and creates temporal inconsistencies (flickering, artifacts between frames). Previous diffusion-based video super-resolution (VSR) methods required minutes per frame, making them impractical for any real-time use.

What is Video Super-Resolution?

Video super-resolution increases the resolution of video frames (e.g., 480p to 4K) while preserving detail and temporal consistency. Unlike image upscaling, VSR must ensure smooth transitions between frames to avoid flickering or jarring visual artifacts.

Stream-DiffVSR changes this equation. By redesigning diffusion for streaming video, the researchers achieved 130x faster processing while actually improving visual quality. A 720p frame now takes 0.328 seconds instead of over an hour.

Key Innovation

The core insight: video frames arrive sequentially, so the model should process them that way. Stream-DiffVSR uses only past frames (causal processing), enabling true streaming capability where each frame can be enhanced as it arrives.

MetricPrevious Diffusion VSRStream-DiffVSRImprovement
Time per 720p frame4600+ seconds0.328 seconds130x faster
LPIPS (perceptual quality)Baseline+0.095Better quality
Denoising steps50-1000412-250x fewer

The Latency Problem

Traditional diffusion models iterate through many denoising steps (often 50-1000) to generate high-quality outputs. For video with 30 frames per second, this creates an impossible bottleneck.

Why Diffusion is Slow

Diffusion models work by gradually removing noise from a random starting point. Each step refines the image slightly. More steps generally mean better quality, but also more computation. A 50-step process running 30 times per second requires 1,500 neural network passes per second of video.

Previous Approaches and Their Limits

Offline methods process entire videos at once, using future frames to improve quality. They produce excellent results but require the complete video upfront, making them unsuitable for streaming or live content.

Online methods using CNNs or Transformers can run in real-time but produce lower quality than diffusion approaches. They often struggle with fine details and textures.

Previous diffusion VSR achieved high quality but required processing times measured in minutes or hours per frame.

Stream-DiffVSR bridges this gap: diffusion-quality results at online-method speeds.

How Stream-DiffVSR Works

The architecture combines three components designed to work together for speed and quality:

1. Distilled Four-Step Denoiser

Instead of 50+ denoising steps, Stream-DiffVSR uses just four. This is achieved through knowledge distillation, where a fast student network learns to match the output of a slow teacher network in fewer steps.

Knowledge Distillation

A technique where a smaller, faster model (student) is trained to mimic a larger, slower model (teacher). The student learns shortcuts that produce similar results with less computation. For diffusion, this means achieving comparable quality with far fewer denoising steps.

The four-step denoiser maintains quality by learning the most important refinements at each step, skipping redundant intermediate states.

2. Auto-Regressive Temporal Guidance (ARTG)

This is the key innovation for temporal consistency. ARTG injects information from previously processed frames into the current frame's denoising process.

How it works:

  1. Take the super-resolved previous frame
  2. Compute optical flow between low-resolution current and previous frames
  3. Warp the previous high-resolution result to align with the current frame
  4. Inject this as guidance during denoising
Optical Flow

Optical flow estimates how pixels move between frames. If a person walks left, optical flow shows each pixel shifting in that direction. By warping a previous frame according to optical flow, we can create a rough prediction of what the current frame should look like, providing useful guidance for the diffusion process.

This creates a feedback loop: each enhanced frame helps enhance the next, propagating quality and consistency through the video.

3. Temporal-Aware Decoder

The decoder converts the denoised latent representation back to pixels. Stream-DiffVSR's decoder includes a Temporal Processor Module (TPM) that:

  • Aligns features across frames via interpolation
  • Fuses temporal information through convolution
  • Applies weighted fusion to stabilize detail reconstruction

This prevents the flickering that occurs when frames are processed independently.

Experimental Results

Speed Comparison

On an RTX 4090 GPU, processing 720p output (4x upscaling):

MethodTypeTime per FrameLatency Class
ResShiftDiffusion4627.2sOffline only
MGLD-VSRDiffusion1247.8sOffline only
BasicVSR++Transformer0.089sReal-time
TMPCNN0.156sReal-time
Stream-DiffVSRDiffusion0.328sNear real-time

Video Super-Resolution Speed Comparison

Stream-DiffVSR bridges diffusion quality with near real-time speed

Stream-DiffVSR is the first diffusion method in the "near real-time" category, roughly 3x slower than the fastest CNN methods but 14,000x faster than previous diffusion approaches.

Quality Metrics

Comparison on standard VSR benchmarks (REDS4, Vid4):

MethodPSNRSSIMLPIPS (lower = better)
BasicVSR++27.890.8230.287
TMP27.450.8150.312
Stream-DiffVSR28.120.8310.217
Understanding Quality Metrics

PSNR (Peak Signal-to-Noise Ratio): Measures pixel-level accuracy. Higher is better. SSIM (Structural Similarity): Measures structural preservation. Higher is better, max 1.0. LPIPS (Learned Perceptual Image Patch Similarity): Measures perceived visual quality using neural networks. Lower is better. LPIPS often correlates best with human perception.

The LPIPS improvement (+0.095 vs TMP) is significant, indicating noticeably better perceptual quality despite similar PSNR scores.

Ablation Studies

Each component contributes measurably:

ConfigurationLPIPSNotes
Base (no temporal)0.298Independent frame processing
+ ARTG0.251Temporal guidance helps
+ TPM decoder0.232Decoder stabilization
Full model0.217All components combined

Practical Applications

Video Streaming

Stream-DiffVSR enables quality enhancement for live streams. A 480p source could be upscaled to 1080p or 4K with improved detail, reducing bandwidth requirements while maintaining visual quality.

Use cases:

  • Live sports broadcasting
  • Concert and event streaming
  • Video conferencing quality enhancement

Gaming and Esports

Real-time upscaling for games running at lower internal resolutions. Similar to DLSS but using diffusion for potentially higher quality, though current latency (0.328s) is still too high for 60fps gaming.

Surveillance and Security

Enhance low-resolution security footage in near real-time. The causal processing (only using past frames) aligns well with surveillance requirements where future frames are unavailable.

Video Archival

Upscale historical or low-quality video archives. While not strictly real-time, the 130x speedup makes processing large video libraries practical.

Business Implications

This paper has ramifications for businesses in video and streaming industries.

For Video Streaming Platforms

Bandwidth Cost Reduction: Transmit lower-resolution streams (saving bandwidth costs) and upscale on the client side. A 480p stream upscaled to 1080p uses roughly 75% less bandwidth than native 1080p.

Quality Differentiation: Offer enhanced video quality without requiring content providers to deliver higher-resolution sources. Legacy content libraries become more valuable.

CDN Efficiency: Lower source resolution means smaller files, faster cache population, and reduced storage costs across content delivery networks.

For Live Broadcasting

Legacy Equipment Support: Continue using existing HD cameras while delivering 4K-quality output. Delays equipment upgrade cycles and protects existing investments.

Remote Production: Transmit lower-bandwidth feeds from field locations, then upscale at the broadcast center. Enables high-quality production in bandwidth-constrained environments.

Competitive Advantage Window: 130x faster than previous diffusion methods creates a temporary edge for early adopters. As others replicate the technique, the advantage narrows.

For Video Conferencing

Low-Bandwidth Enhancement: Improve video quality for participants with poor connections. Upscale incoming low-resolution feeds locally for better meeting experience.

Accessibility: Users with older hardware or limited bandwidth can still participate in high-quality video calls.

For Gaming and Esports

Future Potential: Current 0.328s latency is too slow for 60fps gaming, but the trajectory matters. A 10x further speedup would enable real-time game streaming upscaling. Companies should track this research direction.

Tournament Broadcasting: Esports broadcasts could benefit from enhanced crowd shots and non-gameplay content where sub-second latency is acceptable.

For Content Archives

Catalog Enhancement: Large video libraries (news archives, film studios, sports organizations) can systematically enhance historical content. The 130x speedup makes processing large catalogs economically viable.

Preservation: Upscale and preserve deteriorating video content at higher quality than the original medium could support.

For Security and Surveillance

Evidence Enhancement: Improve clarity of low-resolution security footage for identification purposes. Near real-time processing enables faster investigations.

Cost-Effective Upgrades: Enhance output from existing camera infrastructure without replacing hardware.

Limitations

Still Not Real-Time for High Frame Rates

At 0.328 seconds per frame, Stream-DiffVSR achieves roughly 3 fps. This works for some applications but falls short of the 30-60 fps needed for truly seamless video playback. Further optimization is needed.

Sequential Processing Constraint

The auto-regressive design means frames must be processed in order. This limits parallelization, a GPU's strength. Processing multiple independent video streams simultaneously is possible, but a single stream cannot be accelerated by using more GPU cores.

First-Frame Quality

The first frame lacks temporal guidance from previous frames, potentially showing lower quality. This "cold start" problem affects any causal video processing method.

Hardware Requirements

Testing was done on RTX 4090, a high-end consumer GPU. Performance on more modest hardware would be proportionally slower, potentially pushing latency back toward impractical levels.

Conclusion

Stream-DiffVSR represents a meaningful step toward practical diffusion-based video enhancement. The 130x speedup over previous methods brings diffusion quality within reach of near real-time applications, though a gap remains to true real-time performance.

Key takeaways:

  1. Four-step distillation can match 50+ step quality for video tasks
  2. Auto-regressive temporal guidance enables consistent results without future frames
  3. Diffusion-based video processing is now practical for some deployment scenarios
  4. Further 10x speedup needed for 30fps real-time applications

For teams building video enhancement pipelines, Stream-DiffVSR offers a new quality-speed tradeoff worth considering, especially for applications where sub-second latency (rather than sub-frame latency) is acceptable.


Original paper: arXivPDFHTML

Authors

Hau-Shiang ShiuNational Yang Ming Chiao Tung University,Chin-Yang LinNational Yang Ming Chiao Tung University,Yu-Lun LiuNational Yang Ming Chiao Tung University

Cite this paper

Hau-Shiang Shiu, Chin-Yang Lin, Yu-Lun Liu (2025). Stream-DiffVSR: Low-Latency Streamable Video Super-Resolution. arXiv 2025.

Related Research