Skip to content

SunnyMass/DCVC-DT

Repository files navigation

DCVC-DT

Official code for the ISCAS 2026 paper Neural Video Compression with Domain Transfer.

Environment setup, dataset layout, optional C++ bitstream build, and pretrained checkpoints are the same as DCVC-DC: follow the upstream README, place checkpoints under ./checkpoints, then use the commands below.

This repository adds test_video_y_latent_optimized.py only: copy it into the DCVC-DC project root (next to test_video.py).

TODO

We will gradually open-source the project components in the following order:

  • Release the binaries for HEVC Class C and D. (Note: You can directly use the DCVC-DC decoder to decode these bitstreams and compare the performance.)
  • Release the binaries for HEVC Class B and other 1080p sequences. (Note: You can directly use the DCVC-DC decoder to decode these bitstreams and compare the performance.)
  • Release the training code.

Introduction

Neural codecs are sensitive when test content differs from the training distribution; RD performance and generalization can suffer. DCVC-DT addresses this on top of the DCVC-DC backbone without retraining.

  • Online latent refinement. Encoder and decoder weights stay fixed; only the current-frame latent is adapted at inference with lightweight gradient steps and stochastic Gumbel–annealing style quantization. Decoding cost is unchanged, while the encoder better fits the actual content.

  • Frame-level dynamic RD adjustment. The objective balances rate and distortion; we rescale the weight on the bitrate term from frame to frame using inter-frame PSNR fluctuation, so bitrate is shifted where quality drops and overall RD improves.

The overall framework is illustrated below.

Overall framework of DCVC-DT

The online refinement loop and the dynamic RD adjustment interact as sketched here.

Online latent refinement and frame-level dynamic RD adjustment

Results

We follow DCVC-DC’s protocol on HEVC Class C and D under PSNR (BT.709). With DCVC-DC as anchor, DCVC-DT achieves about 6.21% average BD-rate reduction over the eight sequences, and remains ahead of DCVC-TCM and DCVC-HEM in the same comparison.

BD-rate (%) vs. DCVC-DC (negative is better):

Method BQMall BasketballDrill PartyScene RaceHorses BasketballPass BlowingBubbles BQSquare RaceHorses Avg
DCVC-TCM 103.99 101.37 107.23 77.11 82.13 81.95 140.53 66.42 95.09
DCVC-HEM 38.85 54.14 49.09 30.81 38.84 40.71 63.29 33.70 43.68
DCVC-DC 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00
DCVC-DT (Ours) -7.95 -5.49 -3.22 -8.72 -5.64 -5.48 -5.91 -7.25 -6.21

R–D curves (same setting):

R–D performance on HEVC Class C and D (PSNR, BT.709)

Where error propagation is visible, reconstruction and bitrate use tend to be smoother across frames. On BasketballPass (Class D, first 32 frames at the highest rate), PSNR and BPP evolve more favorably than the baseline.

PSNR and BPP fluctuation on BasketballPass (HEVC Class D)

AVS-EEM Integration Performance:

AVS-EEM (AVS End-to-End Intelligent Video Coding Exploration Model) is a standardization project launched by the AVS video coding working group, aimed at designing and exploring deployable end-to-end intelligent video coding systems under strict computational complexity constraints.

Our online learning enhancement technique has also been proposed and evaluated as an optional encoder-side configuration within the AVS-EEM platform. By integrating Stochastic Gumbel Annealing (SGA-Q), optimizing the YUV Loss ratio to 1:1:1 to prevent chroma degradation, and utilizing frame-level dynamic RD adjustments, the encoder adapts effectively to test sequence characteristics. Operating strictly as a pure encoder-side optimization, this approach guarantees zero additional overhead to the decoding time. When benchmarked against the EEM v8.1 platform, our method achieves comprehensive rate-distortion performance gains of -3.36% (Y), -7.17% (U), and -6.68% (V).

AVS-EEM Integration Performance

Pretrained models

Identical to DCVC-DC: download the official pretrained models into ./checkpoints, or use the download script in that folder.

Test the models

Before running the evaluation, please ensure that you have completed the environment setup according to the Prerequisites section of DCVC-DC and organized your test sequences in RGB format (PNG files) as described in their Test dataset guidelines.

To test the pretrained model with four rate points (using RGB data organization), first copy the test_video_y_latent_optimized.py file into the DCVC-DC project root directory, then run the following command:

python test_video_y_latent_optimized.py --i_frame_model_path ./checkpoints/cvpr2023_image_psnr.pth.tar --p_frame_model_path ./checkpoints/cvpr2023_video_psnr.pth.tar --rate_num 4 --test_config ./dataset_config_example_rgb.json --yuv420 0 --cuda 1 --cuda_device 0,1,2,3 --worker 4 --write_stream 1 --stream_path ./stream --save_decoded_frame 1 --decoded_frame_path ./decoded_frames --output_path output.json --force_intra_period 32 --force_frame_num 96

Acknowledgement

The implementation is based on DCVC-DC (Microsoft), CompressAI, and PyTorchVideoCompression.

Citation

If you find this work useful for your research, please cite our ISCAS paper (or arXiv) and the DCVC-DC reference:

@inproceedings{li2023neural,
  title     = {Neural Video Compression with Diverse Contexts},
  author    = {Li, Jiahao and Li, Bin and Lu, Yan},
  booktitle = {{IEEE/CVF} Conference on Computer Vision and Pattern Recognition,
               {CVPR} 2023, Vancouver, Canada, June 18--22, 2023},
  year      = {2023}
}

@misc{zhang2026neuralvideocompressiondomain,
      title={Neural Video Compression with Domain Transfer}, 
      author={Tiange Zhang and Rongqun Lin and Xiandong Meng and Haofeng Wang and Xing Tian and Qi Zhang and Siwei Ma},
      year={2026},
      eprint={2605.13476},
      archivePrefix={arXiv},
      primaryClass={cs.CV},
      url={https://arxiv.org/abs/2605.13476}, 
}

About

Codes for the ISCAS 2026 paper "Neural Video Compression with Domain Transfer"

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors