In most video platforms, such as Youtube, Kwai, and Tik-Tok, the played videos usually have undergone multiple video encodings such as hardware encoding by recording devices, software encoding by video editing apps, and single/multiple video transcoding by video application servers. Previous works in compressed video restoration typically assume the compression artifacts are caused by one-time encoding. Thus, the derived solution usually does not work very well in practice. In this paper, we propose a new method, temporal spatial auxiliary network (TSAN), for transcoded video restoration. Our method considers the unique traits between video encoding and transcoding, and we consider the initial shallow encoded videos as the intermediate labels to assist the network to conduct self-supervised attention training. In addition, we employ adjacent multi-frame information and propose the temporal deformable alignment and pyramidal spatial fusion for transcoded video restoration. The experimental results demonstrate that the performance of the proposed method is superior to that of the previous techniques.
- Ubuntu 16.04
- CUDA 10.1
- PyTorch 1.6.0+cu101
- Packages: tqdm, lmdb, pyyaml, opencv-python, scikit-image
Build DCNv2.
$ cd ops/dcn/
$ bash build.sh
(Optional) Check if DCNv2 works.
$ python simple_check.py
The DCNv2 source files here is different from the open-sourced version due to incompatibility.[issue]
Test data and pre-trained models: [Google Drive]
$ python test.py