This repository contains the offical implementations along with the experimental splits for the paper "Learning Spatiotemporal Interactions for User-Generated Video Quality Assessment", in IEEE Transactions on Circuits and Systems for Video Technology, vol. 33, no. 3, pp. 1031-1042, Mar. 2023. Hanwei Zhu, Baoliang Chen, Lingyu Zhu, and Shiqi Wang.
The release codes were implemented and have been tested in Ubuntu 18.04 with
- Python = 3.6.13
- PyTorch = 1.8.1
- torchvision = 0.9.0
More details can be found in README.md in the folder of extract_features
.
You can change the paramers in param.py
to train each dataset with intra-/cross-dataset settings:
python main.py --test_only False
You can change the paramers in param.py
to train and test each dataset, and the trained parameters of the proposed model on each dataset can be found at Google Drive:
python main.py --test_only True
The authors would like to thank Dingquan Li for his implementation of VSFA, Yang Li for his code architecture, the BVQA_Benchmark, and the implementation of ViT.
@article{zhu2022learing,
title={Learning Spatiotemporal Interactions for User-Generated Video Quality Assessment},
author={Zhu, Hanwei and Chen, Baoliang and Zhu, lingyu and Wang, Shiqi},
journal={IEEE Transactions on Circuits and Systems for Video Technology},
volume={33},
number={3},
pages={1031-1042},
month={Mar.},
year={2023}
}