Neural Network-Based Video Compression Artifact Reduction Using Temporal Correlation and Sparsity Prior Predictions
| dc.contributor.author | Chen, Wei-Gang | |
| dc.contributor.author | Yu, Runyi | |
| dc.contributor.author | Wang, Xun | |
| dc.date.accessioned | 2026-02-06T18:49:38Z | |
| dc.date.issued | 2020 | |
| dc.department | Doğu Akdeniz Üniversitesi | |
| dc.description.abstract | Quantization in lossy video compression may incur severe quality degradation, especially at low bit-rates. Developing post-processing methods that improve visual quality of decoded images is of great importance, as they can be directly incorporated in any existing compression standard or paradigm. We propose in this article a two-stage method, a texture detail restoration stage followed by a deep convolutional neural network (CNN) fusion stage, for video compression artifact reduction. The first stage performs in a patch-by-patch manner. For each patch in the current decoded frame, one prediction is formed based on the sparsity prior assuming that natural image patches can be represented by sparse activation of dictionary atoms. Under the temporal correlation hypothesis, we search the best matching patch in each reference frame, and select several matches with more texture details to tile motion compensated predictions. The second stage stacks the predictions obtained in the preceding stage along with the decoded frame itself to form a tensor, and proposes a deep CNN to learn the mapping between the tensor as input and the original uncompressed image as output. Experimental results demonstrate that the proposed two-stage method can remarkably improve, both subjectively and objectively, the quality of the compressed video sequence. | |
| dc.description.sponsorship | National Natural Science Foundation of China [61672460]; Public Welfare Technology Research Project of Zhejiang Province [LGG20F020005]; Science and Technology Program of Zhejiang Province (Key Research and Development Plan) [2020C01049] | |
| dc.description.sponsorship | This work was supported in part by the National Natural Science Foundation of China under Grant 61672460, in part by the Public Welfare Technology Research Project of Zhejiang Province under Grant LGG20F020005, and in part by the Science and Technology Program of Zhejiang Province (Key Research and Development Plan) under Grant 2020C01049. | |
| dc.identifier.doi | 10.1109/ACCESS.2020.3020388 | |
| dc.identifier.endpage | 162490 | |
| dc.identifier.issn | 2169-3536 | |
| dc.identifier.scopus | 2-s2.0-85102882116 | |
| dc.identifier.scopusquality | Q1 | |
| dc.identifier.startpage | 162479 | |
| dc.identifier.uri | https://doi.org/10.1109/ACCESS.2020.3020388 | |
| dc.identifier.uri | https://hdl.handle.net/11129/14972 | |
| dc.identifier.volume | 8 | |
| dc.identifier.wos | WOS:000572885700001 | |
| dc.identifier.wosquality | Q2 | |
| dc.indekslendigikaynak | Web of Science | |
| dc.indekslendigikaynak | Scopus | |
| dc.language.iso | en | |
| dc.publisher | IEEE-Inst Electrical Electronics Engineers Inc | |
| dc.relation.ispartof | Ieee Access | |
| dc.relation.publicationcategory | Makale - Uluslararası Hakemli Dergi - Kurum Öğretim Elemanı | |
| dc.rights | info:eu-repo/semantics/openAccess | |
| dc.snmz | KA_WoS_20260204 | |
| dc.subject | Image coding | |
| dc.subject | Image restoration | |
| dc.subject | Dictionaries | |
| dc.subject | Correlation | |
| dc.subject | Video compression | |
| dc.subject | Quantization (signal) | |
| dc.subject | Discrete cosine transforms | |
| dc.subject | Compression artifact reduction | |
| dc.subject | convolutional neural networks | |
| dc.subject | high efficiency video coding | |
| dc.subject | sparse representation | |
| dc.subject | temporal correlation | |
| dc.title | Neural Network-Based Video Compression Artifact Reduction Using Temporal Correlation and Sparsity Prior Predictions | |
| dc.type | Article |










