{"title":"视频压缩中运动预测的深度多帧增强","authors":"N. Prette, D. Valsesia, T. Bianchi","doi":"10.1109/icecs53924.2021.9665523","DOIUrl":null,"url":null,"abstract":"This work proposes a novel Deep Learning technique to increase the efficiency of currently available video compression techniques based on motion compensation. The goal is to improve the frame prediction task, whereby a more accurate prediction of the motion from the reference frames to the target frame allows to reduce the rate needed to encode the residual. This is achieved by means of a convolutional neural network (CNN) architecture that processes the basic block-based motion-compensated prediction of the current frame as well as predictions from past reference frames. This method allows to reduce typical artifacts such as blockiness, and achieves a more accurate prediction of motion thanks to the representation capabilities of CNNs, leading to smaller prediction residuals. Preliminary results show that the proposed approach is capable of providing BD-rate gains up to 6%.","PeriodicalId":448558,"journal":{"name":"2021 28th IEEE International Conference on Electronics, Circuits, and Systems (ICECS)","volume":"75 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-11-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Deep Multiframe Enhancement for Motion Prediction in Video Compression\",\"authors\":\"N. Prette, D. Valsesia, T. Bianchi\",\"doi\":\"10.1109/icecs53924.2021.9665523\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"This work proposes a novel Deep Learning technique to increase the efficiency of currently available video compression techniques based on motion compensation. The goal is to improve the frame prediction task, whereby a more accurate prediction of the motion from the reference frames to the target frame allows to reduce the rate needed to encode the residual. This is achieved by means of a convolutional neural network (CNN) architecture that processes the basic block-based motion-compensated prediction of the current frame as well as predictions from past reference frames. This method allows to reduce typical artifacts such as blockiness, and achieves a more accurate prediction of motion thanks to the representation capabilities of CNNs, leading to smaller prediction residuals. Preliminary results show that the proposed approach is capable of providing BD-rate gains up to 6%.\",\"PeriodicalId\":448558,\"journal\":{\"name\":\"2021 28th IEEE International Conference on Electronics, Circuits, and Systems (ICECS)\",\"volume\":\"75 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-11-28\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2021 28th IEEE International Conference on Electronics, Circuits, and Systems (ICECS)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/icecs53924.2021.9665523\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 28th IEEE International Conference on Electronics, Circuits, and Systems (ICECS)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/icecs53924.2021.9665523","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Deep Multiframe Enhancement for Motion Prediction in Video Compression
This work proposes a novel Deep Learning technique to increase the efficiency of currently available video compression techniques based on motion compensation. The goal is to improve the frame prediction task, whereby a more accurate prediction of the motion from the reference frames to the target frame allows to reduce the rate needed to encode the residual. This is achieved by means of a convolutional neural network (CNN) architecture that processes the basic block-based motion-compensated prediction of the current frame as well as predictions from past reference frames. This method allows to reduce typical artifacts such as blockiness, and achieves a more accurate prediction of motion thanks to the representation capabilities of CNNs, leading to smaller prediction residuals. Preliminary results show that the proposed approach is capable of providing BD-rate gains up to 6%.