Godwin Shen, Woo-Shik Kim, S. K. Narang, Antonio Ortega, Jaejoon Lee, H. Wey
{"title":"Edge-adaptive transforms for efficient depth map coding","authors":"Godwin Shen, Woo-Shik Kim, S. K. Narang, Antonio Ortega, Jaejoon Lee, H. Wey","doi":"10.1109/PCS.2010.5702565","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702565","url":null,"abstract":"In this work a new set of edge-adaptive transforms (EATs) is presented as an alternative to the standard DCTs used in image and video coding applications. These transforms avoid filtering across edges in each image block, thus, they avoid creating large high frequency coefficients. These transforms are then combined with the DCT in H.264/AVC and a transform mode selection algorithm is used to choose between DCT and EAT in an RD-optimized manner. These transforms are applied to coding depth maps used for view synthesis in a multi-view video coding system, and provides up to 29% bit rate reduction for a fixed quality in the synthesized views.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126663786","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
T. Misu, Y. Matsuo, S. Sakaida, Y. Shishikui, E. Nakasu
{"title":"Novel video coding paradigm with reduction/restoration processes","authors":"T. Misu, Y. Matsuo, S. Sakaida, Y. Shishikui, E. Nakasu","doi":"10.1109/PCS.2010.5702537","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702537","url":null,"abstract":"To optimally design distortions in lossy video coding, we propose the use of a novel coding paradigm with adaptive nonlinear transforms as pre/post-processors of a conventional video codec. The preprocessor decimates less important pixels based on an image analysis. A conventional video encoder such as MPEG-4 AVC/H.264 further eliminates the redundancy of the decimated images. On the decoder side, the postprocessor restores small decoded images of the conventional decoder to the original resolution using an inverse mapping including a super-resolution technique that uses a priori knowledge on the decimation in the preprocessing. Results of the experiments showed the proposed coding scheme poses distortion that has a more straightforward appearance than that of a directly encoded/decoded image by a sole conventional H.264 codec.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"40 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126754159","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A novel inloop filter for video-compression based on temporal pixel trajectories","authors":"Marko Esche, A. Krutz, A. Glantz, T. Sikora","doi":"10.1109/PCS.2010.5702551","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702551","url":null,"abstract":"The objective of this work is to investigate the performance of a new inloop filter for video compression, which uses temporal rather than spatial information to improve the quality of reference frames used for prediction. The new filter has been integrated into the H.264/AVC baseline encoder and tested on a wide range of sequences. Experimental results show that the filter achieves a bit rate reduction of up to 12% and more than 4% on average without increasing the complexity of either encoder or decoder significantly.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"148 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115435870","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Takashi Ishibashi, T. Yendo, M. P. Tehrani, T. Fujii, M. Tanimoto
{"title":"3D space representation using epipolar plane depth image","authors":"Takashi Ishibashi, T. Yendo, M. P. Tehrani, T. Fujii, M. Tanimoto","doi":"10.1109/PCS.2010.5702472","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702472","url":null,"abstract":"We propose a novel 3D space representation for multi-view video, using epipolar plane depth images (EPDI). Multi-view video plus depth (MVD) is used as common data format for FTV(Free-viewpoint TV), which enables synthesizing virtual view images. Due to large amount of data and complexity of the multi-view video coding (MVC), compression of MVD is a challenging issue. We address this problem and propose a new representation that is constructed from MVD using rayspace. MVD is converted into image and depth ray-spaces. The proposed representation is obtained by converting each of ray-spaces into a global depth map and a texture map using EPDI. Experiments demonstrate the analysis of this representation, and its efficiency.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125948160","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
D. Doshkov, P. Ndjiki-Nya, H. Lakshman, Martin Köppel, T. Wiegand
{"title":"Towards efficient intra prediction based on image inpainting methods","authors":"D. Doshkov, P. Ndjiki-Nya, H. Lakshman, Martin Köppel, T. Wiegand","doi":"10.1109/PCS.2010.5702539","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702539","url":null,"abstract":"In this paper, novel intra prediction methods based on image inpainting approaches are proposed. The H.264/AVC intra prediction modes are not well suited for processing complex textures at low bit rates. Our algorithm utilizes an efficient combination of partial differential equations (PDEs) and patch-based texture synthesis in addition to the standard directional predictors. Bit rate savings up to 3.5% compared to that of the H.264/AVC standard are shown.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121124527","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
S. Muramatsu, Dandan Han, Tomoya Kobayashi, H. Kikuchi
{"title":"Theoretical analysis of trend vanishing moments for directional orthogonal transforms","authors":"S. Muramatsu, Dandan Han, Tomoya Kobayashi, H. Kikuchi","doi":"10.1109/PCS.2010.5702440","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702440","url":null,"abstract":"This work contributes to investigate theoretical properties of the trend vanishing moment (TVM) which the authors have defined in a previous work and applied to the directional design of 2-D nonsep-arable GenLOTs. The TVM condition can be regarded as a natural extention of 1-D VM to 2-D one and alternative of the conventional directional vanishing moment (DVM). Firstly, the definition of TVM is given to clarify what it is, and then the mathematical meaning is discussed. While the conventional DVM condition requires for the moments to vanish along lines in the frequency domain and restricts the direction to a rational factor, the TVM condition imposes the moments only point-wisely and the direction can be steered flexibly. Some significant properties of TVMs are also shown and a simulation result of zonal coding for an artifical picture is given to verify the trend surface annihilation property.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125992935","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A novel coding scheme for intra pictures of H.264/AVC","authors":"Jin Young Lee, Jaejoon Lee, H. Wey, Du-sik Park","doi":"10.1109/PCS.2010.5702579","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702579","url":null,"abstract":"A novel intra coding scheme is proposed to improve coding performance in intra pictures of H.264/AVC. The proposed method generates two sub-images, which are defined as a sampled image and a prediction error image in this paper, from an original image, and then encodes them separately. Especially, in an intra prediction process of encoding, the sampled image employs the original intra prediction modes, while the prediction error image uses newly defined four intra prediction modes. Experimental results demonstrate that the proposed method achieves significantly higher intra coding performance and reduces encoding complexity with the smaller number of a rate-distortion (RD) cost calculation process, as compared with the original intra coding method of H.264/AVC.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"77 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134254449","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Fractional-sample motion compensation using generalized interpolation","authors":"H. Lakshman, B. Bross, H. Schwarz, T. Wiegand","doi":"10.1109/PCS.2010.5702555","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702555","url":null,"abstract":"Typical interpolation methods in video coding perform filtering of reference picture samples using FIR filters for motion-compensated prediction. This process can be viewed as a signal decomposition using basis functions which are restricted by the interpolating constraint. Using the concept of generalized interpolation provides a greater degree of freedom for selecting basis functions. We implemented generalized interpolation using a combination of IIR and FIR filters. The complexity of the proposed scheme is comparable to that of an 8-tap FIR filter. Bit rate savings up to 20% compared to the H.264/AVC 6-tap filter are shown.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"35 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133091209","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
T. Miyachi, Keita Takahashi, Madoka Hasegawa, Yuichi Tanaka, Shigeo Kato
{"title":"A study on memorability and shoulder-surfing robustness of graphical password using DWT-based image blending","authors":"T. Miyachi, Keita Takahashi, Madoka Hasegawa, Yuichi Tanaka, Shigeo Kato","doi":"10.1109/PCS.2010.5702441","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702441","url":null,"abstract":"Graphical passwords are an authentication method that uses pictures as passwords instead of using alphanumeric characters. We propose a graphical password method which is difficult to steal original pass-image by using characteristics of human vision system. In our method, we combine low frequency components of a decoy picture with high frequency components of a pass-image. It is easy for legitimate users to recognize the pass-image in the blended image. On the other hand, this task is difficult for attackers. We used discrete wavelet transform (DWT) to blend a decoy image and a pass-image. User studies are conducted to evaluate memorability and shoulder-surfing robustness of this method. We also compared our method with other existing methods in terms of the authentication time and the success ratio by the user test. The results show that our method is more usable and secure against shoulder-surfing.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133893504","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A background model based method for transcoding surveillance videos captured by stationary camera","authors":"Xianguo Zhang, Luhong Liang, Qian Huang, Tiejun Huang, Wen Gao","doi":"10.1109/PCS.2010.5702583","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702583","url":null,"abstract":"Real-world video surveillance applications require storing videos without neglecting any part of scenarios for weeks or months. To reduce the storage cost, the high bit-rate videos from cameras should be transcoded into a more efficient compressed format with as little quality loss as possible. In this paper, we propose a background model based method to improve the transcoding efficiency for surveillance videos captured by stationary cameras, and objectively measure it. The background model is trained by pre-decoded I frames, and then used to transcode the source stream. Following this method, an H.264/AVC based transcoder employing the background model as long-term reference frame and a difference frame coding based transcoder are implemented and evaluated. Experimental results show that both trancoders save nearly half the used bits while maintaining quality compared with the full-decoding-full-encoding method, and the latter one has slightly better performance.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129519781","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}