M. Becker, M. Baron, D. Kondermann, M. Bußler, V. Helzle
{"title":"Movie dimensionalization via sparse user annotations","authors":"M. Becker, M. Baron, D. Kondermann, M. Bußler, V. Helzle","doi":"10.1109/3DTV.2013.6676633","DOIUrl":null,"url":null,"abstract":"We present a workflow to semi-automatically create depth maps for monocular movie footage. Artists annotate relevant depth discontinuities in a single keyframe. Depth edges are then learned and predicted for the whole shot. We use structure from motion where possible for sparse depth cues, while the artist optionally provides scribbles to improve the intended visual effect. Finally, all three sources of information are combined via variational inpainting scheme. As the outcome of our method is artistic and cannot be evaluated quantitively, we apply our method to a current movie production, showing good results on different scenes. We further evaluate the depth edge localization compared to the “ground truth” provided by artists. To enable experimentation with our approach, we offer our source code.","PeriodicalId":111565,"journal":{"name":"2013 3DTV Vision Beyond Depth (3DTV-CON)","volume":"2135 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2013-12-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"4","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2013 3DTV Vision Beyond Depth (3DTV-CON)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/3DTV.2013.6676633","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 4
Abstract
We present a workflow to semi-automatically create depth maps for monocular movie footage. Artists annotate relevant depth discontinuities in a single keyframe. Depth edges are then learned and predicted for the whole shot. We use structure from motion where possible for sparse depth cues, while the artist optionally provides scribbles to improve the intended visual effect. Finally, all three sources of information are combined via variational inpainting scheme. As the outcome of our method is artistic and cannot be evaluated quantitively, we apply our method to a current movie production, showing good results on different scenes. We further evaluate the depth edge localization compared to the “ground truth” provided by artists. To enable experimentation with our approach, we offer our source code.