Ting Yu, Cha Zhang, Michael F. Cohen, Y. Rui, Ying Wu
{"title":"基于空间-颜色高斯混合模型的单眼视频前景/背景分割","authors":"Ting Yu, Cha Zhang, Michael F. Cohen, Y. Rui, Ying Wu","doi":"10.1109/WMVC.2007.27","DOIUrl":null,"url":null,"abstract":"This paper presents a new approach to segmenting monocular videos captured by static or hand-held cameras filming large moving non-rigid foreground objects. The foreground and background objects are modeled using spatialcolor Gaussian mixture models (SCGMM), and segmented using the graph cut algorithm, which minimizes a Markov random field energy function containing the SCGMM models. In view of the existence of a modeling gap between the available SCGMMs and segmentation task of a new frame, one major contribution of our paper is the introduction of a novel foreground/background SCGMM joint tracking algorithm to bridge this space, which greatly improves the segmentation performance in case of complex or rapid motion. Specifically, we propose to combine the two SCGMMs into a generative model of the whole image, and maximize the joint data likelihood using a constrained Expectation- Maximization (EM) algorithm. The effectiveness of the proposed algorithm is demonstrated on a variety of sequences.","PeriodicalId":177842,"journal":{"name":"2007 IEEE Workshop on Motion and Video Computing (WMVC'07)","volume":"9 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2007-02-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"78","resultStr":"{\"title\":\"Monocular Video Foreground/Background Segmentation by Tracking Spatial-Color Gaussian Mixture Models\",\"authors\":\"Ting Yu, Cha Zhang, Michael F. Cohen, Y. Rui, Ying Wu\",\"doi\":\"10.1109/WMVC.2007.27\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"This paper presents a new approach to segmenting monocular videos captured by static or hand-held cameras filming large moving non-rigid foreground objects. The foreground and background objects are modeled using spatialcolor Gaussian mixture models (SCGMM), and segmented using the graph cut algorithm, which minimizes a Markov random field energy function containing the SCGMM models. In view of the existence of a modeling gap between the available SCGMMs and segmentation task of a new frame, one major contribution of our paper is the introduction of a novel foreground/background SCGMM joint tracking algorithm to bridge this space, which greatly improves the segmentation performance in case of complex or rapid motion. Specifically, we propose to combine the two SCGMMs into a generative model of the whole image, and maximize the joint data likelihood using a constrained Expectation- Maximization (EM) algorithm. The effectiveness of the proposed algorithm is demonstrated on a variety of sequences.\",\"PeriodicalId\":177842,\"journal\":{\"name\":\"2007 IEEE Workshop on Motion and Video Computing (WMVC'07)\",\"volume\":\"9 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2007-02-23\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"78\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2007 IEEE Workshop on Motion and Video Computing (WMVC'07)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/WMVC.2007.27\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2007 IEEE Workshop on Motion and Video Computing (WMVC'07)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/WMVC.2007.27","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Monocular Video Foreground/Background Segmentation by Tracking Spatial-Color Gaussian Mixture Models
This paper presents a new approach to segmenting monocular videos captured by static or hand-held cameras filming large moving non-rigid foreground objects. The foreground and background objects are modeled using spatialcolor Gaussian mixture models (SCGMM), and segmented using the graph cut algorithm, which minimizes a Markov random field energy function containing the SCGMM models. In view of the existence of a modeling gap between the available SCGMMs and segmentation task of a new frame, one major contribution of our paper is the introduction of a novel foreground/background SCGMM joint tracking algorithm to bridge this space, which greatly improves the segmentation performance in case of complex or rapid motion. Specifically, we propose to combine the two SCGMMs into a generative model of the whole image, and maximize the joint data likelihood using a constrained Expectation- Maximization (EM) algorithm. The effectiveness of the proposed algorithm is demonstrated on a variety of sequences.