{"title":"Hand tracking and segmentation via graph cuts and dynamic model in sign language videos","authors":"Jun Wan, Q. Ruan, Gaoyun An, Wei Li","doi":"10.1109/ICOSP.2012.6491778","DOIUrl":null,"url":null,"abstract":"In this paper, we propose a new method for hands tracking and segmentation based on augmented graph cuts and dynamic model in sign language videos. We focus on resolving three problems which are fast hand motion capture, hand over face and hand occlusions. At first, an effective dynamic model for state prediction is used. This dynamic model can correctly predict the location of hand which has a rapid movement and quick shape deformation. Then, new energy terms are augmented into the energy function in graph cuts. The additional terms are inspired by multi cues, such as color, motion and spatial-temporal information. Finally, we construct the graph and achieve the hand segmentation in successive frames using min-cut/max-flow algorithm. We evaluate our algorithm in a real American Sign Language video from Purdue ASL Database. Besides, our method can be easily extended to track objects with similar color.","PeriodicalId":143331,"journal":{"name":"2012 IEEE 11th International Conference on Signal Processing","volume":"11 9 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2012-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"4","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2012 IEEE 11th International Conference on Signal Processing","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICOSP.2012.6491778","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 4
Abstract
In this paper, we propose a new method for hands tracking and segmentation based on augmented graph cuts and dynamic model in sign language videos. We focus on resolving three problems which are fast hand motion capture, hand over face and hand occlusions. At first, an effective dynamic model for state prediction is used. This dynamic model can correctly predict the location of hand which has a rapid movement and quick shape deformation. Then, new energy terms are augmented into the energy function in graph cuts. The additional terms are inspired by multi cues, such as color, motion and spatial-temporal information. Finally, we construct the graph and achieve the hand segmentation in successive frames using min-cut/max-flow algorithm. We evaluate our algorithm in a real American Sign Language video from Purdue ASL Database. Besides, our method can be easily extended to track objects with similar color.