Robust Video Segment Proposals with Painless Occlusion Handling
Abstract
We propose a robust algorithm to generate video segment proposals. The proposals generated by our method can start from any frame in the video and are robust to complete occlusions. Our method does not assume specific motion models and even has a limited capability to generalize across videos. We build on our previous least squares tracking framework, where image segment proposals are generated and tracked using learned appearance models. The innovation in our new method lies in the use of two efficient moves, the merge move and free addition, to efficiently start segments from any frame and track them through complete occlusions, without much additional computation. Segment size interpolation is used for effectively detecting occlusions. We propose a new metric for evaluating video segment proposals on the challenging VSB-100 benchmark and present state-of-the-art results. Preliminary results are also shown for the potential use of our framework to track segments across different videos.
Paper
Code
Code will be released in the future! Bookmark for future visit!
Video Comparison

Salsa

Hockey_Goals

Ballet

White_Tiger
Results
Authors
Zhengyang Wu
Georgia Tech
Fuxin Li
Georgia Tech
Rahul Sukthankar
Google
James M. Rehg
Georgia Tech
Citation
@inproceedings{ZWuCVPR2015,
  author = {Zhengyang Wu and Fuxin Li and Rahul Sukthankar and James M. Rehg},
  title = { Video Segment Proposals with Painless Occlusion Handling},
  booktitle = {CVPR},
  year = {2015} }
Acknowledgements
This research is partially supported by NSF IIS-1320348 and ARO MURI W911NF-11-1-0046.