Learning Efficient Single-stage Pedestrian Detectors by Asymptotic Localization Fitting
Wei Liu, Shengcai Liao, Weidong Hu, Xuezhi Liang, Xiao Chen
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/VideoObjectSearch/ALFNetOfficialIn papertf★ 0
Abstract
Though Faster R-CNN based two-stage detectors have witnessed significant boost in pedestrian detection accuracy, it is still slow for practical applications. One solution is to simplify this working flow as a single-stage detector. However, current single-stage detectors (e.g. SSD) have not presented competitive accuracy on common pedestrian detection benchmarks. This paper is towards a successful pedestrian detector enjoying the speed of SSD while maintaining the accuracy of Faster R-CNN. Specifically, a structurally simple but effective module called emphAsymptotic Localization Fitting (ALF) is proposed, which stacks a series of predictors to directly evolve the default anchor boxes of SSD step by step into improving detection results. As a result, during training the latter predictors enjoy more and better-quality positive samples, meanwhile harder negatives could be mined with increasing IoU thresholds. On top of this, an efficient single-stage pedestrian detection architecture (denoted as ALFNet) is designed, achieving state-of-the-art performance on CityPersons and Caltech, two of the largest pedestrian detection benchmarks, and hence resulting in an attractive pedestrian detector in both accuracy and speed. Code is available at hrefhttps://github.com/VideoObjectSearch/ALFNethttps://github.com/VideoObjectSearch/ALFNet.