Pedestrian detection is a particular issue in both academia and industry. However, most existing pedestrian detection methods usually fail to detect small-scale pedestrians due to the introduction of feeble contrast and motion blur in images and videos. In this paper, we propose a multi-level feature fusion strategy to detect multi-scale pedestrians, which works particularly well with small-scale pedestrians that are relatively far from the camera. We propose a multi-level feature fusion strategy to make the shallow feature maps encode more semantic and global information to detect small-scale pedestrians. In addition, we redesign the aspect ratio of anchors to make it more robust for pedestrian detection task. The extensive experiments on both Caltech and CityPersons datasets demonstrate that our method outperforms the state-of-the-art pedestrian detection algorithms. Our proposed approach achieves a MR−2 of 0.84%, 23.91% and 62.19% under the “Near”, Medium” and “Far” settings respectively on Caltech dataset, and also leads a better speed-accuracy trade-off with 0.28 second per image of 1024×2048 pixel compared with others on CityPersons dataset.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
INSTITUTIONAL Select your institution to access the SPIE Digital Library.
PERSONAL Sign in with your SPIE account to access your personal subscriptions or to use specific features such as save to my library, sign up for alerts, save searches, etc.