Robust autocalibration for a surveillance camera network

We propose a novel approach for multi-camera autocalibration by observing multiview surveillance video of pedestrians walking through the scene. Unlike existing methods, we do NOT require tracking or explicit correspondences of the same person across time/views. Instead, we take noisy foreground blo...

Full description

Saved in:
Bibliographic Details
Published in2013 IEEE Workshop on Applications of Computer Vision (WACV) pp. 433 - 440
Main Authors Jingchen Liu, Collins, R. T., Yanxi Liu
Format Conference Proceeding Journal Article
LanguageEnglish
Published IEEE 01.01.2013
Subjects
Online AccessGet full text
ISBN9781467350532
1467350532
ISSN1550-5790
1550-5790
DOI10.1109/WACV.2013.6475051

Cover

More Information
Summary:We propose a novel approach for multi-camera autocalibration by observing multiview surveillance video of pedestrians walking through the scene. Unlike existing methods, we do NOT require tracking or explicit correspondences of the same person across time/views. Instead, we take noisy foreground blobs as the only input and rely on a joint optimization framework with robust statistics to achieve accurate calibration under challenging scenarios. First, each individual camera is roughly calibrated into its local World Coordinate System (lWCS) based on analysis of relative 3D pedestrian height distribution. Then, all lWCSs are iteratively registered with respect to a shared global World Coordinate System (gWCS) by incorporating robust matching with a partial Direct Linear Transform (pDLT). As demonstrated by extensive evaluation, our algorithm achieves satisfactory results in various camera settings with up to moderate crowd densities with a large proportion of foreground outliers.
Bibliography:ObjectType-Article-2
SourceType-Scholarly Journals-1
ObjectType-Conference-1
ObjectType-Feature-3
content type line 23
SourceType-Conference Papers & Proceedings-2
ISBN:9781467350532
1467350532
ISSN:1550-5790
1550-5790
DOI:10.1109/WACV.2013.6475051