Functional calibration for pan-tilt-zoom cameras in hybrid sensor networks

被引:6
|
作者
Wren, Christopher R.
Erdem, Ugur Murat
Azarbayejani, Ali J.
机构
[1] MERL Res, Cambridge, MA 02139 USA
[2] Boston Univ, Boston, MA 02215 USA
[3] MERL Technol, Cambridge, MA USA
关键词
sensor networks; video surveillance; adaptive systems;
D O I
10.1007/s00530-006-0061-x
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Wide-area context awareness is a crucial enabling technology for next generation smart buildings and surveillance systems. It is not practical to gather this context awareness by covering the entire building with cameras. However, significant gaps in coverage caused by installing cameras in a sparse way can make it very difficult to infer the missing information. As a solution we advocate a class of hybrid perceptual systems that build a comprehensive model of activity in a large space, such as a building, by merging contextual information from a dense network of ultra-lightweight sensor nodes and video from a sparse network of cameras. In this paper we explore the task of automatically recovering the relative geometry between a pan-tilt-zoom camera and a network of one-bit motion detectors. We present results both for the recovery of geometry alone and also for the recovery of geometry jointly with simple activity models. Because we do not believe a metric calibration is necessary, or even entirely useful, for this task, we formulate and pursue the novel goal we term functional calibration. Functional calibration is a blending of geometry estimation and simple behavioral model discovery. Accordingly, results are evaluated by measuring the ability of the system to automatically foveate targets in a large, non-convex space, rather than by measuring, for example, pixel reconstruction error.
引用
收藏
页码:255 / 268
页数:14
相关论文
共 50 条
  • [1] Functional calibration for pan-tilt-zoom cameras in hybrid sensor networks
    Christopher R. Wren
    Ugur Murat Erdem
    Ali J. Azarbayejani
    [J]. Multimedia Systems, 2006, 12 : 255 - 268
  • [2] Dynamic calibration of pan-tilt-zoom cameras for traffic monitoring
    Song, Kai-Tai
    Tai, Jen-Chao
    [J]. IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART B-CYBERNETICS, 2006, 36 (05): : 1091 - 1103
  • [3] Coverage in Visual Sensor Networks with Pan-Tilt-Zoom Cameras: the MaxFoV Problem
    Munishwar, Vikram P.
    Kolar, Vinay
    Abu-Ghazaleh, Nael B.
    [J]. 2014 PROCEEDINGS IEEE INFOCOM, 2014, : 1492 - 1500
  • [4] Calibration of pan-tilt-zoom (PTZ) cameras and omni-directional cameras
    Thirthala, S
    Sinha, SN
    Pollefeys, M
    [J]. 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Vol 2, Proceedings, 2005, : 1198 - 1198
  • [5] Accurate and Practical Calibration of Multiple Pan-Tilt-Zoom Cameras for Live Broadcasts
    Kano, Masanori
    Okubo, Hidehiko
    Takahashi, Masaki
    Ikeya, Kensuke
    Hisatomi, Kensuke
    Mishina, Tomoyuki
    [J]. IEEE ACCESS, 2020, 8 (08): : 153993 - 154006
  • [6] Projection Invariants for Pan-Tilt-Zoom Robotic Cameras
    [J]. Springer Tracts in Advanced Robotics, 2009, 51 : 117 - 137
  • [7] Neural Background Subtraction for Pan-Tilt-Zoom Cameras
    Ferone, Alessio
    Maddalena, Lucia
    [J]. IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2014, 44 (05): : 571 - 579
  • [8] Modeling of Pan-tilt-zoom Cameras for Tracking Measurement
    Shi, Yan-qiong
    Lu, Rong-Sheng
    Chen, Ning
    Xia, Rui-xue
    [J]. 6TH INTERNATIONAL SYMPOSIUM ON PRECISION ENGINEERING MEASUREMENTS AND INSTRUMENTATION, 2010, 7544
  • [9] Linear Auto-calibration of Pan-Tilt-Zoom Cameras With Rotation Center Offset
    Liu, Yu
    Zhang, Hui
    [J]. 2023 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2023), 2023, : 11461 - 11467
  • [10] An adaptive zoom algorithm for tracking targets using pan-tilt-zoom cameras
    Shah, H
    Morrell, D
    [J]. 2004 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, VOL II, PROCEEDINGS: SENSOR ARRAY AND MULTICHANNEL SIGNAL PROCESSING SIGNAL PROCESSING THEORY AND METHODS, 2004, : 721 - 724