Calibrating zooming cameras from multiple images without feature extraction

W. Brent Seales, David Eggert

Research output: Contribution to journalConference articlepeer-review


This work presents a new approach to the problem of calibrating a zoomable camera. The calibration of zooming cameras is central for tasks which employ zoom to improve feature detection and correspondence, such as 3-D stereo reconstruction. Our method solves for the parameters of a camera model using a global optimization technique on a sequence of images of a known calibration target obtained for different mechanical zoom settings. This approach addresses two primary weaknesses in classical camera calibration. First, the process avoids the difficulties of explicit feature detection. Feature localization is instead included as part of the error measure used in the optimization. Second, images are not calibrated independently, as in previous efforts. Rather, the optimization process considers all images simultaneously, representing the final calibrated camera as a function of zoom. We compute a starting point for the optimization using the measured mechanical zoom settings for the images, and certain features identified by either a high-level process or a human operator. This paper describes the details of our approach, showing initial experimental results on real data.

Original languageEnglish
Pages (from-to)274-282
Number of pages9
JournalProceedings of SPIE - The International Society for Optical Engineering
StatePublished - Oct 6 1994
EventSensor Fusion VII 1994 - Boston, United States
Duration: Oct 31 1994Nov 4 1994

Bibliographical note

Publisher Copyright:
© 1994 SPIE. All rights reserved.


  • Active camera
  • Active vision
  • Calibration
  • Zoom

ASJC Scopus subject areas

  • Electronic, Optical and Magnetic Materials
  • Condensed Matter Physics
  • Computer Science Applications
  • Applied Mathematics
  • Electrical and Electronic Engineering


Dive into the research topics of 'Calibrating zooming cameras from multiple images without feature extraction'. Together they form a unique fingerprint.

Cite this