A High-Resolution Panoramic Camera

Similar documents
Active Aperture Control and Sensor Modulation for Flexible Imaging

Single Camera Catadioptric Stereo System

Cameras for Stereo Panoramic Imaging Λ

Panoramic Mosaicing with a 180 Field of View Lens

Depth Perception with a Single Camera

Catadioptric Omnidirectional Camera *

Proc. of DARPA Image Understanding Workshop, New Orleans, May Omnidirectional Video Camera. Shree K. Nayar

Opto Engineering S.r.l.

Digital Photographic Imaging Using MOEMS

IMAGE FORMATION. Light source properties. Sensor characteristics Surface. Surface reflectance properties. Optics

Folded catadioptric panoramic lens with an equidistance projection scheme

UC Berkeley UC Berkeley Previously Published Works

Parity and Plane Mirrors. Invert Image flip about a horizontal line. Revert Image flip about a vertical line.

Novel Hemispheric Image Formation: Concepts & Applications

On Cosine-fourth and Vignetting Effects in Real Lenses*

Panoramic imaging. Ixyzϕθλt. 45 degrees FOV (normal view)

Unit 1: Image Formation

Waves & Oscillations

E X P E R I M E N T 12

Computer Vision. The Pinhole Camera Model

Capturing Omni-Directional Stereoscopic Spherical Projections with a Single Camera

This is an author-deposited version published in: Eprints ID: 3672

Chapters 1 & 2. Definitions and applications Conceptual basis of photogrammetric processing

Chapter Ray and Wave Optics

Catadioptric Stereo For Robot Localization

LENSLESS IMAGING BY COMPRESSIVE SENSING

Colorado School of Mines. Computer Vision. Professor William Hoff Dept of Electrical Engineering &Computer Science.

College of Arts and Sciences

Section 3. Imaging With A Thin Lens

Laboratory 7: Properties of Lenses and Mirrors

CS535 Fall Department of Computer Science Purdue University

Single-view Metrology and Cameras

Image Formation. Light from distant things. Geometrical optics. Pinhole camera. Chapter 36

Projection. Readings. Szeliski 2.1. Wednesday, October 23, 13

ISOMETRIC PROJECTION. Contents. Isometric Scale. Construction of Isometric Scale. Methods to draw isometric projections/isometric views

Image stitching. Image stitching. Video summarization. Applications of image stitching. Stitching = alignment + blending. geometrical registration

Phys 531 Lecture 9 30 September 2004 Ray Optics II. + 1 s i. = 1 f

Lenses. Overview. Terminology. The pinhole camera. Pinhole camera Lenses Principles of operation Limitations

Rectified Mosaicing: Mosaics without the Curl* Shmuel Peleg

Chapter 36. Image Formation

AR 2 kanoid: Augmented Reality ARkanoid

Geometry of Aerial Photographs

Projection. Projection. Image formation. Müller-Lyer Illusion. Readings. Readings. Let s design a camera. Szeliski 2.1. Szeliski 2.

UNIT 5a STANDARD ORTHOGRAPHIC VIEW DRAWINGS

PHYS 160 Astronomy. When analyzing light s behavior in a mirror or lens, it is helpful to use a technique called ray tracing.

Multiviews and Auxiliary Views

CH. 23 Mirrors and Lenses HW# 6, 7, 9, 11, 13, 21, 25, 31, 33, 35

Cameras. CSE 455, Winter 2010 January 25, 2010

Announcements. Image Formation: Outline. The course. How Cameras Produce Images. Earliest Surviving Photograph. Image Formation and Cameras

Image Formation: Camera Model

Geometric optics & aberrations

TSBB09 Image Sensors 2018-HT2. Image Formation Part 1

Sequential Algorithm for Robust Radiometric Calibration and Vignetting Correction

Lecture 2: Geometrical Optics. Geometrical Approximation. Lenses. Mirrors. Optical Systems. Images and Pupils. Aberrations.

Section 11. Vignetting

Compact camera module testing equipment with a conversion lens

Introduction. Strand F Unit 3: Optics. Learning Objectives. Introduction. At the end of this unit you should be able to;

Lens Design I. Lecture 3: Properties of optical systems II Herbert Gross. Summer term

AgilEye Manual Version 2.0 February 28, 2007

Perspective. Announcement: CS4450/5450. CS 4620 Lecture 3. Will be MW 8:40 9:55 How many can make the new time?

Digital deformation model for fisheye image rectification

Lens Design I. Lecture 3: Properties of optical systems II Herbert Gross. Summer term

Lecture 4: Geometrical Optics 2. Optical Systems. Images and Pupils. Rays. Wavefronts. Aberrations. Outline

Panoramas. CS 178, Spring Marc Levoy Computer Science Department Stanford University

Omni-Directional Catadioptric Acquisition System

Overview. Pinhole camera model Projective geometry Vanishing points and lines Projection matrix Cameras with Lenses Color Digital image

Lecture 2: Geometrical Optics. Geometrical Approximation. Lenses. Mirrors. Optical Systems. Images and Pupils. Aberrations.

Mapping cityscapes into cyberspace for visualization

Performance Factors. Technical Assistance. Fundamental Optics

Image Processing & Projective geometry

LENSES. INEL 6088 Computer Vision

OPTICAL SYSTEMS OBJECTIVES

[VR Lens Distortion] [Sangkwon Peter Jeong / JoyFun Inc.,]

(51) Int Cl.: H04N 1/00 ( ) H04N 13/00 ( ) G06T 3/40 ( )

Photographing Long Scenes with Multiviewpoint

Cross Sections of Three-Dimensional Figures

MULTIPLE SENSORS LENSLETS FOR SECURE DOCUMENT SCANNERS

ON THE CREATION OF PANORAMIC IMAGES FROM IMAGE SEQUENCES

The optical analysis of the proposed Schmidt camera design.

PHYSICS FOR THE IB DIPLOMA CAMBRIDGE UNIVERSITY PRESS

IMAGE SENSOR SOLUTIONS. KAC-96-1/5" Lens Kit. KODAK KAC-96-1/5" Lens Kit. for use with the KODAK CMOS Image Sensors. November 2004 Revision 2

PANORAMIC IMAGE ACQUISITION*

Colour correction for panoramic imaging

10.2 Images Formed by Lenses SUMMARY. Refraction in Lenses. Section 10.1 Questions

INTRODUCTION THIN LENSES. Introduction. given by the paraxial refraction equation derived last lecture: Thin lenses (19.1) = 1. Double-lens systems

History of projection. Perspective. History of projection. Plane projection in drawing

Projection. Announcements. Müller-Lyer Illusion. Image formation. Readings Nalwa 2.1

PRINCIPLE PROCEDURE ACTIVITY. AIM To observe diffraction of light due to a thin slit.

Astronomy 80 B: Light. Lecture 9: curved mirrors, lenses, aberrations 29 April 2003 Jerry Nelson

Lecture 3: Geometrical Optics 1. Spherical Waves. From Waves to Rays. Lenses. Chromatic Aberrations. Mirrors. Outline

Eric B. Burgh University of Wisconsin. 1. Scope

ENGINEERING DRAWING. UNIT III - Part A

Instruction Manual for HyperScan Spectrometer

Patents of eye tracking system- a survey

GEOMETRICAL OPTICS Practical 1. Part I. BASIC ELEMENTS AND METHODS FOR CHARACTERIZATION OF OPTICAL SYSTEMS

Dual-fisheye Lens Stitching for 360-degree Imaging & Video. Tuan Ho, PhD. Student Electrical Engineering Dept., UT Arlington

Folded Catadioptric Cameras*

Light: Lenses and. Mirrors. Test Date: Name 1ÿ-ÿ. Physics. Light: Lenses and Mirrors

CS 443: Imaging and Multimedia Cameras and Lenses

This experiment is under development and thus we appreciate any and all comments as we design an interesting and achievable set of goals.

Transcription:

A High-Resolution Panoramic Camera Hong Hua and Narendra Ahuja Beckman Institute, Department of Electrical and Computer Engineering2 University of Illinois at Urbana-Champaign, Urbana, IL, 61801 Email: Honghualn-ahuja@uiuc. edu Abstract Wide field of view (FOV) and high resolution are two desirable properties in many vision-based applications such as tele-conferencing, surveillance, and robot navigation. In some applications such as 30 reconstruction and rendering, it is also desired that all viewing directions share a single viewpoint, the entire FOV be imaged simultaneously, in real-time, and the depth offield be large. In this paper, we review such a panoramic camera proposed by NaIwa in 1996 that uses reflections off planar mirrors to achieve the jirst four of the uforementioned capabilities. He uses a single mirror pyramid (SW) and a number ofcameras that point to the individual pyramid faces. Together the cameras yield a visual,field having a width of 360 degrees and a height same as that of the individual cameras. We propose a double mirror-pyramid (DMP) design that still achieves a 360-degree FO V horizontally but doubles the vertical FOV. It retains the other three capabilities namely high resolution, a single apparent viewpoint across the entire FOV, and real-time panoramic capture. We specijl the visua1,field mapping @om the scene to the sensor realized by the proposed camera. Finally, an implementation of the proposed DMP design is described and examples of preliminary panoramic images obtained are included. 1. Introduction A sensor with a wide field of view (FOV) and high resolution is highly desirable in many applications such as tele-conferencing, surveillance, and robot navigation [ 11. In addition, a single viewpoint for all viewing directions, a large depth-of-field (omni-focus), and real-time acquisition are desired in some imaging applications (e.g. 3D reconstruction and rendering) [2, 16,243. The FOV of a conventional digital camera is limited by the size of the sensor and the focal length of the lens. For example, a typical 16mm lens with 213 CCD sensor has a 30 x 23 FOV. The number of pixels on the sensor ( 640x480 for NTSC camera) determines the resolution. The depth-offield is limited and is determined by various imaging parameters such as aperture, focal length, and the scene location of the object [12]. Many efforts have been made, which have succeeded in achieving various subsets of these properties (wide FOV, high resolution, large depthof-field, a single viewpoint, and real-time acquisition). A summary of these methods is presented in the next paragraph. One of these efforts [ 161, to which the present work is most closely related, uses a right mirror-pyramid, and as many cameras as the number of pyramid faces, each located and oriented to capture the part of the scene reflected off one of the faces (See Fig. 1). Images from the individual cameras are concatenated to yield a 360- degree wide panoramic FOV whose height is the same as that of the FOV of the individual cameras. This paper presents a design that has a 360-degree wide FOV as in [16], but uses two right mirror-pyramids to double the vertical FOV. The taller FOV comes at the cost of twice as many pyramids and cameras, additional complexity in camera geometry, and processing of individual camera images to obtain the single, larger, panoramic image. Before we describe our design, we will summarize some of the past work on panoramic and omni-directional image acquisition. These methods fall into two categories: dioptric methods, where only refractive elements (lenses) are employed, and catadioptric methods, where a combination of reflective and refractive components is used. Typical dioptric systems include: the camera cluster method where each camera points in a different direction and together the cameras cover all different directions [ 11; the fisheye method where a single camera acquires a wide FOV image through a fisheye lens [15, 23, 271; and the rotating camera method where a conventional camera [21] pans to generate mosaics, or a camera with a non-frontal, tilted sensor [12, 13, 141 pans around its viewpoint to acquire panoramic omni-focused images. The catadioptric methods include: sensors in which a single camera captures the scene as reflected off a single non-planar mirror [4, 5, 7, 8, 9, 18,20, 24,261, or sensors in which multiple cameras image the scene as reflected off the faces of a planar right mirror-pyramid [16, 111. The dioptric camera clusters achieve good resolution across a wide FOV at video rate. However, typically the cameras in the cluster do not share a unique 0-7695-1272-0101 $10.00 Q 2001 IEEE 1-960

viewpoint so that there may be uncovered space between adjacent cameras, and therefore, it is difficult or even impossible to seamlessly combine individual images to form a panoramic view without image blending. The sensors with fisheye lens are able to deliver large FOV images at video rate, but suffer from low resolution, irreversible distortion for close-by objects, and nonunique viewpoints for different portions of the FOV. The rotating cameras deliver high-resolution wide FOV via panning, as well as omni-focus when used in conjunction with non-frontal imaging, but they have limited vertical FOV. Furthermore, because they sequentially capture different parts of the FOV, moving objects may be imaged incorrectly. The sensors that use a parabolic- or a hyperbolic-mirror to map an omni-directional view onto a single sensor are able to achieve a single viewpoint at video rate, but the resolution of the acquired image is limited to that of the sensor used, and further, it is greatly reduced in the peripheral fields. Analogous to the dioptric case, this resolution problem can be resolved by replacing the simultaneous imaging of the entire FOV with panning and sequential imaging of its parts, followed by mosaicing the images [6, 19, 221, but at the expense of video rate. The use of the mirror pyramid, proposed by Nalwa [ 161 and mentioned earlier, achieves high resolution across a wide FOV at video-rate and with a single viewpoint for all directions across the FOV. In the next section, we first review Nalwa's panoramic camera system and then describe our design that doubles the vertical FOV in Sec. 3. 2. Panoramic imaging using a single mirror pyramid (SMP) One of the major problems in designing an omnidirectional sensor using a pinhole camera cluster in a straightforward manner [16] is to co-locate the multiple pinholes so that adjacent cameras cover contiguous FOV without obstructing the view of other cameras or itself. Nalwa [16] instead used planar mirrors to co-locate multiple pinholes [lo] for panoramic imaging. Fig. la illustrates this through a pair of planar mirrors and two associated cameras, C, and C,, positioned such that the mirror images of their projection centers coincide at a point C. Point C becomes the re-located common viewpoint of the two cameras. Nalwa proposed an n-sided right mirror-pyramid with a pinhole camera associated with each face such that the mirror image of every pinhole lies at the same location in space [ 161 as in Fig. lb. He reported an implementation using a 4-sided right pyramid and 4 pinhole-cameras. The pyramid stands on its horizontal base. Each triangular face forms a 45-degree angle with the base. The pinholes are positioned in the horizontal plane that contains the pyramid's vertex such that each pinhole is equidistant from the vertex and the mirror images of all the pinholes coincide at a single point on the axis of the pyramid. The cameras are pointed vertically downward at the pyramid faces, effectively viewing the world horizontally outward from the common virtual pinhole C [ 161. As seen in Fig. lb, this common virtual pinhole is at a location on the pyramid axis whose vertical distance from the pyramid apex is determined by the horizontal distance of the actual camera pinholes from the pyramid apex. The virtual optical axes of the cameras are horizontal, all contained in a plane parallel to the pyramid base and intersecting at the common virtual pinhole. The mapping from the scene points to the image points is as usual, e.g., a rectangular planar object perpendicular to the pyramid base is imaged as a rectangle. Kawanishi et al. [ll] used two such sensors to form a stereo pair. Each sensor uses a hexagonal pyramid and six cameras. The two pyramids, and therefore their common virtual pinholes, are separated vertically along their common axis by an amount equal to the desired stereo baseline. Nalwa suggested using the two pyramids back to back, with their bases coinciding, for such stereo viewing [17]. The vertical dimension of the panoramic FOV in each of the aforementioned cases is the same as that of each of the cameras used - only their horizontal FOV's are concatenated to obtain a wider, panoramic view. In the next section, we present an approach that doubles the vertical FOV while retaining the single-viewpoint and other characteristics of the panoramic image. 3. Proposed double mirror-pyramid (DMP) panoramic camera In this section, we describe our proposed new design that uses a double mirror-pyramid, formed by joining two T a/ A. 4 (b) Fig. 1 : Using planar mirrors to co-locate viewpoints C, and C, of two different cameras at aoint C. 1-961

mirror-pyramids such that their bases coincide (Fig. 2b), for panoramic viewing. The faces in each mirror-pyramid form an angle ct with the base, which determines the maximum vertical FOV of a single mirror-pyramid. The base edge of each face subtends an angle y within the base plane, which determines the horizontal FOV covered by each camera, and is given by the number of pyramid faces. We show how the use of such a double mirrorpyramid facilitates adding another layer of cameras parallel to the single layer present in the single mirrorpyramid (SMP) system. With the second layer, the FOV of any arbitrary camera gets extended not only horizontally, by the two cameras associated with adjacent faces in the SMP system, but also vertically, by the camera associated with the adjacent face in the other pyramid. The resulting double mirror-pyramid (DMP) system thus doubles the vertical FOV while preserving the ability to acquire panoramic high-resolution images from an apparent single viewpoint at video rate. To achieve the vertical contiguity in the panoramic image requires that both camera layers share a common virtual viewpoint. Clearly this is not possible by simply CO) Fig. 2: Comparison of mirror-pyramid panoramic cameras. (a) The SMP design consists of a single mirror-pyramid and a single horizontal layer of cameras; (b) The proposed DMP camera consists of a double truncated mirror-pyramid and two horizontal layers of cameras. replicating the SMP system vertically, because the common virtual viewpoint of each of the two SMP systems, although along the common axis of the two pyramids, is away from their common base. To achieve a common viewpoint, these two virtual viewpoints must be relocated to coincide, and the most direct way of achieving this is to relocate them to the center of the common base of the pyramids. To see how this could be achieved, consider Fig. 2a that shows an arbitrary single mirror-pyramid and the associated cameras, with the camera locations and orientations being controllable to relocate the common virtual viewpoint. The right mirror-pyramid TA, A,... A, is comprised of N identical triangular planar mirror surfaces (TA, A,, TA, A,,..., TA, A, ). These mirror surfaces form an identical angle ct with the base polygon A,A,... A,. A camera cluster C, of N cameras (CI,,ClZ,... Cl,) is placed so that each camera points at a different face of the pyramid, and all cameras share the same virtual viewpoint C on the axis OT of the pyramid. The cameras appear as if they are all at point C but pointing in different directions. The offset of point C from base A,A,... A, is h. Unlike the SMP system of [16], we make the cameras (C,,, C,,,... CIN ) point towards the locations (01,,OlZ,...OIN), respectively, on the pyramid faces. The lines connecting points (O,,,O,Z,...O,N) with the viewpoint (CIl,C,,,... CIN) are the optical axes of the corresponding cameras, and the lines connecting points (01,,012,... O,,) with the common virtual viewpoint C are the corresponding virtual optical axes. The optical axis of each camera makes an angle 4 with the normal of its corresponding pyramid face, or equivalently, each virtual optical axis makes an angle of Q0 with the base. The offset h and tilt angle (bo are the parameters whose values determine the relative locations of the two virtual viewpoints, and the angles formed by the two virtual viewing directions with the base. There are three distinct configurations to consider. (1) 40=0 or 4=90-a and h#o: This is the same configuration as used by Nalwa and Kawanishi for a = 45', provided the sensor and pyramid geometry, focal length, and the value of h are such that the entire vertical dimension of the sensor is filled with the SMP FOV (Fig. 3a). The virtual optical axes are parallel to but are not contained in the base plane. (2) 4" =O or 4 = 90 -a, and h = 0 : Here all virtual cameras face outwards from the common virtual pinhole, which is located at the center of the pyramid base. This is the same configuration as Nalwa's except that the virtual optical axes are contained in the base plane and bisect the 1-962

base edges. This is not a desirable solution since the lower half of the sensor does not receive any reflected light from its pyramid facet and therefore is unused (Fig. 3b). (3) @,, # 0 or @ < 90 -a and h = 0 : Unlike cases (1) and (2), in this case the different optical axes are not coplanar. Consequently, vertical planar objects, although equidistant from the pyramid axis, are foreshortened in the vertical direction (Fig. 3c). However, this distortion, also called keystone distortion, can be corrected by reprojecting each acquired image onto a virtual vertical sensor plane. Configuration (3), along with a double mirror-pyramid, leads to the proposed DMP design. A DMP is formed by stacking two truncated right mirror-pyramids, A,A,...A,B,B,...B, and A,A,... A,D,D,...P,, back to back so their bases coincide (Fig. 2(b)). Each of the two truncated pyramids has the same geometry as the single pyramid operating under case (3) just described. The tilt angle 8, is chosen to be 8, = 8,/2, where 8, is the vertical FOV of each camera. This design achieves a 360-degree FOV horizontally and double the individual camera FOV vertically, and also preserves the SMP's Fig. 3: The position and orientation of a camera relative to the mirror facet affect the reflective FOV captured by the sensor: (a) when @, =O and h 2 RI tan(8, /2), the entire vertical extent of the sensor captures the reflected scene and the image is free from keystone distortion; (b) when =O and h = 0, only half of the vertical extent of the sensor captures the reflected scene and the image is free from keystone distortion; (c) when @(, = 8, /2 and h=o, the entire vertical extent of the sensor captures the reflected scene, but the image suffers from keystone distortion ability for acquiring the entire panoramic image in high resolution itom a single apparent viewpoint. In the next section, we obtain the relationships between the parameters of the acquired panoramic image and those of the individual cameras and the imaging geometry. 4. Parameters of the DMP panoramic camera This section describes the parameters of the DMP imaging. We will refer to surface A,A,B,B, and its associated camera C,,, whenever a mirror face and the associated camera are considered. This is without loss of generality because the faces are symmetrically located about the pyramid axis as well as the pyramid base. Furthermore, and again without loss of generality, we use virtual viewpoint 0 and virtual optical axis 00,, to represent viewpoint C,, and optical axis C,,O,, respectively. The derived equations and observations will directly apply to all face-camera pairs. 4.1. DMP parameters The determination of the minimum acceptable number and dimensions of the pyramid faces is a key step in the DMP camera design. The FOV of each camera is computed in terms of the focal length f of the camera lens and the sensor size p(h)mm*q(v)mm. The horizontal, vertical and diagonal FOVs of a specified lens Xr and CCD sensor are given by 8, = 2ar~tan(~.), 8, = 2arctan( Lf ), and 8, = 2 arctan(gxf), respectively. To avoid a visual field gap, the reflective visual field of each mirror face must be equal or smaller than the FOV of an individual camera. For a given lens and a CCD sensor, the shape and size of a pyramid can be uniquely specified by the number of pyramid faces, N, the angle between the mirror faces (e.g. A,A,B,B,) and base polygon A,A,... A,, a, the radius of the polygon A,A,...A,, RI, and the height of 4.. Fig. 4: Parameter definitions in the DMP 1-963

. cos(-) I80 2s 2 ns90-e, N J ~ c O s ( e 12), (1) R, sin e,, tan a H> sin(@, +a) 360 2b), i.e. y = -. N 4.2. Visual field mapping We will now define the visual field mapping fiom the scene to the sensor achieved by the DMP camera. To do so, we first determine the reflective FOV covered by each camera. Then, the visual field mapping from the scene to the image sensor is given by the mapping from the mirror surface to the image sensor. The ray passing through the center of the entrance pupil of an optical system is defined as the principal ray of a scene point [lo]. With the assumption of a thin-lens model, the center of entrance pupil is the viewpoint, or the projection point. A scene point is uniquely defined by the angle that its principal ray makes with its optical axis, known as the field angle in optics. In Fig. 2@), a right-hand coordinate system OXIZ is defined in which axis OZ is perpendicular to polygonal plane A,A,... A, and points upward, axis OX is perpendicular to A,A, and points to the right, and axis OY is defined according to the right-hand rule. As shown in Fig. 5, a scene point Q on a plane P within the FOV of camera C,, maps onto point T on the mirror surface A,A,B, B,. On the other hand, with the assumption that the reflective FOV of the mirror is equal to or narrower than the camera FOV, a line connecting point T(x, y, z) on the mirror surface A, A, B, B, with the virtual viewpoint 0 (or real viewpoint C, I ) uniquely defines a principal ray, and therefore, a scene point Q. The visual angle B(T) subtended by point T(x,y,z) - with O,,C,, is equivalent to the angle between vector OT and virtual optical axis 00,,, and is given by x * XOI, + Y * YOll + z * ZOI1 cosb(t) = loo, I where I - R, sin aces(@, /2) xq, = sin@ + 8, / 2), and Y",, = 0 R, sinasin(@, 12) Z",, = sin(a + 8, / 2) I T(x, y, z) is constrained by the plane equation: X - Z = l RI RI tuna - R, S x S -RI sin acos8, /sin(a + 8, ). - R, tan(%) s y s R, tan(%) Q5zS RI sinasin8,/sin(a+8,) We are particularly interested in the mapping of boundary lines A, A,, B,B,, A,B,, and A,B2. These boundary mappings determine the trapezoidal shape of the rectangular object on the image sensor. For any point T(x, y,z) on line A, A, ( B,B,, A,B,, or A,B, ), its corresponding visual angle is given by Specifically, for points A,, A,, B,, B,, and O,,, the sensor is normal to the optical axis of the camera lens and is aligned so that the central pixel o;, of the sensor is the intersection of the optical axis 00,, with the sensor plane, and the long axis of the sensor is parallel to A,A,, as shown in Fig. 5. We further assume that the camera lens is free from aberrations and the pinhole model applies. Then the (3) 1-964

Fig. 5: Mapping a point on plane P onto image sensor. 0 is the virtual viewpoint of camera C,,, 00,, is the virtual optical axis. A rectangular object A,,A,,B,,B,, is mapped onto a trapezoidal shape A',, At2, B'2p B',,. mapping of point 0,, (i.e. the scene center) on image sensor would be the central pixel O,,', and the mapping point Q' where scene point Q is mapped is given by IO'Q'I =,f tanb(t). The mapping is illustrated in Fig. 6, which shows that a rectangular object is foreshortened and imaged as a trapezoid due to the camera tilt. This phenomenon is referred to as keystone distortion and must be corrected by dewarping to recover the rectangular image. Mlrror Surface -- -- Mapped FOV Fig. 6: Visual field mapping on sensor. and a prototype of the mirror system are shown in Fig. 7. The cameras used are Pulnix with 2/3", 640x480 blacwwhite CCD sensors and 6.5mm lenses. Each of the cameras effectively covers 60 degrees FOV horizontally and 40 degrees FOV vertically. The total FOV of the DMP is 360"(H)*80'(V). The field angles corresponding to comers A,, A,, B,, B,, HI, K, are O(AJ = 35.53", e(a,) = 35.53*, e(b,) = 30.75-, O(B,) = 30.75', Q(H,) = 20", and e(k,) = 20", respectively. In order to capture a panoramic image fiom a single viewpoint using the DMP camera, the camera clusters must be placed properly with respect to the mirrorpyramids. Otherwise, there may be gaps between the visual fields of the different cameras. For lack of space, we will not present the details of the calibration process in this paper. Since the FOV of each camera is more than 60 degrees horizontally, pincushion or barrel distortion is almost inevitable and needs to be compensated for. For this purpose, each camera is calibrated using Zhang's calibration method [25], which images a planar pattern at different orientations to estimate the intrinsic and extrinsic parameters, and therefore radial distortion. Again, for lack of space, we will omit the details of this step. Fig. 8a shows an original image and Fig. 8b shows the image after compensating for the radial distortion. Furthermore, each image, after compensating for the radial distortion, needs to be re-projected onto a virtual sensor that is perpendicular to the pyramid base. This step removes the keystone distortion. An alternative is to avoid the distortion in the first place by tilting the CCD sensor of the camera by -6, degrees with respect to its optical axis, which maps a vertical rectangular object in the scene as a rectangle on the tilted sensor. This avoids the computation required for digital re-projection, and eliminates the loss of quality due to interpolation that accompanies digital re-projection, but is more demanding because it requires a nontrivial manipulation of the camera hardware to adjust sensor orientation. In our 5. Implementation and experimental results Using the equations derived in section 3, we designed a DMP panoramic camera with two right-hexagonal (N = 6) truncated pyramids. The small-circle radius of the right hexagon is R, = 86.6mm..The angle a between each mirror surface and the base of the hexagon is 40 degrees, and the tilt angle of the optical axes is 20 degrees (S,, = 20" and Q, = 40' ). A 3D CAD model of the system (a) (b) Fig. 7: A 6-faced DMP camera: (a) A CAD model; (b) Our implementation.

adjacent in the lower layer (C,,, C,,). C,, and C, are vertically adjacent and so are C,, and C,,. Figures 9(e) and (f) show the seamless mosaics of the images provided by the camera pairs C,, and C,, and C,, and C,, respectively, after post-processing for keystone and radial distortions. Fig. 9(g) shows the seamless cylindrical mosaic of Figs 9(e) and (f) [22]. Due to the mirror effect, the original images acquired by the cameras need to be flipped appropriately before they are mosaiced. 6. Discussion In generating panoramic images using a mirrorpyramid, we have assumed that the cameras are pinholes, i.e., the aperture is very small, and the edges formed by adjacent mirrors are perfect, i.e. knife-shaped and free of rounding. However, these assumptions do not hold in practice. For example, a real camera has a non-pinhole aperture, and its size vanes as the F-number setting is changed. Therefore, a bundle of rays from a scene point, instead of only the principal ray, go through the entrance pupil to form the image of the point. The effect of the non-pinhole aperture minimally affects the parts of the image formed by reflections off the interiors of the pyramid faces, but there is mixing of light arriving directly from the scene and that after reflection off the mirror. This leads to artifacts. Furthermore, since the two adjacent mirrors do not form a perfect knife-edge, the edge curvature further adds to the complexity of the image formed. For example, if two adjacent faces have a flat surface transition (i.e. an approximately planar patch connecting the two faces and having a normal close to the average of the two face normals), this may lead to a loss of light from the corresponding part of the scene, resulting in a dark band, which will occur periodically (g) Fig. 9: Sample images obtained by the DMP camera. (a)-(d) Original images of camera Cl1, CIZ, Czl and Cz2; (e) The mosaic of images (a) and (c) (a> (b) (c) after post-processing of keystone and radial Fig. 8: Compensating for distortions. (a) Original DMP distortions; (f) The mosaic of images (b) and (d) image; (b) Result after correcting radial distortion after post-processing of keystone and radial using Zhang s calibration method; (c) Result after distortions; (g) The cylindrical mosaic of the images compensating for both radial and keystone distortions. (e)-(fh 1-966

across the panoramic image along each boundary between sub-images acquired by adjacent cameras. Such artifacts will occur in all mirror-pyramid based systems. We plan to analyze these effects and explore possible ways of eliminating or softening them. Acknowledgements This work was supported in part by National Science Foundation Grant IIS 00-83037 ITR. We thank Chunyu Gao for his assistance in the design, building and calibration of the camera. References P. I. Anderson, "From telepresence to true immersive imaging: into real-life video-now!", Advanced Imaging, 1995, Vol. 10(7), ~ ~48-50, 1995. S. Baker and S. K. Nayar, "A theory of single-viewpoint catadioptric image formation", in International Journal of Computer Vision, 1999,35(2), pp.1-22. R. Benosman, E. Deforas, and J. Devars, "A new catadioptric sensor for the panoramic vision of mobile robots", in Workshop on Omnidirectional Vision, June 2000, pp. 112-116. J. S. Chahl and M. V. Srinivasan, "Reflective surfaces for panoramic imaging", Applied Optics, November 1997, 36(31):pp. 8275-8285. J. S. Chahl and M. V. Srinivasan, "A complete panoramic vision system, incorporating imaging, ranging, and three dimensional navigation", in Workshop on Omnidirectional Vision, June 2000, pp. 104-111. S. Coorg, N. Master, and S. Teller, "Acquisition of a large pose-mosaic dataset", in Conference on Computer Vision aiid Pattern Recognition, 1998, pp. 872-878. R. A. Hicks, "Reflective surfaces as computational sensors", in Workshop on Perception.for Mobile Agents, 1999, pp. 82-86. R. A. Hicks and R. Bajcsy, "Catadioptric sensors that approximate wide-angle perspective projections", in Workshop on Onznidirectional Vision, June 2000, pp. 97-103. H. Ishiguro, M. Yamamoto, and S. Tsuji, "Omnidirectional stereo", IEEE Transactions On Pattern Analysis and Machine Intelligence, February 1992, 14(2), pp257-262. [lo] B. R. Johnson, "Lenses", Handbook of Optics: Device Measurement and Properties, Michael Bass, Eric W. Van Stryland et. al. Editors, volume 11. McGraw Hill, 2nd edition, 1995. [ 1 I] Takahito Kawanishi, Kazumasa Yamazawa, et al. "Generation of high resolution stereo panoramic images by omnidirectional imaging sensor using hexagonal pyramidal mirrors," 141h International Conference on Pattern Recognition, 16th-20th August, 1998, Brisbane, Australia, pp. 485-489, 1998. [ 121 A. Krishnan and N. Ahuja. "Range estimation from focus using il non-frontal imaging camera," in National Conference on Artijkial Intelligence, pages 830-835, Washington D.C., July 11-15 1993. [ 131 A. Krishnan an N. Ahuja, Range Estimation from Focus using a Nonfrontal Imaging Camera, Int. Journal of Computer Vision, Vol. 20, No. 3, 1996, 169-185. [ 141 A. Krishnan and N. Ahuja, "Panoramic image acquisition", in Conference on Computer Vision and Pattern Recognition, 1996, pp. 379-384. [ 151 K. Miyamoto, "Fish eye lens", Journal of Optical Society of America, August 1964,64:pp. 1060-1061. [16] V. Nalwa, "A true omnidirectional viewer", Technical report, Bell Laboratories, February 1996. 1171 V. Nalwa, "Stereo panoramic viewing system", US Patent 6141145. [ 181 S. K. Nayar, "Catadioptric omnidirectional camera", in Conference on Computer Vision and Pattern Recognition, 1997, pp. 482-488. [19] S. K. Nayar and A. Karmarkar, "360x360 mosaics", in Conference on Computer Vision and Pattern Recognition, June 2000, volume 2, pp. 388-393. [20] S. K. Nayar and V. Peri, "Folded catadioptric cameras", in Conference on Computer Vision and Pattern Recognition, June 1999, volume 2, pp. 217-223. E211 S. Peleg, "Panoramic mosaics by manifold projection", in Conference on Computer Vision and Pattern Recognition, June 1997, pp 338-343. [22] H.-Y. Shum and R. Szeliski, "Panoramic image mosaics", Technical Report MSR-TR-97-23, Microsofi Research, 1997. [23] Y. Xiong and K. Turkowski, "Creating image-based VR using a self-calibrating fisheye lens", in Conference on Computer Vision and Pattern Recognition, 1997, pp. 237-243. [24] K. Yamazawa, Y. Yagi, and M. Yachida, "Omnidirectional imaging with hyperboloidal projection", in International Conference on Intelligent Robots and Systems, July 1993, pp 1029-1034. [25] Zhengyou Zhang, "A flexible new technique for camera calibration", Technical Report MSR-TR-98-71, Microsoft Research, 1998. [26] Z. Zhu, K. D. Rajasekar, E. M. Riseman, and A. R. Hanson, "Panoramic virtual stereo vision of cooperative mobile robots for localizing 3D moving objects", in Workshop on Omnidirectional Vision, 2000, pp 29-36. E271 S. Zimmerman and D. Kuban, "A video pan/tilt/magnify/rotate system with no moving parts", in IEEE/AIAA Digital Avionics Systems Conference, 1992, pp. 523-531. 1-967