VLSI DESIGN OF A HIGH-SPEED CMOS IMAGE SENSOR WITH IN-SITU 2D PROGRAMMABLE PROCESSING

Similar documents
Design and Simulation of High Speed Multi-Processing CMOS Image Sensor

A vision sensor with on-pixel ADC and in-built light adaptation mechanism

A CMOS Imager with PFM/PWM Based Analogto-digital

TODAY, improvements in the growing digital imaging

A CMOS Image Sensor with Ultra Wide Dynamic Range Floating-Point Pixel-Level ADC

Research Article An SIMD Programmable Vision Chip with High-Speed Focal Plane Image Processing

Adaptive sensing and image processing with a general-purpose pixel-parallel sensor/processor array integrated circuit

EE 392B: Course Introduction

Optical Flow Estimation. Using High Frame Rate Sequences

A Dynamic Range Expansion Technique for CMOS Image Sensors with Dual Charge Storage in a Pixel and Multiple Sampling

IT IS widely expected that CMOS image sensors would

Image Processing Vision System Implementing a Smart Sensor

Trend of CMOS Imaging Device Technologies

Image toolbox for CMOS image sensors simulations in Cadence ADE

Characterization of CMOS Image Sensors with Nyquist Rate Pixel Level ADC

Fundamentals of CMOS Image Sensors

Interpixel crosstalk in a 3D-integrated active pixel sensor for x-ray detection

A high speed programmable focal-plane SIMD vision chip

Computational Sensors

A flexible compact readout circuit for SPAD arrays ABSTRACT Keywords: 1. INTRODUCTION 2. THE SPAD 2.1 Operation 7780C - 55

A Digital High Dynamic Range CMOS Image Sensor with Multi- Integration and Pixel Readout Request

A 1.3 Megapixel CMOS Imager Designed for Digital Still Cameras

Polarization-analyzing CMOS image sensor with embedded wire-grid polarizers

Digital Photographs and Matrices

IN RECENT years, we have often seen three-dimensional

Digital Photographs, Image Sensors and Matrices

A Sorting Image Sensor: An Example of Massively Parallel Intensity to Time Processing for Low Latency Computational Sensors

ABSTRACT. Section I Overview of the µdss

A 3MPixel Multi-Aperture Image Sensor with 0.7µm Pixels in 0.11µm CMOS

CMOS Active Pixel Sensor Technology for High Performance Machine Vision Applications

A Foveated Visual Tracking Chip

NEW CIRCUIT TECHNIQUES AND DESIGN METHODES FOR INTEGRATED CIRCUITS PROCESSING SIGNALS FROM CMOS SENSORS

Design of an Integrated Image Sensor System

Comparison between Analog and Digital Current To PWM Converter for Optical Readout Systems

Column-Parallel Architecture for Line-of-Sight Detection Image Sensor Based on Centroid Calculation

ELEN6350. Summary: High Dynamic Range Photodetector Hassan Eddrees, Matt Bajor

Active Pixel Sensors Fabricated in a Standard 0.18 um CMOS Technology

Pixel Level Processing Why, What, and How?

CHARGE-COUPLED device (CCD) technology has been. Photodiode Peripheral Utilization Effect on CMOS APS Pixel Performance Suat Utku Ay, Member, IEEE

Low Voltage Low Power CMOS Image Sensor with A New Rail-to-Rail Readout Circuit

A High Image Quality Fully Integrated CMOS Image Sensor

Low Power Sensors for Urban Water System Applications

Cameras CS / ECE 181B

Fully depleted, thick, monolithic CMOS pixels with high quantum efficiency

A 1Mjot 1040fps 0.22e-rms Stacked BSI Quanta Image Sensor with Cluster-Parallel Readout

Photons and solid state detection

A Digital High Dynamic Range CMOS Image Sensor with Multi- Integration and Pixel Readout Request

IT FR R TDI CCD Image Sensor

A 120dB dynamic range image sensor with single readout using in pixel HDR

CCD1600A Full Frame CCD Image Sensor x Element Image Area

IEEE. Proof. CHARGE-COUPLED device (CCD) technology has been

DURING the past few years, fueled by the demands of multimedia

Techniques for Pixel Level Analog to Digital Conversion

STA1600LN x Element Image Area CCD Image Sensor

A new Photon Counting Detector: Intensified CMOS- APS

Comparative Analysis of SNR for Image Sensors with Enhanced Dynamic Range

Chapter 3 Wide Dynamic Range & Temperature Compensated Gain CMOS Image Sensor in Automotive Application. 3.1 System Architecture

A new Photon Counting Detector: Intensified CMOS- APS

Characterisation of a CMOS Charge Transfer Device for TDI Imaging

A CMOS Image Sensor With Dark-Current Cancellation and Dynamic Sensitivity Operations

Charged Coupled Device (CCD) S.Vidhya

Picosecond time measurement using ultra fast analog memories.

[2] Brajovic, V. and T. Kanade, Computational Sensors for Global Operations, IUS Proceedings,

Front-End and Readout Electronics for Silicon Trackers at the ILC

Ultra-high resolution 14,400 pixel trilinear color image sensor

All-digital ramp waveform generator for two-step single-slope ADC

RECENTLY, CMOS imagers, which integrate photosensors, A New CMOS Pixel Structure for Low-Dark-Current and Large-Array-Size Still Imager Applications

DAT175: Topics in Electronic System Design

Low Power Design of Successive Approximation Registers

Yet, many signal processing systems require both digital and analog circuits. To enable

444 Index. F Fermi potential, 146 FGMOS transistor, 20 23, 57, 83, 84, 98, 205, 208, 213, 215, 216, 241, 242, 251, 280, 311, 318, 332, 354, 407

Introduction to Computer Vision

CAFE: User s Guide, Release 0 26 May 1995 page 18. Figure 13. Calibration network schematic. p-strip readout IC

Application of CMOS sensors in radiation detection

Power and Area Efficient Column-Parallel ADC Architectures for CMOS Image Sensors

ON CHIP ERROR COMPENSATION, LIGHT ADAPTATION, AND IMAGE ENHANCEMENT WITH A CMOS TRANSFORM IMAGE SENSOR

On-Chip Binary Image Processing with CMOS Image Sensors

8.2 IMAGE PROCESSING VERSUS IMAGE ANALYSIS Image processing: The collection of routines and

Device design for global shutter operation in a 1.1-um pixel image sensor and its application to nearinfrared

A 19-bit column-parallel folding-integration/cyclic cascaded ADC with a pre-charging technique for CMOS image sensors

Low Noise Wide Dynamic Range Image Sensor Readout using Multiple Reads During Integration (MRDI)

An Analog Phase-Locked Loop

ELEC Dr Reji Mathew Electrical Engineering UNSW

DELIVERABLE!D60.4! 1k!x!1k!pnCCD!Conceptual!Design! WP60!Advanced!Instrumentation!Development! 1 ST Reporting Period.

Megapixels and more. The basics of image processing in digital cameras. Construction of a digital camera

TRIANGULATION-BASED light projection is a typical

Variability-Aware Optimization of Nano-CMOS Active Pixel Sensors using Design and Analysis of Monte Carlo Experiments

A Pixel Silicon Retina for Gradient Extraction With Steering Filter Capabilities and Temporal Output Coding

Image sensor combining the best of different worlds

e2v Launches New Onyx 1.3M for Premium Performance in Low Light Conditions

A 3D Multi-Aperture Image Sensor Architecture

Demonstration of a Frequency-Demodulation CMOS Image Sensor

Advanced output chains for CMOS image sensors based on an active column sensor approach a detailed comparison

A Parallel Analog CCD/CMOS Signal Processor

Review of ADCs for imaging

Analog Circuit for Motion Detection Applied to Target Tracking System

The Electronics Readout and Measurement of Parameters of. a Monitor System

Highly Efficient Ultra-Compact Isolated DC-DC Converter with Fully Integrated Active Clamping H-Bridge and Synchronous Rectifier

Digital Imaging Rochester Institute of Technology

DESIGN AND PERFORMANCE VERIFICATION OF CURRENT CONVEYOR BASED PIPELINE A/D CONVERTER USING 180 NM TECHNOLOGY

Transcription:

VLSI DESIGN OF A HIGH-SED CMOS IMAGE SENSOR WITH IN-SITU 2D PROGRAMMABLE PROCESSING J.Dubois, D.Ginhac and M.Paindavoine Laboratoire Le2i - UMR CNRS 5158, Universite de Bourgogne Aile des Sciences de l Ingenieur, BP 47870, 21078 Dijon cedex, France phone: +33(0)3.80.39.36.03, fax: +33(0)3.80.39.59.69, email: jerome.dubois@u-bourgogne.fr web: http://www.le2i.com ABSTRACT A high speed VLSI image sensor including some preprocessing algorithms is described in this paper. The sensor implements some low-level image processing in a massively parallel strategy in each pixel of the sensor. Spatial gradients, various convolutions as Sobel or Laplacian operators are described and implemented on the circuit. Each pixel includes a photodiode, an amplifier, two storage capacitors and an analog arithmetic unit. The systems provides address-event coded outputs on tree asynchronous buses, the first output is dedicated to the result of image processing and the two others to the frame capture at very high speed. Frame rates to 10 000 frames/s with only image acquisition and 1000 to 5000 frames/s with image processing have been demonstrated by simulations. Keywords: CMOS Image Sensor, parallel architecture, High-speed image processing, analog arithmetic unit. 1. INTRODUCTION Today, improvements continue to be made in the growing digital imaging world with two main image sensor technologies: the charge-coupled devices (CCD) and CMOS sensors. The continuous advances in CMOS technology for processors and DRAMs have made CMOS sensor arrays a viable alternative to the popular CCD sensors. New technologies provide the potential for integrating a significant amount of VLSI electronics onto a single chip, greatly reducing the cost, power consumption and size of the camera [1-4]. This advantage is especially important for implementing full image systems requiring significant processing such as digital cameras and computational sensors [5]. Most of the work on complex CMOS systems talks about the integration of sensors integrating a chip level or column level processing [6, 7]. Indeed, pixel level processing is generally dismissed because pixel sizes are often too large to be of practical use. However, integrating a processing element at each pixel or group of neighbouring pixels seems to be promising. More significantly, employing a processing element per pixel offers the opportunity to achieve massively parallel computations. This benefits traditional high-speed image capture [8-9] and enables the implementation of several complex applications at standard video rates [10-11]. This paper describes the design and implementation of a 64 x 64 active-pixel sensor with per-pixel programmable processing element fabricated in a standard 0.35-µm CMOS technology. The main objectives of our design are: (1) to evaluate the speed of the sensor, and, in particular, to reach 10 000 frames/s, (2) to demonstrate a versatile and programmable processing unit at pixel level, (3) to provide an original platform dedicated to embedded image processing. The remainder of this paper is organized as follows: first, the section II describes the main characteristics of the overall architecture. The section III is dedicated to the description of the image processing algorithms embedded at pixel level in the sensor. Then, in the section IV, we describe the details of the pixel design. Finally, we present some simulation results of high speed image acquisition with or without processing at pixel level. 2. OVERALL ARCHITECTURE Figure 1 shows a photograph of the image sensor and the main chip characteristics are listed in Table 1. The chip consists of a 64 x 64 element array of photodiode active pixels and contains 160 000 transistors on a 3.5 x 3.5 mm die. Each pixel is 35 µm on a side with a fill factor of about 25 %. It contains 38 transistors including a photodiode, an Analog Memory Amplifier and Multiplexer ([AM]²), an Analog Arithmetic Unit (A²U). The chip also contains test structures used for detailed characterization of pixels and processing units (on the bottom left of the chip). Figure 1 - Photograph of the image sensor in a standard 0.35µm CMOS technology

Technology 0.35µm 4-metal CMOS Array size 64 x 64 Chip size 11 mm2 Number of transistors 160 000 Number of transistors / pixel 38 Pixel size 35 µm x 35 µm Sensor Fill Factor 25 % Dynamic power consumption 750 mw Supply voltage 3.3 V Frame Rate without processing 10000 fps Frame Rate with processing 1000 to 5000 fps x (1) (2) (4) (3) y y β x Table 1 - Chip Characteristics 3. EMBEDDED ALGORITHMS Our main objective is the implementation of various in-situ image processing using local neighbourhood (spatial gradients, Sobel and Laplacian operators). So, the processing element structure must be reconsidered. Indeed, in a traditional point of view, a CMOS sensor can be seen as an array of independent pixels, each including a photodetector (PD) and a processing element () built upon few transistors mainly dedicated to improve sensor performance as in an Active Pixel Sensor or APS (Figure 2.a). (a) Figure 2 - (a) Photosite with intrapixel processing, (b) photosite with interpixel processing In our point of view, the electronics, which takes place around the photodetector, has to be a real on-chip analog signal processor which improves the functionality of the sensor. This typically allows local and/or global pixel calculations, implementing some early vision processing. This obliges to allocate the processing resources, so that each computational unit can easily use a neighbourhood of various pixels. In our design, each processing element takes place in the centre of 4 adjacent pixels (Figure 2.b). This distribution is more propitious to the algorithms - embedded in the sensor - which will be described in the following sections. 3.1 Spatial gradients PD: Photo Detector : Processing Element The structure of our processing unit is perfectly adapted to the computation of spatial gradients (Figure 3). (b) Figure 3 Computation of spatial gradients The main idea for evaluating these gradients [12] in-situ is based on the definition of the first-order derivative of a 2-D function performed in the vector directionξ, which can be expressed as: = cos( β ) + sin( β ) ξ x' y' Where β is the vector s angle. A discretization of the equation (1) at the pixel level, according to the Figure 3, would be given by: = ξ ( 3 V2 V4)cos( β) + ( V1 V )sin( β) Where V i (with 1 i 4 ) is the luminance at the pixel i i.e. the photodiode output. In this way, the local derivative in the direction of vector ξ is continuously computed as a linear combination of two basis functions, the derivatives in the x and y directions. Using a four-quadrant multiplier, the product of the derivatives by a circular function in cosines can be easily computed. The output P is given by: P= V1 sin( β) + V2 V3 sin( β) V4 (3) Coef1 Coef2 Coef3 Coef4 (1) V1 V2 (2) Multipliers (3) V3 V4 (4) Product Figure 4 - Implementation of four multipliers (1) (2)

According to the Figure 4, the processing element implemented at the pixel level carries out a linear combination of the four adjacent pixels by the four associated weights (coef1, coef2, coef3 and coef4). To evaluate the equation 3, the following values have to be given to the coefficients: coef1 coef 3 3.2 Sobel operator coef 2 sin( β) = coef 4 sin( β) The structure of our architecture is also adapted to various algorithms based on convolutions using binary masks on a neighbourhood of pixels. As example, we describe here the Sobel operator. With our chip, the evaluation of the Sobel algorithm leads to the result directly centred on the photosensor and directed along the natural axes of the image (Figure 5.a). In order to compute the mathematical operation, a 3x3 neighbourhood (Figure 5.b) is applied on the whole image. (4) and the horizontal axis (I h2 =I 21 +I 22 +I 23 +I 24 ) can be computed. 3.3 Second-order detector: Laplacian Edge detection based on some second-order derivatives as the Laplacian can also be implemented on our architecture. Unlike spatial gradients previously described, the Laplacian is a scalar (see equation 7) and does not provide any indication about the edge direction. 0 1 1 = 0 1-41 0 1 0 From this 3x3 mask, the following series of operations can be extracted according to the principles used previously for the evaluation of the Sobel operator; (7) 1 : I 11 =I 4 -I 5 I 12 =I 2 -I 5 I 13 =I 6 -I 5 I 14 =I 8 -I 5 (8) i i+1 i+2 j j+1 j+2 a1 a2 a3 (1) (2) a4 a5 a6 (4) (3) a7 a8 a9 (a) (b) Figure 5 - (a) Array architecture, (b) 3x3 mask used by the four processing elements 4.1 Pixel Design 4. IMAGE SENSOR DESIGN The proposed chip is based on a classical architecture widely used in the literature as shown on the figure 6. The CMOS image sensor consists of an array of pixels that are typically selected a row at a time by a row decoder. The pixels are read out to vertical column buses that connect the selected row of pixels to an output multiplexer. The chip includes three asynchronous output buses, the first one is dedicated to the image processing results whereas the two others provides parallel outputs for full high rate acquisition of raw images. In order to compute the derivatives in the two dimensions, two 3x3 matrices called h1 and h2 are needed: -1 0 h 1 = 1-202 -1 0 1-1 -2 et h 2 = -1 0 0 0 1 2 1 Within the four processing elements numbered from 1 to 4 (see Figure 5.a), two 2x2 masks are locally applied. According to (5), this allows the evaluation of the following series of operations: (5) h 1 : I 11 =-(I 1 +I 4 ) h 2 : I 21 =-(I 1 +I 2 ) I 12 = I 3 +I 6 I 22 =-(I 2 +I 3 ) I 13 = I 6 +I 9 I 23 = I 8 +I 9 I 14 =-(I 4 +I 7 ) I 24 = I 7 +I 8 (6) With the I 1i and I 2i provided by the processing element (i). Then, from these trivial operations, the discrete amplitudes of the derivatives along the vertical axis (I h1 =I 11 +I 12 +I 13 +I 14 ) Figure 6 - Image sensor system architecture As a classical APS, all reset transistor gates are connected in parallel, so that the whole array is reset when the reset line is

active. In order to supervise the integration period (i.e. the time when incident light on each pixel generates a photocurrent that is integrated and stored as a voltage in the photodiode), the global output called Out_int on the Figure 6 provides the average incidental illumination of the whole matrix of pixels. If the average level of the image is too low, the exposure time may be increased. On the contrary, if the scene is too luminous, the integration period may be reduced. Operations Capture sequence Memory 1 Sequential Readout Memory 2 Capture sequence Memory 2 Sequential Readout Memory 1 t0 t1 t2 100 µs 100 µs Figure 7 Parallelism between capture sequence and readout sequence time which represents the first stage of the capture circuit. The pixel array is held in a reset state until the reset signal goes high. Then, the photodiode discharges according to incidental luminous flow. While the read signal remains high, the analog switch is open and the capacitor C AM of the analog memory stores the pixel value. The C AM capacitors are able to store, during the frame capture, the pixel values, either from the switch 1 or the switch 2. The following inverter, polarized on V DD /2, serves as an amplifier of the stored value and provides a level of tension proportional to the incidental illumination to the photosite. 4.2 Analog Arithmetic Unit (A²U) design The analog arithmetic unit (A²U) represents the central part of the pixel and includes four multipliers (M1, M2, M3 and M4), as illustrated on the Figure 9. The four multipliers are all interconnected with a diode-connected load (i.e. a NMOS transistor with gate connected to drain). The operation result at the node point is a linear combination of the four adjacent pixels. To increase the algorithmic possibilities of the architecture, the acquisition of the light inside the photodiode and the readout of the stored value at pixel level are dissociated and can be independently executed. So, two specific circuits, including an analog memory, an amplifier and a multiplexer are implemented at pixel level. With these circuits called [AM]² (Analogical Memory, Amplifier and Multiplexer), the capture sequence can be made in the first memory in parallel with a readout sequence and/or processing sequence of the previous image stored in the second memory (as shown on Figure 7). With this strategy, the frame rate can be increased without reducing the exposure time. Simulations of the chip show that frame rates up to 10 000 fps can be achieved. Figure 9 (a) Architecture of the A²U (b) Four-quadrant multiplier schematic 5. LAYOUT AND PRELIMINARY RESULTS Figure 8 Frame capture circuit schematic In each pixel, as seen on Figure 8, the photosensor is a nmos photodiode associated with a PMOS transistor reset, The layout of a 2x2 pixel block is shown in Figure 10. This layout is symmetrically built in order to reduce fixed pattern noise among the four pixels and to ensure uniform spatial sampling. An experimental 64x64 pixel image sensor has been developed in a 0.35µm, 3.3 V, standard CMOS technology with poly-poly capacitors. This prototype has been sent to foundry at the beginning of 2006 and will be available at the end of the first quarter of the year. The Figure 11 represents a simulation of the capture operation. Various levels of illumination are simulated by activating different readout signals (read 1 and read 2). The two outputs (output 1 and output 2) give the levels between OV and V DD, corresponding to incidental illumination on the pixels. The calibration of the structure is ensured by the biasing (Vbias=1,35V). Moreover, in this simulation, the node output is the result of the difference operation (out1- out2). The factors were fixed at the following values: coef1=-coef2=v DD and coef3= coef4=v DD /2.

14th European Signal Processing Conference (EUSIPCO 2006), Florence, Italy, September 4-8, 2006, copyright by EURASIP MOS transistors operate in sub-threshold region. There is no energy spent for transferring information from one level of processing to another level. According to the simulation s results, the voltage gain of the amplifier stage of the two [ AM]² is Av=10 and the disparities on the output levels are about 4.3%. Figure 11 Simulation of (1) high speed sequence capture and (2) basic processing between neighbouring pixels Figure 10 - Layout of a pixel 6. CONCLUSION An experimental pixel sensor implemented in a standard digital CMOS 0.35µm process was described. Each 35 µm x 35 µm pixel contains 38 transistors implementing a circuit with photo-current s integration, two [AM]² (Analog Memory, Amplifier and Multiplexer), and a A²U (Analogical Arithmetic Unit). Simulations of the chip show that frame rates up to 10000 fps with no image processing and frame rates from 1000 to 5000 fps with basic image processing can be easily achieved using the parallel A²U implemented at pixel level. The next step in our research will be the characterization of the real chip as soon as possible and the integration on-situ of a fast analog digital converter able to provide digital images with embedded image processing at thousands of frames by second. REFERENCES [1] E. R. Fossum, Active pixel sensors (APS): Are CCDs dinosaurs?, In Proc. SPIE, vol. 1900, pp. 2-14, 1992. [2] E. R. Fossum, CMOS image sensors: Electronic cameraon-a-chip, IEEE Transactions on Electron Devices, vol. 44, no. 10, pp. 1689-1698, 1997. [3] D. Litwiller, CCD vs. CMOS: facts and fiction, Photonics Spectra, pp. 154.158, 2001. [4] P. Seitz, Solid-state image sensing, Handbook of Computer Vision and Applications, vol. 1, 165-222, 2000. [5] A. El Gamal, D. Yang, and B. Fowler, Pixel level processing Why, what and how?, Proc. SPIE, vol. 3650, pp. 213, 1999. [6] M. J. Loinaz, K. J. Singh, A. J. Blanksby, D. A. Inglis, K. Azadet, and B. D. Ackland, A 200mW 3.3V CMOS color camera IC producing 352x288 24b Video at 30 frames/s, IEEE Journal of Solid-State Circuits, vol. 33, pp. 2092-2103, 1998. [7] S. Smith, J. Hurwitz, M. Torrie, D. Baxter, A. Holmes, M. Panaghiston, R. Henderson, A. Murrayn S. Anderson, and P. Denyer, A single-chip 306x244-pixel CMOS NTSC video camera, In ISSCC Digest of Technical Papers, San Fransisco, CA, pp. 170-171, 1998. [8] B. M. Coaker, N. S. Xu, R. V. Latham, and F. J. Jones, High-speed imaging of the pulsed-field flashover of an alumina ceramic in vacuum, IEEE Trans. of Dielect; Electric. Insul., vol. 2, pp. 210-217, 1995. [9] B. T. Turko, and M. Fardo, High-speed imaging with a tapped solid state sensor, IEEE Transactions on Nuclear Science, vol. 37, pp. 320-325, 1990. [10] D. Yang, A. El Gamal, B. Fowler, and H. Tian, A 640 x 512 CMOS image sensor with ultra wide dynamic range floating point pixel-level ADC, IEEE Journal of Solid-State Circuits, vol.34, pp. 1821-1834, 1999. [11] S. H. Lim, and A. El Gamal, Integrating image capture and processing Beyond single chip digital camera, In Proc. SPIE, vol. 4306, 2001. [12] Massimo Barbaro et al, A 100x100 Pixel Silicon Retina for Gradient Extraction With Steering Fiter Capabilities and Temporal Output Coding, IEEE Journal of solid-state circuits, vol. 37, no. 2, feb. 2002.