Date post: | 07-Apr-2018 |
Category: |
Documents |
Upload: | aleksandarha |
View: | 218 times |
Download: | 0 times |
of 12
8/3/2019 Cmos Primer
1/12
An Introduction to CMOS Image Sensor Technology
Preface
THE PURPOSE of this primer is to introduce the technology used in CMOS image sensors
as well as promoting a better understanding of the consumer product potential unleashed
as the Internet, digital imaging and integrated information appliances converge. It isassumed that the reader is familiar with some basic physics and semiconductor
technology, but most of this overview covers entry level imaging technology information.
THIS DOCUMENT starts with the basics of light and builds on this information to bring anunderstanding of CMOS imagers. Also included are some terms and phrases you may run
into in the digital imaging industry.
Table of Contents
LightPhotons
Color
The Human Eye
Color Mixing and Reproduction
Binary Representation of PicturesStandard Spatial Resolutions
Digital Representation of Pixels
Image SensorsCMOS vs. CCD
CMOS Imager Characteristics
Light
Visible light is the band of electromagnetic radiation that can be sensed by the human
eye. Electromagnetic radiation is the type of energy, traveling in a wave that is produced
by an oscillating charge or energy source. Electromagnetic ("EM") waves also include
radio waves, x-rays and gamma rays.
An electromagnetic wave can be defined by its wavelength (measure of lengthpeak-to-peak) or its frequency (number of cycles per second). The multiplication of these
two characteristics is a constant - the speed of light - so the two are inversely
proportional to one another. That is, the shorter the wavelength, the faster the frequency,the longer the wavelength, the slower the frequency.
OS PRIMER http://www.siliconimaging.com/ARTICLES/CMOS PRIMER.ht
12 28.9.2011 14:10
8/3/2019 Cmos Primer
2/12
Visible light is usually referred to in wavelength (instead of frequency) and includes
wavelengths of 400 nanometers (10
-9
meters, abbreviated "nm") to 750 nm. Ultravioletand infrared refer to the wavelengths just beyond the ends of the visible electromagnetic
spectrum:
Wavelength Band Type of RadiationFrequency (cycles persecond)
~10 -14 meters Gamma Rays ~10 22
~10 -9 meters(nanometer)
X Rays ~10 17
~10 -7 meters Ultraviolet ~10 15
~10 -6 meters Visible Light ~10 14
~10 -5 meters Infrared ~10 13
~10 -2 meter (centimeter) Short Wave Radio ~10 10
~1 meter TV and FM Radio ~10 8
~10 2 meters AM Radio ~10 6 (megahertz)
~10 6 meters Long Wave Radios ~10 2
Measurements approximate. The product of the wavelength times the frequency equals
3x108 meters per second, the speed of light.
Everyday usage of this chart is heard in the terms "short wave radio" and "900MHzcordless phone".
Photons
While light has properties of waves, the energy carried by light is not distributed in a
wave, but carried in discrete bundles (or "quantized"), giving light some properties likeparticles. These light "particles" are called photons, and are referred to when explaining
how light transfers energy and are used to explain how CMOS Imagers transfer light
energy to electrical energy.
OS PRIMER http://www.siliconimaging.com/ARTICLES/CMOS PRIMER.ht
12 28.9.2011 14:10
8/3/2019 Cmos Primer
3/12
Color
The Visible Light band in the EM spectrum can be broken down into a series of colors,each color corresponding to a different wavelength. The typical spectrum that is
displayed is seven colors - red, orange, yellow, green, blue, indigo, and violet. In reality
the band represents a continuum of colors, each corresponding to a differentwavelength, but seven colors are historically displayed. The bands outside this region -
ultraviolet and infrared - are said to be beyond the range of the human eye, although inexperiments both ultraviolet and infrared light can be seen unaided in certain conditions.
Visible Band
400nm 500nm 600nm
The Human Eye
The human eye can discriminate between hundreds of wavelengths as well as the
intensity of the light source being received. The ability to distinguish these characteristics
is through two main types of sensory cells in the retina:
rods - Rods convert photons into an electrical impulse that is processed by the brain.Rod cells are stimulated by the intensity of light and are responsible for perceiving the
size, shape, and brightness of visual images. They do not perceive color and fine detail;tasks performed by the other major type of light-sensitive cell, the cone. The rods are
what are in use when you are in the dark, meaning a red stop sign looks gray when you
look at one without the aid of your car's headlights.
cones - Cones are less sensitive to low illumination levels, but give us our color vision.
There are three types of cones, each of which contains a distinctive type of pigment. One
cone absorbs red light, another green, and the third type blue. A given color stimulatesall three types of receptors with varying effectiveness; the pattern of these responses
determines the color perceived. This color breakdown is explained more in color
reproduction, below.
The mnemonic to remember which cells does what is that Cones receive Color and both
start with a "C"! Rod cells are not only much more sensitive to light than cones but arealso much more numerous. The human eye contains about 130 million rods to about 7
million cones. This means that the human eye is much more sensitive to the intensity of
the light than its color. This characteristic is taken advantage of in color processing,
which will be covered later.
Color Mixing and Reproduction
This basic problem of color reproduction comes down to the question of how to create all
the colors that are possible in the color spectrum. Over the years, it has been discovered
OS PRIMER http://www.siliconimaging.com/ARTICLES/CMOS PRIMER.ht
12 28.9.2011 14:10
8/3/2019 Cmos Primer
4/12
that all the colors in the spectrum can be recreated from only a sub sample of only three
other colors by mixing them in varying degrees.
The fact that the whole range of colors may be synthesized from only three primary
colors is essentially a description of the process by which the eye processes colors. Thisis a fortunate property of vision since it allows three colors to represent any of the
10,000 or more colors (and brightness) that may be distinguished by human vision. If
this was not the case and vision was dependent on the energy and wavelength
relationship of light described above, it is doubtful that color reproduction could beincorporated in any mass-communication system.
The three main ways to reproduce color are as follows:
Primary Colors - Red, Green, Blue or "RGB" - Most people remember this from their childhood art
classes. This is an "additive" method of adding the three primary colors in different amounts to
recreate other colors, usually used in systems that project light. A mixture of these three primaries
- red, green, and blue - may match any other color if their relative intensities are varied. White is
made by adding ALL the colors (remember that "white light" represents the entire visible EM
spectrum). The RGB scheme is used by televisions, computer monitors and other devices that
project with light.
1.
Complementary Colors - Cyan, Magenta, Yellow or "CMY" - This method is "subtractive" and is
primarily used in printing since ink pigment "subtracts" the light falling on it. For example, a yellow
pigment "absorbs" blue and violet light and "reflects" yellow (along with green and blue which
together make more yellow). Since RGB is the best method for adding colors, then "negative red",
or a pigment which absorbs the most red, "negative green", or a pigment which absorbs the most
green, and "negative blue", which absorbs the most blue, are the best colors for subtracting. These
colors, respectively, are Cyan, Magenta and Yellow. This method is used in inkjet printers and other
methods that print (rather than project, which uses RGB). In practice, most inkjet printers not only
use C, M, and Y ink, but also black ink since black, the combination of all the colors, would use up
those inks very quickly. Since "B" already means "Blue", the last letter of the word "black" is used,meaning this method is referred to as "CMYk".
Subtractive filters are used in consumer cameras since they absorb less light. Professional cameras
use additive filters since additive produces more accurate color.
2.
YCRCB - Luminance, Chrominance (Red), Chrominance (Blue) - The third way to characterize l ight
makes use of the RGB concept above, but breaks down the color components in different fashion.Any color can be broken down into two qualities:
3.
Luminance - Its brightness or intensity. Remember that the human eye that is more
sensitive to brightness than to color. The luminance value, stated with the letter "Y", is
1.
OS PRIMER http://www.siliconimaging.com/ARTICLES/CMOS PRIMER.ht
12 28.9.2011 14:10
8/3/2019 Cmos Primer
5/12
the brightness breakdown of the color.
Chrominance - the color "remaining" once luminance is removed. This can be broken
down again into two independent qualities:
2.
Hue - This is the color of the light, in this case red or blue.
Saturation - the relative purity, or the amount of white light mixed with a hue. High
saturation colors contain little or no white light.
The translation from RGB to YCRCB is done with a "look-up table" which takes any RGB value and
matches it to its corresponding YCRCB components.
Binary Representation of Pictures
The fact that color can be broken down into individual components is extremely
important to digital imaging - the process of breaking down a picture into the "1"s and
"0" of digital communications. The process of breaking down a picture into individual
components can be done in two basic steps:
Breaking down the picture into a pixel grid - For a picture to be described as a series of1s and 0s, it first must be broken down into a grid, or array. This process simply places a
grid over a picture and assigns a single color for each square in the grid. This single color
grid square is called a "pixel" (short for picture element).
The number of pixels used for the picture breakdown is called the "spatial resolution"
and is usually referred to by its horizontal and vertical number of pixels, such as"640x480", meaning 640 pixels horizontally and 480 pixels vertically.
For a given picture, the number of pixels will determine the quality of the digital picture.That is, the smaller number of pixels, the larger they must be and the lower the picture
quality. The higher number of pixels, the smaller each pixel is and the better the picturequality:
Low Spatial
Resolution
High Spatial
Resolution
Large Pixel Size
Fewer PixelsLow Picture Quality
Small Pixel Size
More PixelsHigh Picture Quality
OS PRIMER http://www.siliconimaging.com/ARTICLES/CMOS PRIMER.ht
12 28.9.2011 14:10
8/3/2019 Cmos Primer
6/12
Standard Spatial Resolutions
There are a number of standard resolutions, or arrays, in the sensor industry. Most of
these formats come from the display monitor industry, which drives the number of pixels
you see on computer monitors. Since sensors typically display on monitors, they
commonly match monitor resolutions. Term one will hear in regards to spatial format
include:
CIF - Common Intermediate Format - 352 x 288 pixels for a total of 101,376pixels (commonly referred rounded at 100,000 pixels). This format was
developed for PC video conferencing. The number of pixels is fairly small, butwas needed in order to get full motion video at 30 frames per second.
QCIF - Quarter CIF - One quarter of a CIF format, so 176x144 for a total of
about 25,000.
VGA - Video Graphics Array - 640x480 pixels for a total of 307,200 pixels. TheVGA format was developed for computer monitors by IBM and become the
standard for monitors for many years. Although monitor resolutions today are
higher, VGA is still lowest "common" display which all PCs will support.
SVGA - Super VGA - 800x600 pixels for a total of 480,000 pixels. The next
highest monitor resolution developed for PCs.
XGA - "Xtended" Graphics Array - 1024x768 for a total of 786,432 pixels.
Another monitor standard.
If a sensor is not one of these standards, its resolution is simply displayed as vertical byhorizontal (200x300, for example). Typically, if a sensor has more than 1 million total
pixels (anything more than 1000x1000 pixels), it is termed a "megapixel" sensor, which
has come to mean any sensor with more than one million pixels.
Digital Representation of Pixels
Now that the picture is represented as an array of pixels, each pixel needs to be
described digitally. To do this, each pixel is assigned two main components: its location in
the picture and its color. Its location is usually just represented by its "x and y"coordinate in the grid. Its color is represented by its color resolution, which is the
method of describing a color digitally.
Using the RGB method of color representation, a color can be divided into an arbitrary
number of levels of that color. For example, red can be broken down from total red to no
red (or white):
No Red Total Red
Each step in the arbitrary breakdown is called a "gray level" (even though the color isnot gray). The same breakdown can be done for green and blue.
OS PRIMER http://www.siliconimaging.com/ARTICLES/CMOS PRIMER.ht
12 28.9.2011 14:10
8/3/2019 Cmos Primer
7/12
By experiment, the naked eye can distinguish about 250 shades of each color. Using
binary math, the closest binary number is 256, which is 28, gray levels can be used for
each color. This means for each color component of a picture, there are 8-bits used foreach R, G, B element, for a total of 24 bits of color representation. The complete R, G,
breakdown of 224 colors represents about 16.7 million colors that can be represented
digitally. The number of colors represented by a pixel is called its "tonal resolution" orits "color dynamic range". If fewer bits are used, the number of colors represented is
smaller, so its dynamic range is smaller.
Image Sensors
Image sensors are devices that take an image and directly convert it to a digital image.Referred to in marketing literature as "silicon firm" or "silicon eyes", these devices are
made of silicon since silicon has the properties of both being sensitive to light in the
visible spectrum and being able to have circuitry integrated on-board. Silicon imagesensors come in two broad classes:
Charge-Coupled Devices (CCD) - Currently the most commonly used image sensor,
CCDs capture light onto an array of light-sensitive diodes, each diode representing onepixel. For color imagers, each pixel is coated with a film of red, green, or blue (or
complementary color scheme) so that each particular pixel captures that one particularcolor.
The pixel, made up of a light sensitive diode, converts the light photon into a charge, and
the value of that charge is moved to a single location in a manner similar to a row ofpeople passing buckets of water. At the end, the charge is amplified. Since this "bucket
brigade" is accomplished by applying different voltages to the pixels in a succession, the
process is called charge-coupling. Because the value in the pixel is moved by applyingdifferent voltages, CCD sensors must be supported by several external voltage
generators. In addition, CCDs require a specialized manufacturing process that cannot be
used by any other device.
Graphical representation
of CCD
Image source: Digital Photography
Review
CMOS Imagers - Like CCDs, these imagers are made from silicon, but as the name
implies, the process they are made in is called CMOS, which stands for Complementary
OS PRIMER http://www.siliconimaging.com/ARTICLES/CMOS PRIMER.ht
12 28.9.2011 14:10
8/3/2019 Cmos Primer
8/12
Metal Oxide Semiconductor. This process is today the most common method of making
processors and memories, meaning CMOS Imagers take advantage of the process and
cost advancements created by these other high-volume devices.
Like CCDs, CMOS imagers include an array of photo-sensitive diodes, one diode withineach pixel. Unlike CCDs, however, each pixel in a CMOS imager has its own individual
amplifier integrated inside. Since each pixel has its own amplifier, the pixel is referred to
as an "active pixel". (note: There are also "passive pixel sensors" (pps) that do not
contain this amplifier). In addition, each pixel in a CMOS imager can be read directly onan x-y coordinate system, rather than through the "bucket-brigade" process of a CCD.This means that while a CCD pixel always transfers a charge, a CMOS pixel always
detects a photon directly, converts it to a voltage and transfers the information directlyto the output. This fundamental difference in how information is read out of the imager,
coupled with the manufacturing process, gives CMOS Imagers several advantages over
CCDs.
CMOS Sensor Array
CMOS vs. CCD
Due to both design and manufacturing considerations, there are a number of advantages
that CMOS Imagers have over CCD:
Integration - Because CMOS Imagers are created in the same process as
processors, memories and other major components, CMOS Imagers can integrated
with these same components onto a single piece of silicon. In contrast, CCDs aremade in a specialized process and require multiple clocks and inputs. This feature
limits CCDs to discrete systems, which in the long run will put CMOS Imagers at a
cost advantage, as well as limit what kinds of portable devices CCDs can beintegrated into.
1.
OS PRIMER http://www.siliconimaging.com/ARTICLES/CMOS PRIMER.ht
12 28.9.2011 14:10
8/3/2019 Cmos Primer
9/12
Reduced Power Consumption - because of all the external clocks needed to "bucketbrigade" each pixel, CCDs are inherently power hungry. Every clock is essentiallycharging and discharging large capacitors in the CCD array. In contrast CMOS
imagers require only a single voltage input and clock, meaning they consume muchless power than CCDs, a feature that is critical for portable, battery operated
devices.
Pixel Addressibility - CCDs use of the bucket brigade to transfer pixel values means
that individual pixels in a CCD cannot be read individually. CMOS imagers on the
other hand have the pixels in an x-y grid allowing pixels to be read individually. This
means that CMOS imagers will be able to do functions such as "windowing", whereonly a small sample of the imager is read, image stabilization to remove jitters from
camcorders, motion tracking and other advanced imaging techniques internally that
CCDs cannot do.
2.
Manufacturing Cost - Since CMOS imagers are manufactured in the same process as
memories, processors and other high-volume devices, CMOS imagers can take
advantage of process improvements and cost reductions these devices drivethroughout the industry.
3.
CMOS Imager Characteristics
OS PRIMER http://www.siliconimaging.com/ARTICLES/CMOS PRIMER.ht
12 28.9.2011 14:10
8/3/2019 Cmos Primer
10/12
There are a number of phrases and terms for describing the functional capability, physical
features or competitive characteristics of an imager:
Active Pixel Sensor (also APS) - As explained above, an active CMOS Imager pixel has
its own amplifier for boosting the pixel's signal. Active Pixels are the dominant type ofCMOS Imagers in the commercial market today. The other type of CMOS Imager, a
passive pixel sensor (PPS), consists of only the photo detector without a local amplifier.
While very sensitive to low light conditions, these types of sensors are not suitable for
commercial applications due to their high amount of noise and poor picture quality whencompared to active pixels.
Fill Factor - The amount of a CMOS Pixel that is actually capturing light. In an activepixel, both the photo detector and the amplifier take up "real estate" in the pixel. The
amplifier is not sensitive to light, so this part of the pixel area is lost when taking a
picture.
The fill factor is simply the percentage of the area of the pixel that is sensitive to light. In
the picture above, this is about 40%. As semiconductor process technologies get smallerand smaller, the amount of area taken up by the amplifier is taking up less space, so low
fill factors are becoming less of an issue with active pixels. Note that in passive pixels -
where there is no amplifier at all - fill factors typically reach over 80%. The reason they
do not reach 100% is due to routing and pixel selection circuitry that are also needed in a
CMOS imager.
Microlenses - In some pixel designs, the fill factor becomes too small to be effective.
For example, if a fill factor in an imager were 25%, this would mean that 75% of the lightfalling on a pixel would be lost, reducing the pixel's capability. To get around thissituation, some CMOS imagers have small lenses manufactured directly above the pixel
to focus the light towards the active portion that would otherwise fall on the non-lightsensitive portion of the pixel. Microlenses typically can increase the effective fill factor by
two to three times.
OS PRIMER http://www.siliconimaging.com/ARTICLES/CMOS PRIMER.ht
f 12 28.9.2011 14:10
8/3/2019 Cmos Primer
11/12
Color Filter Array (also CFA or just "color filter") - CMOS Pixels are sensitive to light
photons but are not, by themselves, sensitive to color. Unaided, the pixels will capture
any kind of light, creating a black and white image. In order to distinguish betweencolors, filters are put on top of a pixel to allow only certain colors to pass, turning the
"rods" of the array into "cones". Since all colors can be broken down into an RGB orCMYk pattern, individual primary or complementary color schemes are deposited on top
of the pixel array. After being read from the sensor, software takes the different values ofthe pattern and recombines the colors to match the original picture. There are a variety
of different filters, the most popular being the Bayer Filter Pattern (also known as
RGBG). Note the large amount of green in the pattern, due to the fact that the eye ismost sensitive to color in the green part of the spectrum.
Bayer Color FilterPattern
Noise - The same as static in a phone line or "snow" in a television picture, noise is any
unwanted electrical signal that interferes with the image being read and transferred bythe imager. There are two main types of noise associated with CMOS Sensors:
Read Noise (also called temporal noise) - This type of noise occurs randomly and isgenerated by the basic noise characteristics of electronic components. This type of
noise looks like the "snow" on a bad TV reception.
1.
Fixed Pattern Noise (also FPN) - This noise is a result of each pixel in an imager
having its own amplifier. Even though the design of each amplifier is the same, whenmanufactured, these amplifiers may have slightly different offset and gain
characteristics. This means for any picture given, if certain pixels are boosting thesignal for every picture taken, they will create the same pattern again and again,
hence the name.
2.
Blooming - The situation where too many photons are being produced to be received by
a pixel. The pixel overflows and causes the photons to go to adjacent pixels. Blooming is
similar to overexposure in film photography, except that in digital imaging, the result is a
OS PRIMER http://www.siliconimaging.com/ARTICLES/CMOS PRIMER.ht
f 12 28.9.2011 14:10
8/3/2019 Cmos Primer
12/12
number of vertical and/or horizontal streaks appearing from the light source in the
picture.
This photo illustrates two undesirable characteristics:
blooming, the slight vertical line running from the topto the bottom of the picture and lens flare, the star
shape light which is a function of the lens and not the
imager.
Optical Format - is a number in inches that is calculated by taking the diagonal
measurement of a sensor array in millimeters and dividing by 16. For example, a CMOS
Imager that has a diagonal measurement of 4mm has an optical format of 4/16, or ".
What Optical Format calculates is the type of lens system that must be used with theimager. In the lens industry, there are standard sets of ", ", ", etc. lens systems. By
using Optical Format, a user of imagers can use standard, mass-produced (andinexpensive) lens systems rather than having to design and custom build a special lens
system. The terms and measurement comes from the days of electron tubes and
pre-dates solid-state electronics. Generally speaking, larger optics are more expensive,so a " lens system is less than a 1/3" lens system.
Aspect Ratio - The ratio between the height and width of a sensor or display. It is found
by dividing the vertical number of pixels (height) by the horizontal number of pixels
(width) leaving it in fractional format.
For example, a pixel with resolution of 640x480 would have an aspect ration of 480/640=
.
The most common aspect ratios are and 9/16. The aspect ratio is the ratio for
computer monitors and TVs. The newer 9/16 aspect ratio is used for High Definition
Television (HDTV)
Quantum Efficiency (or QE) - Imagers create digital images by converting photonenergy to electrical energy. The efficiency in which each photon is converted to an
electron is the imager's quantum efficiency. The number is calculated by simply dividing
electrons by photons, or E/P. If no electrons are created, the efficiency is obviously zero,
while if each photon creates one electron the efficiency is 100%. Typically, a sensor has
different efficiency at different light frequencies, so a graph of the quantum efficiencyover the different wavelengths is typically shown:
Dark Current - A situation in CMOS imagers where the pixels fill with thermally created
electrons without any illumination. This problem is a function of the manufacturing
process and layout and increases with increasing temperature.
OS PRIMER http://www.siliconimaging.com/ARTICLES/CMOS PRIMER.ht