Solutions Manual for Digital Image Processing
and Analysis Computer Vision and Image
Analysis, 4e by Scott Umbaugh (All Chapters)
Solutions for Chapter 1: Digital Image Processing and Analysis
1. Digital image processing is also referred to as computer imaging and can be defined as the
acquisition and processing of visual information by computer. It can be divided into application
areas of computer vision and human vision; where in computer vision applications the end user
is a computer and in human vision applications the end user is a human. Image analysis ties these
two primary application areas together, and can be defined as the examination of image data to
solve a computer imaging problem. A computer vision system can be thought of as a deployed
image analysis system.
2. In general, a computer vision system has an imaging device, such as a camera, and a computer
running analysis software to perform a desired task. Such as: A system to inspect parts on an
assembly line. A system to aid in the diagnosis of cancer via MRI images. A system to
automatically navigate a vehicle across Martian terrain. A system to inspect welds in an
automotive assembly factory.
3. The image analysis process requires the use of tools such as image segmentation, image
transforms, feature extraction and pattern classification. Image segmentation is often one of the
first steps in finding higher level objects from the raw image data. Feature extraction is the
process of acquiring higher level image information, such as shape or color information, and may
Downloaded by: tutorsection | Want to earn $1.236
Distribution of this document is illegal extra per year?
, 3
require the use of image transforms to find spatial frequency information. Pattern classification
is the act of taking this higher level information and identifying objects within the image.
4. hardware and software.
5. Gigabyte Ethernet, USB 3.2, USB 4.0, Camera Link.
6. It samples an analog video signal to create a digital image. This sampling is done at a fixed
rate when it measures the voltage of the signal and uses this value for the pixel brightness. It uses
the horizontal synch pulse to control timing for one line of video (one row in the digital image),
and the vertical synch pulse to tell the end of a field or frame
7. A sensor is a measuring device that responds to various parts of the EM spectrum, or other
signal that we desire to measure. To create images the measurements are taken across a two-
dimensional gird, thus creating a digital image.
8. A range image is an image where the pixel values correspond to the distance from the imaging
sensor. They are typically created with radar, ultrasound or lasers.
9. The reflectance function describes the way an object reflects incident light. This relates to
what we call color and texture it determines how the object looks.
10. Radiance is the light energy reflected from, or emitted by, an object; whereas irradiance is
the incident light falling on a surface. So radiance is measured in Power/(Area)(SolidAngle), and
irradiance is measure in Power./Area.
11. A photon is a massless particle that is used to model EM radiation. A CCD is a charge-
coupled device. Quantum efficiency is a measure of how effectively a sensing element converts
photonic energy into electrical energy, and is given by the ratio of electrical output to photonic
input.
Downloaded by: tutorsection | Want to earn $1.236
Distribution of this document is illegal extra per year?
, 4
1 1 1
12. See fig 1.4-5 and use the lens equation: + = . If the object is at infinity:
a b f
1 1 1 1 1
+ = = 0 + = ; f = b
b f b f
Downloaded by: tutorsection | Want to earn $1.236
Distribution of this document is illegal extra per year?
, 5
13. Yes, solid state because the quantum efficiency is 95%
N = At b( )q( )d
700
= 20 (10 x10 −3 ) 600 (0.95)d
400
700
700 2
= 114 d = 114 = 114 (165,000 ) = 1.881x10 7
400 2 400
14. With interlaced scanning, a frame in 1/30 of a second is a field rate of 1/60 of a second. Here
we 240 lines per field and 640 pixels per line which gives (240)(640) = 153,600 pixels in 1/60 of
a second. So the sampling rate must be:
153,600 pixels
= 9.216 x10 6 ; or about 9 megahertz
1 sec
60
15. Gamma rays have the most energy, radio waves have the least. For human life, more energy
is more dangerous.
16. UV is used in fluorescence microscopy, and IR images are used in remote sensing, law
enforcement, medical thermography and fire detection.
17. Acoustic imaging works by sending out pulses of sonic energy (sound) at various
frequencies, and then measuring the reflected waves. The time it takes for the reflected signal to
appear contains distance information, and the amount of energy reflected contains information
about the object’s density and material. The measured information is then used to create a two or
three dimensional image. It is used in geological applications, for example oil and mineral
exploration, typically use low frequency sounds (around hundreds of hertz). Ultrasonic, or high
frequency sound, imaging is often used in manufacturing to detect defects and in medicine to
“see” inside opaque objects such as a woman’s womb to image a developing baby.
Downloaded by: tutorsection | Want to earn $1.236
Distribution of this document is illegal extra per year?