Documente Academic
Documente Profesional
Documente Cultură
UNIT I
DIGITAL IMAGE FUNDAMENTALS
1.1. Define Image.
An Image may be defined as a two dimensional function f(x,y) where x & y are spatial
(plane) coordinates, and the amplitude of f at any pair of coordinates (x,y) is called intensity or
gray level of the image at that point. When x,y and the amplitude values of f are all finite,
discrete quantities we call the image as Digital Image.
1.2. Define Image Sampling.
Digitization of spatial coordinates (x,y) is called Image Sampling. To be suitable for
computer processing, an image function f(x,y) must be digitized both spatially and in
magnitude.
1.3. Define Quantization.
Digitizing the amplitude values is called Quantization. Quality of digital image is
determined to a large degree by the number of samples and discrete gray levels used in sampling
and quantization.
1.4. What is Dynamic Range?
The range of values spanned by the gray scale is called dynamic range of an image.
Image will have high contrast, if the dynamic range is high, and image will have dull washed
out gray look if the dynamic range is low.
1.5. Define Mach band effect.
The spatial interaction of Luminance from an object and its surround creates a
phenomenon called the mach band effect.
1.6. Define Brightness.
Brightness of an object is the perceived luminance of the surround. Two objects with
different surroundings would have identical luminance but different brightness.
1.7. What is meant by Tapered Quantization?
If gray levels in a certain range occur frequently while others occurs rarely, the
quantization levels are finely spaced in this range and coarsely spaced outside of it. This method
is sometimes called Tapered Quantization.
1.8. What do you meant by Gray level?
Gray level refers to a scalar measure of intensity, that ranges from black to grays and
finally to white.
Page 1
Page 2
1.16. Write the expression to find the number of bits to store a digital image?
The number of bits required to store a digital image is
b=M X N X k
When M=N, this equation becomes
b=N^2k
1.17. What is meant by pixel?
A digital image is composed of a finite number of elements, each of which has a
particular location of value. These elements are referred to as pixels or image elements or
picture elements or pels elements.
1.18. Define digital image.
A digital image is an image f(x,y), that has been discretized both in spatial coordinates
and brightness.
1.19. List the steps involved in digital image processing.
The steps involved in digital image processing are,
i.Image Acquisition.
ii.Preprocessing.
iii.Segmentation.
iv.Representation and description.
v.Recognition and interpretation.
1.20. What is recognition and interpretation?
Recognition is a process that assigns a label to an object based on the information
provided by its descriptors. Interpretation means assigning to a recognized object.
1.21. Specify the elements of DIP system.
The elements of DIP system are,
i.Image acquisition.
ii.Storage.
iii.Processing.
iv.Communication.
v.Display.
1.22. List the categories of digital storage.
The categories of digital storage are,
i.Short term storage for use during processing.
ii.Online storage for relatively fast recall.
iii. Archical storage for frequent access.
1.23. Write the two types of light receptors.
The two types of light receptors are,
i.Cones.
ii.Rods.
Page 3
Scotopic vision
Several rods are connected to one nerve
end. So it gives the overall picture of the
image.
This is also known as dim light vision.
million.
Page 4
Page 5
WN^(K+N/2)= -WN^K
1.40. Give the Properties of one-dimensional DFT.
The properties of one-dimensional DFT are,
i. The DFT and unitary DFT matrices are symmetric.
ii. The extensions of the DFT and unitary DFT of a sequence and their inverse transforms
are periodic with period N.
iii. The DFT or unitary DFT of a real sequence is conjugate symmetric about N/2.
1.41. Give the Properties of two-dimensional DFT.
The properties of two-dimensional DFT are,
i.Symmetric
ii.Periodic extensions
iii.Sampled Fourier transform
iv.Conjugate symmetry.
1.42. Write the properties of cosine transform:
The properties of cosine transform are as follows,
i.Real & orthogonal.
ii.Fast transform.
iii.Has excellent energy compaction for highly correlated data
1.43. Write the properties of Hadamard transform
The properties of hadamard transform are as follows,
i.Hadamard transform contains any one value.
ii.No multiplications are required in the transform calculations.
iii.The no: of additions or subtractions required can be reduced from N 2 to about Nlog2N
iv.Very good energy compaction for highly correlated images.
1.44. Define Haar transform.
The Haar functions are defined on a continuous interval Xe [0,1] and for K=0,1, N1,where N=2^n..The integer k can be uniquely decomposed as K=2^P+Q-1.
1.45. Write the properties of Haar transform.
The properties of Haar transform are,
i. Haar transform is real and orthogonal.
ii. Haar transform is a very fast transform
iii. Haar transform has very poor energy compaction for images
iv. The basic vectors of Haar matrix sequensly ordered.
1.46.Write the Properties of Slant transform.
The properties of Slant transform are,
i.Slant transform is real and orthogonal.
ii.Slant transform is a fast transform
iii.Slant transform has very good energy compaction for images
iv.The basic vectors of Slant matrix are not sequensely ordered.
Page 6
Page 7
UNIT II
IMAGE ENHANCEMENT
2.1. What is Image Enhancement?
Image enhancement is to process an image so that the output is more suitable
for specific application.
2.2. Name the categories of Image Enhancement and explain?
The categories of Image Enhancement are,
i.Spatial domain
ii.Frequency domain
Spatial domain: It refers to the image plane and it is based on direct
manipulation of pixels of an image.
Frequency domain techniques are based on modifying the Fourier transform of an image.
2.3. What do you mean by Point processing?
Image enhancement at any Point in an image depends only on the gray level at
that point is often referred to as Point processing.
2.4.What is gray level slicing?
Highlighting a specific range of gray levels in an image is referred to as gray level
slicing. It is used in satellite imagery and x-ray images
2.5. What do you mean by Mask or Kernels.
A Mask is a small two-dimensional array, in which the value of the mask
coefficient determines the nature of the process, such as image sharpening.
2.6. What is Image Negative?
The negative of an image with gray levels in the range [0, L-1] is obtained by
using the negative transformation, which is given by the expression.
s = L-1- r ,Where s is output pixel, r is input pixel
2.7. Define Histogram.
The histogram of a digital image with gray levels in the range [0, L-1] is a discrete
function h (rk) = nk, where rk is the kth gray level and nk is the number of pixels in the image
having gray level rk.
2.8.What is histogram equalization
It is a technique used to obtain linear histogram . It is also known as histogram
linearization. Condition for uniform histogram is Ps(s) = 1
2.9. What is contrast stretching?
Contrast stretching reduces an image of higher contrast than the original, by darkening the
levels below m and brightening the levels above m in the image.
Prepared by A.Devasena., Associate Professor., Dept/ECE
Page 8
Page 9
UNIT III
IMAGE RESTORATION
3.1. Define Restoration.
Restoration is a process of reconstructing or recovering an image that has been
degraded by using a priori knowledge of the degradation phenomenon. Thus restoration
techniques are oriented towards modeling the degradation and applying the inverse process in
order to recover the original image.
3.2. How a degradation process id modeled?
A system operator H, which together with an additive white noise term n(x,y) a operates on
an input image f(x,y) to produce a degraded image g(x,y).
3.3. What is homogeneity property and what is the significance of this property?
Homogeneity property states that
H [k1f1(x,y)] = k1H[f1(x,y)]
Where H=operator
K1=constant
f(x,y)=input image.
It says that the response to a constant multiple of any input is equal to the response to that
input multiplied by the same constant.
3.4.What is meant by image restoration?
Restoration attempts to reconstruct or recover an image that has been degraded by using a
clear knowledge of the degrading phenomenon.
3.5. Define circulant matrix.
A square matrix, in which each row is a circular shift of the preceding row and the first
row is a circular shift of the last row, is called circulant matrix.
3.6. What is the concept behind algebraic approach to restoration?
Algebraic approach is the concept of seeking an estimate of f, denoted f^, that
minimizes a predefined criterion of performance where f is the image.
3.7. Why the image is subjected to wiener filtering?
This method of filtering consider images and noise as random process and the objective
is to find an estimate f^ of the uncorrupted image f such that the mean square error between
them is minimized. So that image is subjected to wiener filtering to minimize the error.
3.8. Define spatial transformation.
Spatial transformation is defined as the rearrangement of pixels on an image plane.
3.9. Define Gray-level interpolation.
Prepared by A.Devasena., Associate Professor., Dept/ECE
Page 10
Gray-level interpolation deals with the assignment of gray levels to pixels in the spatially
transformed image.
3.10. Give one example for the principal source of noise.
The principal source of noise in digital images arise image acquisition (digitization)
and/or transmission. The performance of imaging sensors is affected by a variety of factors,
such as environmental conditions during image acquisition and by the quality of the sensing
elements. The factors are light levels and sensor temperature.
3.11. When does the degradation model satisfy position invariant property?
An operator having input-output relationship g(x,y)=H[f(x,y)] is said to position invariant
if H[f(x-,y-_)]=g(x-,y-_) for any f(x,y) and and _. This definition indicates that the response at
any point in the image depends only on the value of the input at that point not on its position.
3.12. Why the restoration is called as unconstrained restoration?
In the absence of any knowledge about the noise n, a meaningful criterion function is
to seek an f^ such that H f^ approximates of in a least square sense by assuming the noise term
is as small as possible.
Where H = system operator.
f^ = estimated input image.
g = degraded image.
3.13. Which is the most frequent method to overcome the difficulty to formulate the
spatial relocation of pixels?
The point is the most frequent method, which are subsets of pixels whose location in
the input (distorted) and output (corrected) imaged is known precisely.
3.14. What are the three methods of estimating the degradation function?
The three methods of degradation function are,
i.Observation
ii.Experimentation
iii.Mathematical modeling.
3.15. How the blur is removed caused by uniform linear motion?
An image f(x,y) undergoes planar motion in the x and y-direction and x0(t) and y0(t)
are the time varying components of motion. The total exposure at any point of the recording
medium (digital memory) is obtained by integrating the instantaneous exposure over the time
interval during which the imaging system shutter is open.
3.16. What is inverse filtering?
The simplest approach to restoration is direct inverse filtering, an estimate F^(u,v) of
thetransform of the original image simply by dividing the transform of the degraded image
G^(u,v) by the degradation function.
F^ (u,v) = G^(u,v)/H(u,v)
Page 11
Page 12
UNIT IV
DATA COMPRESSION
4.1. What is Data Compression?
Data compression requires the identification and extraction of source redundancy. In
other words, data compression seeks to reduce the number of bits used to store or transmit
information.
4.2. What are two main types of Data compression?
The two main types of data compression are lossless compression and lossy
compression.
4.3.What is Lossless compression?
Lossless compression can recover the exact original data after compression. It is used
mainly for compressing database records, spreadsheets or word processing files, where exact
replication of the original is essential.
4.4. What is lossy compression?
Lossy compression will result in a certain loss of accuracy in exchange for a substantial
increase in compression. Lossy compression is more effective when used to compress graphic
images and digitized voice where losses outside visual or aural perception can be tolerated.
4.5. What is the Need For Compression?
In terms of storage, the capacity of a storage device can be effectively increased with
methods that compress a body of data on its way to a storage device and decompresses it when it
is retrieved. In terms of communications, the bandwidth of a digital communication link can be
effectively increased by compressing data at the sending end and decompressing data at the
receiving end.
At any given time, the ability of the Internet to transfer data is fixed. Thus, if data can
effectively be compressed wherever possible, significant improvements of data throughput can
be achieved. Many files can be combined into one compressed document making sending easier.
4.6. What are different Compression Methods?
The different compression methods are,
i.Run Length Encoding (RLE)
ii.Arithmetic coding
iii.Huffman coding and
iv.Transform coding
4.7. What is run length coding?
Run-length Encoding, or RLE is a technique used to reduce the size of a repeating
string of characters. This repeating string is called a run; typically RLE encodes a run of
symbols into two bytes, a count and a symbol. RLE can compress any type of data regardless
Prepared by A.Devasena., Associate Professor., Dept/ECE
Page 13
of its information content, but the content of data to be compressed affects the compression
ratio. Compression is normally measured with the compression ratio:
4.8. Define compression ratio.
Compression ratio is defined as the ratio of original size of the image to compressed size of the
image .It is given as
Compression Ratio = original size / compressed size: 1
4.9. Give an example for Run length Encoding.
Consider a character run of 15 'A' characters, which normally would require 15 bytes to
store: AAAAAAAAAAAAAAA coded into 15A With RLE, this would only require two bytes
to store; the count (15) is stored as the first byte and the symbol (A) as the second byte.
4.10. What is Huffman Coding?
Huffman compression reduces the average code length used to represent the symbols of
an alphabet. Symbols of the source alphabet, which occur frequently, are assigned with short
length codes. The general strategy is to allow the code length to vary from character to
character and to ensure that the frequently occurring characters have shorter codes.
4.11. What is Arithmetic Coding?
Arithmetic compression is an alternative to Huffman compression; it enables characters
to be represented as fractional bit lengths. Arithmetic coding works by representing a number
by an interval of real numbers greater or equal to zero, but less than one. As a message
becomes longer, the interval needed to represent it becomes smaller and smaller, and the
number of bits needed to specify it increases.
4.12. What is JPEG?
The acronym is expanded as "Joint Photographic Expert Group". It is an international
standard in 1992. It perfectly Works with colour and greyscale images, Many applications e.g.,
satellite, medical,...
4.13. What are the basic steps in JPEG?
The Major Steps in JPEG Coding involve:
i. DCT (Discrete Cosine Transformation)
ii. Quantization
iii. Zigzag Scan
iv. DPCM on DC component
v.RLE on AC Components
vi. Entropy Coding
4.14. What is MPEG?
The acronym is expanded as "Moving Picture Expert Group". It is an international
standard in 1992. It perfectly Works with video and also used in teleconferencing
4.15. What is transform coding?
Transform coding is used to convert spatial image pixel values to transform coefficient
values. Since this is a linear process and no information is lost, the number of coefficients
Prepared by A.Devasena., Associate Professor., Dept/ECE
Page 14
produced is equal to the number of pixels transformed. The desired effect is that most of the
energy in the image will be contained in a few large transform coefficients. If it is generally the
same few coefficients that contain most of the energy in most pictures, then the coefficients
may be further coded by loss less entropy coding. In addition, it is likely that the smaller
coefficients can be coarsely quantized or deleted (lossy coding) without doing visible damage
to the reproduced image.
4.16.Define I-frame.
I-frame is intraframe or independent frame. An I-frame is compressed independently of all
frames. It resembles a JPEG encoded image. It is the reference point for the motion estimation
needed to generate subsequent P and P-frame.
4.17. Define p-frame.
P-frame is called predictive frame. A P-frame is the compressed difference between the
current frame and a prediction of it based on the previous I or P-frame.
4.18. Define B-frame.
B-frame is the bidirectional frame. A B-frame is the compressed difference between the
current frame and a prediction of it based on the previous I or P frame or next P-frame.
Accordingly the decoder must have access to both past and future reference frames.
4.19. Draw the JPEG Encoder.
Page 15
Page 16
UNIT V
IMAGE SEGMENTATION
5.1. What is segmentation?
The first step in image analysis is to segment the image. Segmentation subdivides an
image into its constituent parts or objects.
5.2. Write the applications of segmentation.
The applications of segmentation are,
i. Detection of isolated points.
ii. Detection of lines and edges in an image.
5.3. What are the three types of discontinuity in digital image?
Three types of discontinuity in digital image are points, lines and edges.
5.4. How the discontinuity is detected in an image using segmentation?
The steps used to detect the discontinuity in an image using segmentation are
i.Compute the sum of the products of the coefficient with the gray levels contained inthe
region encompassed by the mask.
ii.The response of the mask at any point in the image is R = w1z1+ w2z2 + w3z3 +..+
w9z9
iii.Where zi = gray level of pixels associated with mass coefficient w i.
iv.The response of the mask is defined with respect to its center location.
5.5. Why edge detection is most common approach for detecting discontinuities?
The isolated points and thin lines are not frequent occurrences in most practical
applications, so edge detection is mostly preferred in detection of discontinuities.
5.6. How the derivatives are obtained in edge detection during formulation?
The first derivative at any point in an image is obtained by using the magnitude of the
gradient at that point. Similarly the second derivatives are obtained by using the laplacian.
5.7. Write about linking edge points.
The approach for linking edge points is to analyse the characteristics of pixels in a
small neighborhood (3x3 or 5x5) about every point (x,y)in an image that has undergone edge
detection. All points that are similar are linked, forming a boundary of pixels that share some
common properties.
Page 17
Page 18
i.Preprocessing.
ii.Definition of a set of criteria that markers must satisfy.
Page 19
ii.Perimeter.
iii.Mean and median gray levels
iv.Minimum and maximum of gray levels.
v.Number of pixels with values above and below mean.
5.26. Define texture.
Texture is one of the regional descriptors. It provides measure measures of properties such
as smoothness, coarseness and regularity.
Page 20