You are on page 1of 8

Nofe Aftempt all questions.

The question paper contains 70 MCQ ype quesio


to
question carries equal marks. Select corresponding
the answer and fill the bubble
that question in the attached OMR sheet.

. Data compression is a_ aspect Ratio


in (A)
the number of bits needed to
compression ratio
(B).
represent data. bit ratio
(C)
(A) incrememt (D) none of the above
and a
the model
(B) reduction 6. A description of
data differ
(C) expansion descriptionof how the

(D) none of the above encoded is called


from the model are

2. Data compression schemes can be (A) modelling

divided into broad classes. (B) testing


(A) 4 (C) coding
(B) 3
(D) maintenance

C 2 7. The average number of bits required


known
(D) 5 to represent single sample is
3. Text compression is an example of as
compression. bit rate
(A)
(A lossless compression (B) byte rate

(B) lossy compression (C) compression rate

(C) both (A) and (B) (D) all of the above

D) none of the above 8. The difference between the orginal


Video compression is an example and reconstruction is often called

of compression (A) waveform

(A) lossless compression B distortion

(B lossy compression (C) noise

C) both (A) and (B) (D) error

(D) None of the above The compression of analog signals is


The ratio of the number of bits referred to as

required to represent data before (A) analog coding


compression to the number of bits (B) arithmetic coding
required to represent the data after (C) huffman coding
compression is known as: (D) waveform coding

[190543] [RCS087]
[Page-3]
10. no codeword is a
The difference between the data and 15. A code in which
codeword is called
the model is called as prefix to another

(A) residual a
uniquely decodable code
(B) waveform (A)
huffman code
(C) distortion (B)
arithmetic code
(D) none of the above (C)
11. The first order prefix code
entropy of the (D
difference between
the entropy
following sequenceis 16. The
1 23234 54567898910 and the average length is called
(A) 2.25 bits/symbol (A)
self information

(B) 3.25 bits/symbol (B) redundancy


(C) 1.25 bits/symbol (C) probability
(D) 5.25 bits/symbol (D)
none of the above

NYT stands
12 The probability models also known 17. In Huffman tree

with another name for


(A) knowledge Model No Yet Transmitted
(A)
(B) reconstructed Model Not Yet Transmission
(B)
Not Yet Transmitted
( ignorance Model (C)
(D) none of the above (D) None of the above
13. ASCII code uses the same number of 18. In Golomb code the unary code for
bits to represent each symbol such a positive integer n is simply

code is called (A) n 0s followed by 1


variable length code ls followed by 0
(A) (B)n
(B) fixedxed code (C) n lsfollowed by 00

(C fixed length code (D) n Os followed by 11

(D) none of the above 19. In Golomb code the unary code of 5

14. Test the following codes are IS

uniquely decodable or not: (A) 111110


{0,01,11} (B) 000001
(A) uniquely decodable codes C) 1111100
(B) not uniquely decodable (D) 0000011
(C) neither uniquely decodable nor 20. In Rice codes the CCSDS stands for
not uniquely decodable (A) Consultative Committee on
(D) none of the following Standard IData Science

(190543 [RCS087] Page-4)


rows
of
(B) Consultative Commission on 25, The image consists of 256
uncompressed
Standard Data Science so the
256 pixcls ,

bytes.
(C) Consultative Committee on representation uses_
Space Data Standards
(A 65536
(D) Consultative Committee on 65426
(B)
Specialized Data Science
C) 66536
21. In CCSDS algorithm the (D) 66325
for
removes correlation from the
input audio signal
Z6. At what rate the
and generates a sampled?
sequence of each stereo channel i
nonnegative integers. 42.2 kHz
(A)
(A) a binary coder (B) 43.1 kHz
(B) a binary generator
(C) 44.1 kHz
(C) a parity checker 44.2 kHz
(D)
(D). a pre-processor when
27. In Rice codes the ROS is used
22. The binary coder generates a blocks in a
the last five or more

to represent the segment are all


integer sequence. (A) one

(A) byte stream


(B), zero

(B) bit Stream


(C) same

C) character stream
D) none of the above
(D) none of the above 28. What is the value of redundancy in
23. In which codes all codewords have
Huffman codes when the
equal length? probabilities are negative powers
(A) Huffiman codes of two?
(B) Golomb codes (A) 1

C) Prefix Codes (B) 2


(D) Tunstall codes (C) 3
24. In n-bit Tunstall code for a source
D)0
the number of codewords is: 29. Who was the developer of Rice
(A)logn codes?
(B) 2n (A) Robert F. Rice
(C) 2n-1 (B) Robin S. Rice
(D) 2n+1 (C)Robert S. Rice
(D) None of the above

[190543] RCSO87
Page-5]
30, Which is the most popular method 35. Which one of the following is an
for generating variable length codes? application of LZW compression
(A) Huffman Coding algorithm?
(B) arithmetie Coding (A) Audio compression
run Length coding Video compression
(B)
(D) none of the above
C) Text compression
31 Which one of the following is not (D) Unix compress
application of arithmetic coding? 36. GIF stands for
(A) bi-level image compression
(A) Graphics Interchange Format
(B) JBIG
(B) Graphical Image Format
(C) JBIG-2
C) Graphics Image Format
(D) text compression
(D) None of the above
32 A static dictionary technique that is
37. Which coding scheme use the
less specific to a single application history of the sequence to determine
its encoding?
(A)y digram coding (A) Huffman coding
(B) arithmetic coding
(B) Arithmetic coding
(C) huffman coding
C) Prefix coding
D) all of the above
(D) None of the above
33. In LZ77 a buffer that contains 38. Which is the best known context
one
portion of the recently encoded
based algorithm?
sequence is called
(A) Prediction with Partial Match
(A) look ahead buffer
(B)LZ7
(B) search buffer
(C) LZ78
(C) stackk buffer
(D) LZW
(D) none of the above 39. CALIC stands for
34. In LZ77 a buffer that contains next (A) Context Adaptive Lossless
portion of the sequence to be
Image Compression
encoded is called a
(B) Content Adaptive Lossy Image
(A) stack buffer
compression
(B) look ahead buffer (C) Context Adaptive Lossy Image
(C) search buffer
compression
(D all of the above (D) none of the above

[190543] [RCS087] [Page-6]


40 Which coding scheme that takes
advantages of long runs of identical
C multiresolution model

(D) capon model

symbols? takes the


45. Which compression
(A) arithmetic coding and
advantage of relationships
(B, move to front coding correlations that extend beyond a

(C) move to back


coding single symbol?
(D) predictive coding (A) Image compression
41. In which transmission a page is Markov
(B) Dynamic
Scanned and converted into a Compression
sequence of black or white pixels? C) Audio compression

(A) audio transmission D) All of the above


measures
video transmission 46. Name of the two popular
(B)
facsimile transmission of distortion are
(C).
and absolute
error
D) all ofthe above (A) squared
Which algorithm requires that the difference
42.
entire sequence to be coded be (B) fidelity and quality
and vector
available to the encoder before the (C)scalar quantization

coding takes place? quantization


squared error and signal
(A LZW (D) mean

to noise ratio
(B) LZ8
47. Rate is defined as the
(C) The Burrows-Wheeler

number of bits used to represent each


Transform
sample value.
(D) None of the above

rise the run (A) Maximum


43. Which model gives
(B) Minimum
length coding?
(A) Physical model (C) Average
B) Probability model (D) all of the above

Composite source model 48. The rate distortion function R(D)


(C)
Capon model specifiesthe rate at which
(D)
44. Which type of model generate the output of a source can be

representations of an image with encoded while keeping the distortion


varying spatial resolution? less than or equal to D.

(A) physical model (A) highest


B) probability model (B) lowest

[190543] [RCS087]
[Page-7]
(C) smallest (B). encoder mapping

(D) average (C) both (A) and (B)


49. The models play an important role in (D) none of the above

the of lossy compression 54. A mapping that generates a

algorithms. reconstruction value is

(A) testing called


(B) maintenance
(A)encodermapping
(C2 design (B) decoder mapping
(D) coding
(C) multiplexer mapping
50. Which model is used to for the
D) demultiplexer mapping
design and analysis of lossy 55. A quantizer that does not have zero
compression schemes? as one of its representation levels is
(A), probability model
called
(B) linear System Model
(A) midrise qunatizer
(C) physical model (B) midtread quantizer
(D) none of the above
scalar quantizer
(C)
51. The models are based the
on
physics D) vector quantizer
of the source
output production are 56 Control Systems and audio coding
called schemes
of
are the examples
(A) probability models quantizer.
(B) physical models (A) midrise quantizer
C) composite source models (B) midtread quantizer
(D) none of the above
C) nonuniform quantizer
52. The process of representing large
a
(D) none of the above
possibly infinite set of values with a 57. In which quntization approach there
much smaller set is called is a need for side information?
(A) uniform quantization
(B) (A)forwardadaptive
scalar quantization (B) backward adaptive
(C non uniform quantization
(C) uniform
D) quantization (D) scalar
53. A mapping which divides the range 58. In quantization the
of values that the source generates quantizer output is available to both
into number of intervals is called transmitter and receiver, there is no
(A) decoder mapping need for side information.

[190543] [RCS087] [Page-8]


(A) scalar 03, In vector quantization, a set ot

vector called
(B) dimensional vectors

(C) forward adaptive of the vector quantizer.

(D)backwardadaptive (A) code vectors

59. A quantizer also known with another (B) directory


name " quantization with one word
(C) vectors

(D codebook
memory" is called
a
code vector is assigned
midrise quantizer 64. Each
(A)
index.
(B) midtread quantizer
binary
(C) Jayant quantizer (A)
decimal
(D) vector quantizer (B)
octal
60. A nonuniform quantizer provides (C)
none of the above
average distortion. D) codebook
is
vectors in the
lowest 65 The
(A) known as

(B) lower (A) b l o c k vectors

(C) average (B)quantizedvectors


D) higher sampled vectors

-Max quantizer,
C)
61. For a given Lloyd (D) code-vectors

output and ne
the quantizer at the
In vector quantization
66.
to each
quantization noise are vector is compared
the input
find code
diagonal in order to
(A) code vector

B) square to the input vector.


vector closest

orthogonal Decoder
C) (A)
parallel Encoder
(D) he (B)
companded quantization,

62. n (C)Multiplexer
None of the above
compressor

he (D)
of the code-vector are
function The elements
close to the
67.
regions values of the source
high-probability
the
origin.
Output.
compresses
(A) sampled
(A)
(B)
expands
B) binary
streches (C) quantized
(C)
above
all of the none of the above
(D) (D)

[RCSO87 [Page-9
[190543
arrangements
of output
70 The regular
68. The LBZ stands for
are called
points in space
Linde-Buze-Gray
(A) (A) cell
(B) Linde-Buzo-Girey
lattices
(C) Linde-Buzo-Green
(B)
None of the above (C) Pyramid

(D) the above


codebook the (D) none of
69. In tree structured

removal o subgroups is called

(A) pruning
(B) parsing

(C) updation
none of the above
(D)

******

You might also like