Hadoop Questions and Answers Part-15

1. Avro schemas are defined with _____
a) JSON
b) XML
c) JAVA
d) All of the mentioned

Answer: a
Explanation: JSON implementation facilitates implementation in languages that already have JSON libraries.

2. Point out the correct statement.
a) Avro provides functionality similar to systems such as Thrift
b) When Avro is used in RPC, the client and server exchange data in the connection handshake
c) Apache Avro, Avro, Apache, and the Avro and Apache logos are trademarks of The Java Foundation
d) None of the mentioned

Answer: a
Explanation: Avro differs from these systems in the fundamental aspects like untagged data.

3. __________ facilitates construction of generic data-processing systems and languages.
a) Untagged data
b) Dynamic typing
c) No manually-assigned field IDs
d) All of the mentioned

Answer: b
Explanation: Avro does not require that code be generated

4. With ______ we can store data and read it easily with various programming languages.
a) Thrift
b) Protocol Buffers
c) Avro
d) None of the mentioned

Answer: c
Explanation: Avro is optimized to minimize the disk space needed by our data and it is flexible.

5. ________ are a way of encoding structured data in an efficient yet extensible format.
a) Thrift
b) Protocol Buffers
c) Avro
d) None of the mentioned

Answer: b
Explanation: Google uses Protocol Buffers for almost all of its internal RPC protocols and file formats.

6. Thrift resolves possible conflicts through _________ of the field.
a) Name
b) Static number
c) UID
d) None of the mentioned

Answer: b
Explanation: Avro resolves possible conflicts through the name of the field.

7. Avro is said to be the future _______ layer of Hadoop.
a) RMC
b) RPC
c) RDC
d) All of the mentioned

Answer: b
Explanation: When Avro is used in RPC, the client and server exchange schemas in the connection handshake.

8. When using reflection to automatically build our schemas without code generation, we need to configure Avro using?
a) AvroJob.Reflect(jConf);
b) AvroJob.setReflect(jConf);
c) Job.setReflect(jConf);
d) None of the mentioned

Answer: c
Explanation: For strongly typed languages like Java, it also provides a generation code layer, including RPC services code generation.

9. We can declare the schema of our data either in a ______ file.
a) JSON
b) XML
c) SQL
d) R

Answer: c
Explanation: Schema can be declared using an IDL or simply through Java beans by using reflection-based schema building.

10. ________ job usually splits the input data-set into independent chunks which are processed by the map tasks in a completely parallel manner.
a) Tasker
b) MapReduce
c) Tasktrack
d) None of the mentioned

Answer: b
Explanation: Hadoop MapReduce is a software framework for easily writing applications that process vast amounts of data.