Generic selectors
Exact matches only
Search in title
Search in content
Search in posts
Search in pages
Filter by Categories
nmims post
Objective Type Set
Online MCQ Assignment
Question Solution
Solved Question
Uncategorized

Hadoop MCQ Set 1

1. Which hdfs command is used to check for various inconsistencies ?
a) fsk
b) fsck
c) fetchdt
d) none of the mentioned

View Answer

Answer: b [Reason:] fsck is designed for reporting problems with various files, for example, missing blocks for a file or under-replicated blocks.

2. Point out the correct statement :
a) All hadoop commands are invoked by the bin/hadoop script
b) Hadoop has an option parsing framework that employs only parsing generic options
c) Archive command creates a hadoop archive
d) All of the mentioned

View Answer

Answer: a [Reason:] Running the hadoop script without any arguments prints the description for all commands.

3. HDFS supports the ____________ command to fetch Delegation Token and store it in a file on the local system.
a) fetdt
b) fetchdt
c) fsk
d) rec

View Answer

Answer: b [Reason:] Delegation token can be later used to access secure server from a non secure client.

4. In ___________ mode, the NameNode will interactively prompt you at the command line about possible courses of action you can take to recover your data.
a) full
b) partial
c) recovery
d) commit

View Answer

Answer: c [Reason:] Recovery mode can cause you to lose data, you should always back up your edit log and fsimage before using it.

5. Point out the wrong statement :
a) classNAME displays the class name needed to get the Hadoop jar
b) Balancer Runs a cluster balancing utility
c) An administrator can simply press Ctrl-C to stop the rebalancing process
d) None of the mentioned

View Answer

Answer: a [Reason:] classpath prints the class path needed to get the Hadoop jar and the required libraries.

6. _________ command is used to copy file or directories recursively.
a) dtcp
b) distcp
c) dcp
d) distc

View Answer

Answer: b [Reason:] Usage of the distcp command: hadoop distcp .

7. __________ mode is a Namenode state in which it does not accept changes to the name space.
a) Recover
b) Safe
c) Rollback
d) None of the mentioned

View Answer

Answer: c [Reason:] dfsadmin runs a HDFS dfsadmin client.

8. __________ command is used to interact and view Job Queue information in HDFS.
a) queue
b) priority
c) dist
d) all of the mentioned

View Answer

Answer: a [Reason:] Hadoop script can be used to invoke any class.

9. Which of the following command runs the HDFS secondary namenode ?
a) secondary namenode
b) secondarynamenode
c) secondary_namenode
d) none of the mentioned

View Answer

Answer: b [Reason:] The secondary NameNode merges the fsimage and the edits log files periodically and keeps edits log size within a limit.

10. Which of the following is used for the MapReduce job Tracker node ?
a) mradmin
b) tasktracker
c) jobtracker
d) none of the mentioned

View Answer

Answer: c [Reason:] tasktracker runs a MapReduce task Tracker node.

Hadoop MCQ Set 2

1. Which of the following is a common hadoop maintenance issue ?
a) Lack of tools
b) Lack of configuration management
c) Lack of web interface
d) None of the mentioned

View Answer

Answer: b [Reason:] Without a centralized configuration management framework, you end up with a number of issues that can cascade just as your usage picks up.

2. Point out the correct statement :
a) RAID is turned off by default
b) Hadoop is designed to be a highly redundant distributed system
c) Hadoop has a networked configuration system
d) None of the mentioned

View Answer

Answer: b [Reason:] Hadoop deployment is sometimes difficult to implement.

3. ___________ mode allows you to suppress alerts for a host, service, role, or even the entire cluster.
a) Safe
b) Maintenance
c) Secure
d) All of the mentioned

View Answer

Answer: b [Reason:] Maintenance mode can be useful when you need to take actions in your cluster and do not want to see the alerts that will be generated due to those actions.

4. Which of the following is a configuration management system ?
a) Alex
b) Puppet
c) Acem
d) None of the mentioned

View Answer

Answer: b [Reason:] Administrators may use configuration management systems such as Puppet and Chef to manage processes.

5. Point out the wrong statement :
a) If you set the HBase service into maintenance mode, then its roles (HBase Master and all Region Servers) are put into effective maintenance mode
b) If you set a host into maintenance mode, then any roles running on that host are put into effective maintenance mode
c) Putting a component into maintenance mode prevent events from being logged
d) None of the mentioned

View Answer

Answer: c [Reason:] Maintenance mode only suppresses the alerts that those events would otherwise generate.

6. Which of the following is a common reason to restart hadoop process ?
a) Upgrade Hadoop
b) React to incidents
c) Remove worker nodes
d) All of the mentioned

View Answer

Answer: d [Reason:] The most common reason administrators restart Hadoop processes is to enact configuration changes.

7. __________ Manager’s Service feature monitors dozens of service health and performance metrics about the services and role instances running on your cluster.
a) Microsoft
b) Cloudera
c) Amazon
d) None of the mentioned

View Answer

Answer: b [Reason:] Manager’s Service feature presents health and performance data in a variety of formats.

8. Which of the tab shows all the role instances that have been instantiated for this service ?
a) Service
b) Status
c) Instance
d) All of the mentioned

View Answer

Answer: c [Reason:] The Instances page displays the results of the configuration validation checks it performs for all the role instances for this service.

9. __________ is a standard Java API for monitoring and managing applications.
a) JVX
b) JVM
c) JMX
d) None of the mentioned

View Answer

Answer: c [Reason:] Hadoop includes several managed beans (MBeans), which expose Hadoop metrics to JMX-aware applications.

10. NameNode is monitored and upgraded in a __________ transition.
a) safemode
b) securemode
c) servicemode
d) none of the mentioned

View Answer

Answer: b [Reason:] The HDFS service has some unique functions that may result in additional information on its Status and Instances pages.

Hadoop MCQ Set 3

1. Apache Hadoop Development Tools is an effort undergoing incubation at :
a) ADF
b) ASF
c) HCC
d) AFS

View Answer

Answer: b [Reason:] The Apache Software Foundation(ASF) is sponsored by the Apache Incubator PMC.

2. Point out the correct statement :
a) HDT tool allows you to allow working with only 1.1 version of Hadoop
b) HDT tool allows you to allow working with multiple versions of Hadoop
c) HDT tool allows you to allow working with multiple versions of Hadoop from multiple IDE
d) All of the mentioned

View Answer

Answer: b [Reason:] HDT project is currently a member of the Apache Incubator.

3. HDT project works with eclipse version ________ and above.
a) 3.4
b) 3.5
c) 3.6
d) 3.7

View Answer

Answer: c [Reason:] The user should be able to install using a single update site for all Hadoop-related Eclipse tools.

4. HDT has been tested on __________ and Juno, and can work on Kepler as well.
a) Rainbow
b) Indigo
c) Indiavo
d) Hadovo

View Answer

Answer: b [Reason:] HDT aims at bringing plugins in eclipse to simplify development on Hadoop platform.

5. Point out the wrong statement :
a) There is support for creating Hadoop project in HDT
b) HDT aims at bringing plugins in eclipse to simplify development on Hadoop platform
c) HDT is based on eclipse plugin architecture and can possibly support other versions like 0.23, CDH4 etc in next releases
d) None of the mentioned

View Answer

Answer: d [Reason:] HDT aims to simplify the Hadoop platform for developers.

6. Which of the following tool is intended to be more compatible with HDT ?
a) Git
b) Juno
c) Indigo
d) None of the mentioned

View Answer

Answer: c [Reason:] The HDT uses a git repository, which anyone is free to checkout.

7. Which of the following has the core Eclipse PDE tools for HDT development ?
a) RVP
b) RAP
c) RBP
d) RVP

View Answer

Answer: b [Reason:] RCP/RAP developers package has the core Eclipse PDE tools.

8. HDT provides plugin for inspecting ________ nodes.
a) LocalWriter
b) HICC
c) HDFS
d) All of the mentioned

View Answer

Answer: c [Reason:] The Hadoop Development Tools (HDT) is a set of plugins for the Eclipse IDE for developing against the Hadoop platform.

9. HDT is used for listing running Jobs on __________ Cluster
a) MR
b) Hive
c) Pig
d) None of the mentioned

View Answer

Answer: a [Reason:] HDT can be used for launching Map-Reduce programs on a Hadoop cluster.

10. HDT provides wizards for creating Java Classes for :
a) Mapper
b) Reducer
c) Driver
d) All of the mentioned

View Answer

Answer: d [Reason:] HDT provides wizards for creation of Hadoop Based Projects.

Hadoop MCQ Set 4

1. IBM and ________ have announced a major initiative to use Hadoop to support university courses in distributed computer programming.
a) Google Latitude
b) Android (operating system)
c) Google Variations
d) Google

View Answer

Answer: d [Reason:] Google and IBM Announce University Initiative to Address Internet-Scale.

2. Point out the correct statement :
a) Hadoop is an ideal environment for extracting and transforming small volumes of data
b) Hadoop stores data in HDFS and supports data compression/decompression
c) The Giraph framework is less useful than a MapReduce job to solve graph and machine learning
d) None of the mentioned

View Answer

Answer: b [Reason:] Data compression can be achieved using compression algorithms like bzip2, gzip, LZO, etc. Different algorithms can be used in different scenarios based on their capabilities.

3. What license is Hadoop distributed under ?
a) Apache License 2.0
b) Mozilla Public License
c) Shareware
d) Commercial

View Answer

Answer: a [Reason:] Hadoop is Open Source, released under Apache 2 license.

4. Sun also has the Hadoop Live CD ________ project, which allows running a fully functional Hadoop cluster using a live CD.
a) OpenOffice.org
b) OpenSolaris
c) GNU
d) Linux

View Answer

Answer: b [Reason:] The OpenSolaris Hadoop LiveCD project built a bootable CD-ROM image.

5. Which of the following genres does Hadoop produce ?
a) Distributed file system
b) JAX-RS
c) Java Message Service
d) Relational Database Management System

View Answer

Answer: a [Reason:] The Hadoop Distributed File System (HDFS) is designed to store very large data sets reliably, and to stream those data sets at high bandwidth to user.

6. What was Hadoop written in ?
a) Java (software platform)
b) Perl
c) Java (programming language)
d) Lua (programming language)

View Answer

Answer: c [Reason:] The Hadoop framework itself is mostly written in the Java programming language, with some native code in C and command line utilities written as shell-scripts.

7. Which of the following platforms does Hadoop run on ?
a) Bare metal
b) Debian
c) Cross-platform
d) Unix-like

View Answer

Answer: c [Reason:] Hadoop has support for cross platform operating system.

8. Hadoop achieves reliability by replicating the data across multiple hosts, and hence does not require ________ storage on hosts.
a) RAID
b) Standard RAID levels
c) ZFS
d) Operating system

View Answer

Answer: a [Reason:] With the default replication value, 3, data is stored on three nodes: two on the same rack, and one on a different rack.

9. Above the file systems comes the ________ engine, which consists of one Job Tracker, to which client applications submit MapReduce jobs.
a) MapReduce
b) Google
c) Functional programming
d) Facebook

View Answer

Answer: a [Reason:] MapReduce engine uses to distribute work around a cluster.

10. The Hadoop list includes the HBase database, the Apache Mahout ________ system, and matrix operations.
a) Machine learning
b) Pattern recognition
c) Statistical classification
d) Artificial intelligence

View Answer

Answer: a [Reason:] The Apache Mahout project’s goal is to build a scalable machine learning tool.

Hadoop MCQ Set 5

1. __________ is a fully integrated, state-of-the-art analytic database architected specifically to leverage strengths of Hadoop.
a) Oozie
b) Impala
c) Lucene
d) BigTop

View Answer

Answer: b [Reason:] Impala provides scalability and flexibility to hadoop.

2. Point out the correct statement :
a) With Impala, more users, whether using SQL queries or BI applications, can interact with more data
b) Technical support for Impala is not available via a Cloudera Enterprise subscription
c) Impala is proprietary tool for Hadoop
d) None of the mentioned

View Answer

Answer: a [Reason:] It is possible through a single repository and metadata store from source through analysis.

3. Impala is an integrated part of a ____________ enterprise data hub.
a) MicroSoft
b) IBM
c) Cloudera
d) All of the mentioned

View Answer

Answer: c [Reason:] Impala is open source (Apache License), so you can self-support in perpetuity if you wish.

4. For Apache __________ users, Impala utilizes the same metadata.
a) cTakes
b) Hive
c) Pig
d) Oozie

View Answer

Answer: b [Reason:] You don’t have to worry about re-inventing the implementation wheel.

5. Point out the wrong statement :
a) For Apache Hive users, Impala utilizes the same metadata, ODBC driver, SQL syntax, and user interface as Hive
b) Impala provides high latency and low concurrency
c) Impala also scales linearly, even in multitenant environments
d) All of the mentioned

View Answer

Answer: b [Reason:] Impala provides low latency and high concurrency.

6. Impala is integrated with native Hadoop security and Kerberos for authentication via __________ module
a) Sentinue
b) Sentry
c) Sentinar
d) All of the mentioned

View Answer

Answer: b [Reason:] Via the Sentry module, you can ensure that the right users and applications are authorized for the right data.

7. Which of the following companies shipped Impala ?
a) Amazon
b) Oracle
c) MapR
d) All of the mentioned

View Answer

Answer: d [Reason:] Impala is shipped by Cloudera, MapR, Oracle, and Amazon.

8. ____________ analytics is a work in progress with Impala.
a) Reproductive
b) Exploratory
c) Predictive
d) All of the mentioned

View Answer

Answer: a [Reason:] Impala is the de facto standard for open source interactive business intelligence and data discovery.

9. Which of the following features is not provided by Impala ?
a) SQL functionality
b) ACID
c) Flexibility
d) None of the mentioned

View Answer

Answer: b [Reason:] Impala combines all of the benefits of other Hadoop frameworks, including flexibility, scalability, and cost-effectiveness, with the performance, usability, and SQL functionality necessary for an enterprise-grade analytic database.

10. Which of the following hadoop file formats is supported by Impala ?
a) SequenceFile
b) Avro
c) RCFile
d) All of the mentioned

View Answer

Answer: [Reason:] Impala reads Hadoop file formats, including text, LZO, SequenceFile, Avro, RCFile, and Parquet.