MCQOPTIONS
Saved Bookmarks
This section includes 230 Mcqs, each offering curated multiple-choice questions to sharpen your Technical Programming knowledge and support exam preparation. Choose a topic below to get started.
| 201. |
HDFS and NoSQL file systems focus almost exclusively on adding nodes to : |
| A. | Scale out |
| B. | Scale up |
| C. | Both Scale out and up |
| D. | None of the mentioned |
| Answer» B. Scale up | |
| 202. |
Which of the following phases occur simultaneously ? |
| A. | Shuffle and Sort |
| B. | Reduce and Sort |
| C. | Shuffle and Map |
| D. | All of the mentioned |
| Answer» B. Reduce and Sort | |
| 203. |
The output of the _______ is not sorted in the Mapreduce framework for Hadoop. |
| A. | Mapper |
| B. | Cascader |
| C. | Scalding |
| D. | None of the mentioned |
| Answer» E. | |
| 204. |
The ________ class provides the getValue() method to read the values from its instance. |
| A. | Get |
| B. | Result |
| C. | Put |
| D. | Value |
| Answer» C. Put | |
| 205. |
__________ class adds HBase configuration files to its object. |
| A. | Configuration |
| B. | Collector |
| C. | Component |
| D. | None of the mentioned |
| Answer» B. Collector | |
| 206. |
The standard output (stdout) and error (stderr) streams of the task are read by the TaskTracker and logged to : |
| A. | ${HADOOP_LOG_DIR}/user |
| B. | ${HADOOP_LOG_DIR}/userlogs |
| C. | ${HADOOP_LOG_DIR}/logs |
| D. | None of the mentioned |
| Answer» C. ${HADOOP_LOG_DIR}/logs | |
| 207. |
During the execution of a streaming job, the names of the _______ parameters are transformed. |
| A. | vmap |
| B. | mapvim |
| C. | mapreduce |
| D. | mapred |
| Answer» E. | |
| 208. |
The ___________ executes the Mapper/ Reducer task as a child process in a separate jvm. |
| A. | JobTracker |
| B. | TaskTracker |
| C. | TaskScheduler |
| D. | None of the mentioned |
| Answer» B. TaskTracker | |
| 209. |
__________ is the primary interface for a user to describe a MapReduce job to the Hadoop framework for execution. |
| A. | JobConfig |
| B. | JobConf |
| C. | JobConfiguration |
| D. | All of the mentioned |
| Answer» C. JobConfiguration | |
| 210. |
The right level of parallelism for maps seems to be around _________ maps per-node |
| A. | 1-10 |
| B. | 10-100 |
| C. | 100-150 |
| D. | 150-200 |
| Answer» C. 100-150 | |
| 211. |
Applications can use the ____________ to report progress and set application-level status messages |
| A. | Partitioner |
| B. | OutputSplit |
| C. | Reporter |
| D. | All of the mentioned |
| Answer» D. All of the mentioned | |
| 212. |
Reducer is input the grouped output of a : |
| A. | Mapper |
| B. | Reducer |
| C. | Writable |
| D. | Readable |
| Answer» B. Reducer | |
| 213. |
Interface ____________ reduces a set of intermediate values which share a key to a smaller set of values. |
| A. | Mapper |
| B. | Reducer |
| C. | Writable |
| D. | Readable |
| Answer» C. Writable | |
| 214. |
________ is the slave/worker node and holds the user data in the form of Data Blocks. |
| A. | DataNode |
| B. | NameNode |
| C. | Data block |
| D. | Replication |
| Answer» B. NameNode | |
| 215. |
The need for data replication can arise in various scenarios like : |
| A. | Replication Factor is changed |
| B. | DataNode goes down |
| C. | Data Blocks get corrupted |
| D. | All of the mentioned |
| Answer» E. | |
| 216. |
Which of the following scenario may not be a good fit for HDFS ? |
| A. | HDFS is not suitable for scenarios requiring multiple/simultaneous writes to the same file |
| B. | HDFS is suitable for storing data related to applications requiring low latency data access |
| C. | HDFS is suitable for storing data related to applications requiring low latency data access |
| D. | None of the mentioned |
| Answer» B. HDFS is suitable for storing data related to applications requiring low latency data access | |
| 217. |
_________ identifies filesystem pathnames which work as usual with regular expressions. |
| A. | archiveName |
| B. | source |
| C. | destination |
| D. | None of the mentioned |
| Answer» E. | |
| 218. |
Which of the following parameter describes destination directory which would contain the archive ? |
| A. | archiveName |
| B. | source |
| C. | destination |
| D. | None of the mentioned |
| Answer» D. None of the mentioned | |
| 219. |
_________ is a pluggable Map/Reduce scheduler for Hadoop which provides a way to share large clusters. |
| A. | Flow Scheduler |
| B. | Data Scheduler |
| C. | Capacity Scheduler |
| D. | None of the mentioned |
| Answer» D. None of the mentioned | |
| 220. |
On a tasktracker, the map task passes the split to the createRecordReader() method on InputFormat to obtain a _________ for that split. |
| A. | InputReader |
| B. | RecordReader |
| C. | OutputReader |
| D. | None of the mentioned |
| Answer» C. OutputReader | |
| 221. |
InputFormat class calls the ________ function and computes splits for each file and then sends them to the jobtracker. |
| A. | puts |
| B. | gets |
| C. | getSplits |
| D. | All of the mentioned |
| Answer» D. All of the mentioned | |
| 222. |
Hadoop achieves reliability by replicating the data across multiple hosts, and hence does not require ________ storage on hosts. |
| A. | RAID |
| B. | Standard RAID levels |
| C. | ZFS |
| D. | Operating system |
| Answer» B. Standard RAID levels | |
| 223. |
Which of the following platforms does Hadoop run on ? |
| A. | Bare metal |
| B. | Debian |
| C. | Cross-platform |
| D. | Unix-like |
| Answer» D. Unix-like | |
| 224. |
What was Hadoop written in ? |
| A. | Java (software platform) |
| B. | Perl |
| C. | Java (programming language) |
| D. | Lua (programming language) |
| Answer» D. Lua (programming language) | |
| 225. |
Which of the following genres does Hadoop produce ? |
| A. | Distributed file system |
| B. | JAX-RS |
| C. | Java Message Service |
| D. | Relational Database Management System |
| Answer» B. JAX-RS | |
| 226. |
___________ is general-purpose computing model and runtime system for distributed data analytics. |
| A. | Mapreduce |
| B. | Drill |
| C. | Oozie |
| D. | None of the mentioned |
| Answer» B. Drill | |
| 227. |
________ is the most popular high-level Java API in Hadoop Ecosystem |
| A. | Scalding |
| B. | HCatalog |
| C. | Cascalog |
| D. | Cascading |
| Answer» E. | |
| 228. |
All of the following accurately describe Hadoop, EXCEPT: |
| A. | Open source |
| B. | Real-time |
| C. | Java-based |
| D. | Distributed computing approach |
| Answer» C. Java-based | |
| 229. |
What was Hadoop named after? |
| A. | Creator Doug Cutting’s favorite circus act |
| B. | Cutting’s high school rock band |
| C. | The toy elephant of Cutting’s son |
| D. | A sound Cutting’s laptop made during Hadoop’s development |
| Answer» D. A sound Cutting’s laptop made during Hadoop’s development | |
| 230. |
Point out the wrong statement : |
| A. | Hardtop’s processing capabilities are huge and its real advantage lies in the ability to process terabytes & petabytes of data |
| B. | Hadoop uses a programming model called “MapReduce”, all the programs should confirms to this model in order to work on Hadoop platform |
| C. | The programming model, MapReduce, used by Hadoop is difficult to write and test |
| D. | All of the mentioned |
| Answer» D. All of the mentioned | |