Description : In a Hadoop cluster, what is true for a HDFS block that is no longer available due to disk corruption or machine failure? a)It is lost for ever b)It can be replicated form its alternative locations ... to read it. d)The Mapreduce job process runs ignoring the block and the data stored in it.
Last Answer : It can be replicated form its alternative locations to other live machines.
Description : Point out the correct statement. a) Hadoop is an ideal environment for extracting and transforming small volumes of data b) Hadoop stores data in HDFS and supports data compression/decompression c) The ... less useful than a MapReduce job to solve graph and machine learning d) None of the mentioned
Last Answer : Hadoop stores data in HDFS and supports data compression/decompression
Description : Point out the correct statement. a) Hadoop is an ideal environment for extracting and transforming small volumes of data b) Hadoop stores data in HDFS and supports data compression/decompression c) The ... useful than a MapReduce job to solve graph and machine learning d) None of the mentioned
Description : Above the file systems comes the ________ engine, which consists of one Job Tracker, to which client applications submit MapReduce jobs. a) MapReduce b) Google c) Functional programming d) Facebook
Last Answer : MapReduce
Description : _________ is a web programming model for scalable data processing on large clusters over large data sets a) MapReduce b) HDFS c) GFS d) DFS
Description : The Hadoop MapReduce framework spawns one map task for each __________ generated by the InputFormat for the job. a) OutputSplit b) InputSplit c) InputSplitStream d) All of the mentioned
Last Answer : ) InputSplit
Description : Which of the following is built on top of a Hadoop framework using the Elastic Compute Cloud? a) Amazon Elastic MapReduce b) Amazon Mechanical Turkc) Amazon DevPay d) Multi-Factor Authentication
Last Answer : Amazon Elastic MapReduce
Description : What are the advantages of HDFS federation in Hadoop? a) Isolation b) Namespace scalability c) Improves throughput d) All of the above
Last Answer : All of the above
Description : The client reading the data from HDFS filesystem in Hadoop a) gets the data from the namenode b) gets the block location from the datanode c) gets only the block locations form the namenode d) gets both the data and block location from the namenode
Last Answer : gets only the block locations form the namenode
Description : In Hadoop, the files are stored in a) Directory b) DFS c) GFS d) HDFS
Last Answer : HDFS
Description : Very large sustainable reading and writing bandwidth, mostly continuous accessing instead of random accessing. The programming interface is similar to that of the POSIX file system accessing interface. This is ... : Google File System c) HDFS: Hadoop Distributes File System d) None of the above
Last Answer : GFS: Google File System
Description : Sun also has the Hadoop Live CD ________ project, which allows running a fully functional Hadoop cluster using a live CD. a) OpenOffice.org b) OpenSolaris c) GNU d) Linux
Last Answer : OpenSolaris
Description : During Safemode Hadoop cluster is in a) Read-only b) Write-only c) Read-Write d) None of the above
Last Answer : Read-only
Description : The gateway administrator provides the _________ template information when the infrastructure is set up. a) Cloud b) HDFS c) better d) VM
Last Answer : VM
Description : The building blocks of Google’s App Engine architecture comprises of a) GFS, MapReduce, Chubby, BigTable b) GFS, Scheduler, Chubby, BigTable c) GFS, MapReduce, Chunk Server, BigTable d) Scheduler, MapReduce, Chubby, Chunk Server
Last Answer : GFS, MapReduce, Chubby, BigTable
Description : MapReduce engine has a master/slave architecture consisting of a single ______ as the master and a number of _______ as the slaves a) Name Node, Data Node b) JobTracker, TaskTrackers c) TaskTrackers, JobTracker d) Data Node, Name Node
Last Answer : JobTracker, TaskTrackers
Description : . ___________ was built primarily as the fundamental storage service for Google’s search engine. a) BFS b) HDFS c) GFS d) GAE
Last Answer : 200 TB
Description : ___________ was built primarily as the fundamental storage service for Google’s search engine. a) BFS b) HDFS c) GFS d) GAE
Last Answer : GFS
Description : _________ software library to write and run large user applications on vast data sets in business applications a) Apache Tomcat b) Hadoop c) Open Stack d) Open Nebula
Last Answer : Hadoop
Description : Point out the wrong statement. a) A MapReduce job usually splits the input data-set into independent chunks which are processed by the map tasks in a completely parallel manner b) The ... implement the Mapper and Reducer interfaces toprovide the map and reduce methods d) None of the mentioned
Last Answer : None of the mentioned
Description : Which one of the following statements is true regarding pairs of a MapReduce job? a)A key class must implement Writable. b)A key class must implement WritableComparable. c)A value class must implement WritableComparable. d)A value class must extend WritableComparable.
Last Answer : A key class must implement WritableComparable.
Description : Which one of the following statements is false regarding the Distributed Cache? a) The Hadoop framework will ensure that any files in the Distributed Cache are distributed to all map and reduce tasks. b) ... Cache on to the slave node before any tasks for the job are executed on that node.
Last Answer : MapReduce tries to place the data and the compute as close as possible
Description : Job tracker runs on a) Namenode b) Datanode c) Secondary namenode d) Secondary datanode
Last Answer : )Sqoop
Description : Which of the following is a means for accessing human researchers or consultants to help solve problems on a contractual or temporary basis? a) Amazon Elastic MapReduce b) Amazon Mechanical Turk c) Amazon DevPay d) Multi-Factor Authentication
Last Answer : Amazon Mechanical Turk
Description : Which of the following is a billing and account management service? a) Amazon Elastic MapReduce b) Amazon Mechanical Turk c) Amazon DevPay d) Multi-Factor Authentication
Last Answer : Amazon DevPay
Description : .A ________ node acts as the Slave and is responsible for executing a Task assigned to it by the JobTracker. a) MapReduce b) Mapper c) TaskTracker d) JobTracker
Last Answer : TaskTracker
Description : _______ in GAE provides coarse-grained locking service a) Bigtable b) Chubby c) MapReduce d) GFS
Last Answer : Chubby
Description : The comman used to copy a directory form one node to another in HDFS is a) rcp b) dcp c) drcp d) distcp
Last Answer : distcp
Description : The hdfs command to create the copy of a file from a local system is a) CopyFromLocal b) copyfromlocal c) CopyLocal d) copyFromLocal
Last Answer : copyFromLocal
Description : The architecture for _______ is built on the concepts of shared-nothing and messaging-based information exchange. a) HDFS b) Chabby c) Nova d) BigTable
Last Answer : Nova
Description : To read a file in HDFS, a user sends an _______ request to the NameNode toget the location of file blocks. a) access b) open c) authentication d) create
Last Answer : open
Description : The main goal of HDFS High availability is a) Faster creation of the replicas of primary namenode. b) To reduce the cycle time required to bring back a new primary namenode after existing ... data loss due to failure of primary namenode. d) Prevent the primary namenode form becoming single point
Last Answer : To reduce the cycle time required to bring back a new primary namenode after existing primary fails.
Description : In HDFS, reliability of storing data was maintained by a) Block Operations b) Block Replication c) Block Storage d) All the above
Last Answer : Block Replication
Description : The Master of the HDFS Architecture is a) Node Manager b) Data Node c) Name Node d) None of the above
Last Answer : Name Node
Description : . _____________ The open source clone of GFS. Written in Java. The programming interfaces are similar to POSIX but not identical. a) HDFS b) SAMIL c) SCIM d) DVE
Description : Which of the following can be identified as cloud? a) Web Applicationsb) Intranet c) Hadoop d) All of the mentioned
Description : The Hadoop list includes the HBase database, the Apache Mahout ________ system, and matrix operations. a) Machine learning b) Pattern recognition c) Statistical classification d) Artificial intelligence
Last Answer : Machine learning
Description : Hadoop achieves reliability by replicating the data across multiple hosts and hence does not require ________ storage on hosts. a) RAID b) Standard RAID levels c) ZFS d) Operating system
Last Answer : RAID
Description : Which of the following platforms does Hadoop run on? a) Bare metal b) Debian c) Cross-platform d) Unix-like
Last Answer : Cross-platform
Description : What was Hadoop written in? a) Java (software platform) b) Perl c) Java (programming language) d) Lua (programming language)
Last Answer : Java (programming language)
Description : Which of the following genres does Hadoop produce? a) Distributed file system b) JAX-RS c) Java Message Service d) Relational Database Management System
Last Answer : Distributed file system
Description : What license is Hadoop distributed under? a) Apache License 2.0 b) Mozilla Public License c) Shareware d) Commercia
Last Answer : Apache License 2.0
Description : IBM and ________ have announced a major initiative to use Hadoop to support university courses in distributed computer programming. a) Google Latitude b) Android (operating system) c) Google Variations d) Google
Last Answer : Google
Description : Which of the following can be identified as cloud? a) Web Applications b) Intranet c) Hadoop d) All of the mentioned
Description : Which of the following command is used to enter Safemode a) hadoop dfsadmin –safemode get b) bin dfsadmin –safemode get c) hadoop dfsadmin –safemode enter d) None of the above
Last Answer : hadoop dfsadmin –safemode enter
Description : Under Hadoop High Availability, Fencing means a)Preventing a previously active namenode from start running again. b)Preventing the start of a failover in the event of network failure with the ... previously active namenode. d)Preventing a previously active namenode from writing to the edit log.
Last Answer : Preventing a previously active namenode from writing to the edit log.
Description : ________ is a utility which allows users to create and run jobs with any executables as the mapper and/or the reducer. a) Hadoop Strdata b) Hadoop Streaming c) Hadoop Stream d) None of the mentioned
Last Answer : Hadoop Streaming