In a Hadoop cluster, what is true for a HDFS block that is no longer
available due to disk corruption or machine failure?
a)It is lost for ever
b)It can be replicated form its alternative locations to other live
machines.
c)The namenode allows new client request to keep trying to read it.
d)The Mapreduce job process runs ignoring the block and the data stored in
it.

1 Answer

Answer :

It can be replicated form its alternative locations to other live
machines.

Related questions

Description : The client reading the data from HDFS filesystem in Hadoop a) gets the data from the namenode b) gets the block location from the datanode c) gets only the block locations form the namenode d) gets both the data and block location from the namenode

Last Answer : gets only the block locations form the namenode

Description : The topmost layer of Hadoop is the _________ engine a) HDFS b) Cluster c) MapReduce d) Job Tracker

Last Answer : MapReduce

Description : Point out the correct statement. a) Hadoop is an ideal environment for extracting and transforming small volumes of data b) Hadoop stores data in HDFS and supports data compression/decompression c) The ... less useful than a MapReduce job to solve graph and machine learning d) None of the mentioned

Last Answer : Hadoop stores data in HDFS and supports data compression/decompression

Description : Point out the correct statement. a) Hadoop is an ideal environment for extracting and transforming small volumes of data b) Hadoop stores data in HDFS and supports data compression/decompression c) The ... useful than a MapReduce job to solve graph and machine learning d) None of the mentioned

Last Answer : Hadoop stores data in HDFS and supports data compression/decompression

Description : Point out the correct statement. a) Hadoop is an ideal environment for extracting and transforming small volumes of data b) Hadoop stores data in HDFS and supports data compression/decompression c) The ... useful than a MapReduce job to solve graph and machine learning d) None of the mentioned

Last Answer : Hadoop stores data in HDFS and supports data compression/decompression

Description : The main goal of HDFS High availability is a) Faster creation of the replicas of primary namenode. b) To reduce the cycle time required to bring back a new primary namenode after existing ... data loss due to failure of primary namenode. d) Prevent the primary namenode form becoming single point

Last Answer : To reduce the cycle time required to bring back a new primary namenode after existing primary fails.

Description : To read a file in HDFS, a user sends an _______ request to the NameNode toget the location of file blocks. a) access b) open c) authentication d) create

Last Answer : open

Description : Under Hadoop High Availability, Fencing means a)Preventing a previously active namenode from start running again. b)Preventing the start of a failover in the event of network failure with the ... previously active namenode. d)Preventing a previously active namenode from writing to the edit log.

Last Answer : Preventing a previously active namenode from writing to the edit log.

Description : Sun also has the Hadoop Live CD ________ project, which allows running a fully functional Hadoop cluster using a live CD. a) OpenOffice.org b) OpenSolaris c) GNU d) Linux

Last Answer : OpenSolaris

Description : In Hadoop, the files are stored in a) Directory b) DFS c) GFS d) HDFS

Last Answer : HDFS

Description : Which one of the following statements is false regarding the Distributed Cache? a) The Hadoop framework will ensure that any files in the Distributed Cache are distributed to all map and reduce tasks. b) ... Cache on to the slave node before any tasks for the job are executed on that node.

Last Answer : MapReduce tries to place the data and the compute as close as possible

Description : _________ is a web programming model for scalable data processing on large clusters over large data sets a) MapReduce b) HDFS c) GFS d) DFS

Last Answer : MapReduce

Description : The Hadoop MapReduce framework spawns one map task for each __________ generated by the InputFormat for the job. a) OutputSplit b) InputSplit c) InputSplitStream d) All of the mentioned

Last Answer : ) InputSplit

Description : During Safemode Hadoop cluster is in a) Read-only b) Write-only c) Read-Write d) None of the above

Last Answer : Read-only

Description : Which of the following is built on top of a Hadoop framework using the Elastic Compute Cloud? a) Amazon Elastic MapReduce b) Amazon Mechanical Turkc) Amazon DevPay d) Multi-Factor Authentication

Last Answer : Amazon Elastic MapReduce

Description : Job tracker runs on a) Namenode b) Datanode c) Secondary namenode d) Secondary datanode

Last Answer : )Sqoop

Description : What are the advantages of HDFS federation in Hadoop? a) Isolation b) Namespace scalability c) Improves throughput d) All of the above

Last Answer : All of the above

Description : Very large sustainable reading and writing bandwidth, mostly continuous accessing instead of random accessing. The programming interface is similar to that of the POSIX file system accessing interface. This is ... : Google File System c) HDFS: Hadoop Distributes File System d) None of the above

Last Answer : GFS: Google File System

Description : Point out the correct statement. a) A client can request access to a cloud service from any location b) A cloud has multiple application instances and directs requests to an instance based on ... a set of virtual machines with each machine being assigned a workload d) All of the mentioned

Last Answer : All of the mentioned

Description : Point out the correct statement. a) A client can request access to a cloud service from any location b) A cloud has multiple application instances and directs requests to an instance based on ... a set of virtual machines with each machine being assigned a workload d) All of the mentioned

Last Answer : All of the mentioned

Description : Point out the correct statement. a) A client can request access to a cloud service from any location b) A cloud has multiple application instances and directs requests to an instance based on ... a set of virtual machines with each machine being assigned a workload d) All of the mentioned

Last Answer : All of the mentioned

Description : Point out the correct statement. a) A client can request access to a cloud service from any location b) A cloud has multiple application instances and directs requests to an instance based on ... a set of virtual machines with each machine being assigned a workload d) All of the mentioned

Last Answer : All of the mentioned

Description : Which statement is true about NameNode High Availability a) Solve Single point of failure b) For high scalability c) Reduce storage overhead to 50% d) None of the above

Last Answer : Solve Single point of failure

Description : Above the file systems comes the ________ engine, which consists of one Job Tracker, to which client applications submit MapReduce jobs. a) MapReduce b) Google c) Functional programming d) Facebook

Last Answer : MapReduce

Description : In HDFS, reliability of storing data was maintained by a) Block Operations b) Block Replication c) Block Storage d) All the above

Last Answer : Block Replication

Description : .In computing, ________ improves the distribution of workloads across multiple computing resources, such as computers, a computer cluster, network links, central processing units, or disk drives. a) Virtual machine b) Virtual computing c) Virtual cloud d) Load balancer

Last Answer : Load balancer

Description : Which statement best describes the Software as a Service cloud delivery model? A. A virtual machine provisioned and provided from the cloud which allows the customer to deploy custom ... and run the application on the customer's own computers and simplifying maintenance and support.

Last Answer : An application delivered to the client from the cloud which eliminates the need to install and run the application on the customer's own computers and simplifying maintenance and support.

Description : Point out the wrong statement. a) A MapReduce job usually splits the input data-set into independent chunks which are processed by the map tasks in a completely parallel manner b) The ... implement the Mapper and Reducer interfaces toprovide the map and reduce methods d) None of the mentioned

Last Answer : None of the mentioned

Description : ________ is a utility which allows users to create and run jobs with any executables as the mapper and/or the reducer. a) Hadoop Strdata b) Hadoop Streaming c) Hadoop Stream d) None of the mentioned

Last Answer : Hadoop Streaming

Description : ________ is a utility which allows users to create and run jobs with any executables as the mapper and/or the reducer. a) Hadoop Strdata b) Hadoop Streaming c) Hadoop Stream d) None of the mentioned

Last Answer : Hadoop Streaming

Description : Which one of the following statements is true regarding pairs of a MapReduce job? a)A key class must implement Writable. b)A key class must implement WritableComparable. c)A value class must implement WritableComparable. d)A value class must extend WritableComparable.

Last Answer : A key class must implement WritableComparable.

Description : The Hadoop list includes the HBase database, the Apache Mahout ________ system, and matrix operations. a) Machine learning b) Pattern recognition c) Statistical classification d) Artificial intelligence

Last Answer : Machine learning

Description : MapReduce engine has a master/slave architecture consisting of a single ______ as the master and a number of _______ as the slaves a) Name Node, Data Node b) JobTracker, TaskTrackers c) TaskTrackers, JobTracker d) Data Node, Name Node

Last Answer : JobTracker, TaskTrackers

Description : Which statement is true about a bare metal hypervisor? A. It can only be hosted on an existing operating system. B. It requires a separate license for the native operating system. C ... system. D. It runs directly on server hardware to provide virtual machines with timesharing resources.

Last Answer : It runs directly on server hardware to provide virtual machines with timesharing resources.

Description : What are all the modular architecture for OpenStack are: a) OpenStack Compute: For overseeing extensive systems of the virtual machine. b) OpenStack Object Storage: The storage framework that offers help for both object storage ... to virtual disk pictures. d) A & B e) B & C f) All of the above

Last Answer : All of the above

Description : What are all the modular architecture for OpenStack are: a) OpenStack Compute: For overseeing extensive systems of the virtual machine. b) OpenStack Object Storage: The storage framework that offers help for both object storage ... to virtual disk pictures. d) A & B e) B & C f) All of the above

Last Answer : All of the above

Description : The Master of the HDFS Architecture is a) Node Manager b) Data Node c) Name Node d) None of the above

Last Answer : Name Node

Description : Point out the wrong statement. a) Securing data sent to, received from, and stored in the cloud is the single largest security concern b) The problem with the data you store in the cloud is that ... the cloud from direct client access is to create layered access to the data d) All of the mentioned

Last Answer : One and only approach to isolating storage in the cloud from direct client access is to create layered access to the data

Description : Point out the wrong statement. a) XML checks the ID request against its policies and either allows or denies the request b) SPML in most instances operates as a reply/request mechanism c) User centric authentication takes place usually in the form of knowledge cards d) All of the mentioned

Last Answer : SPML in most instances operates as a reply/request mechanism

Description : Hadoop achieves reliability by replicating the data across multiple hosts and hence does not require ________ storage on hosts. a) RAID b) Standard RAID levels c) ZFS d) Operating system

Last Answer : RAID

Description : _________ software library to write and run large user applications on vast data sets in business applications a) Apache Tomcat b) Hadoop c) Open Stack d) Open Nebula

Last Answer : Hadoop

Description : The comman used to copy a directory form one node to another in HDFS is a) rcp b) dcp c) drcp d) distcp

Last Answer : distcp

Description : Which of the following is a means for accessing human researchers or consultants to help solve problems on a contractual or temporary basis? a) Amazon Elastic MapReduce b) Amazon Mechanical Turk c) Amazon DevPay d) Multi-Factor Authentication

Last Answer : Amazon Mechanical Turk

Description : Which of the following is a billing and account management service? a) Amazon Elastic MapReduce b) Amazon Mechanical Turk c) Amazon DevPay d) Multi-Factor Authentication

Last Answer : Amazon DevPay

Description : The building blocks of Google’s App Engine architecture comprises of a) GFS, MapReduce, Chubby, BigTable b) GFS, Scheduler, Chubby, BigTable c) GFS, MapReduce, Chunk Server, BigTable d) Scheduler, MapReduce, Chubby, Chunk Server

Last Answer : GFS, MapReduce, Chubby, BigTable

Description : .A ________ node acts as the Slave and is responsible for executing a Task assigned to it by the JobTracker. a) MapReduce b) Mapper c) TaskTracker d) JobTracker

Last Answer : TaskTracker

Description : .A ________ node acts as the Slave and is responsible for executing a Task assigned to it by the JobTracker. a) MapReduce b) Mapper c) TaskTracker d) JobTracker

Last Answer : TaskTracker

Description : _______ in GAE provides coarse-grained locking service a) Bigtable b) Chubby c) MapReduce d) GFS

Last Answer : Chubby

Description : Google cloud SQL Features a) Lets your application read files from and write files to buckets in Google cloud storage b) A Fully -managed web service that allows you to create , ... large applications into logical components that can be share stateful services and communicate in secure fashion.

Last Answer : A Fully –managed web service that allows you to create , configure , and use relational databases that live in Google’s cloud

Description : The hdfs command to create the copy of a file from a local system is a) CopyFromLocal b) copyfromlocal c) CopyLocal d) copyFromLocal

Last Answer : copyFromLocal