_________ is a web programming model for scalable data processing on
large clusters over large data sets
a) MapReduce
b) HDFS
c) GFS
d) DFS

1 Answer

Answer :

MapReduce

Related questions

Description : In Hadoop, the files are stored in a) Directory b) DFS c) GFS d) HDFS

Last Answer : HDFS

Description : Very large sustainable reading and writing bandwidth, mostly continuous accessing instead of random accessing. The programming interface is similar to that of the POSIX file system accessing interface. This is ... : Google File System c) HDFS: Hadoop Distributes File System d) None of the above

Last Answer : GFS: Google File System

Description : . _____________ The open source clone of GFS. Written in Java. The programming interfaces are similar to POSIX but not identical. a) HDFS b) SAMIL c) SCIM d) DVE

Last Answer : HDFS

Description : The topmost layer of Hadoop is the _________ engine a) HDFS b) Cluster c) MapReduce d) Job Tracker

Last Answer : MapReduce

Description : Point out the correct statement. a) Hadoop is an ideal environment for extracting and transforming small volumes of data b) Hadoop stores data in HDFS and supports data compression/decompression c) The ... less useful than a MapReduce job to solve graph and machine learning d) None of the mentioned

Last Answer : Hadoop stores data in HDFS and supports data compression/decompression

Description : Point out the correct statement. a) Hadoop is an ideal environment for extracting and transforming small volumes of data b) Hadoop stores data in HDFS and supports data compression/decompression c) The ... useful than a MapReduce job to solve graph and machine learning d) None of the mentioned

Last Answer : Hadoop stores data in HDFS and supports data compression/decompression

Description : Point out the correct statement. a) Hadoop is an ideal environment for extracting and transforming small volumes of data b) Hadoop stores data in HDFS and supports data compression/decompression c) The ... useful than a MapReduce job to solve graph and machine learning d) None of the mentioned

Last Answer : Hadoop stores data in HDFS and supports data compression/decompression

Description : In a Hadoop cluster, what is true for a HDFS block that is no longer available due to disk corruption or machine failure? a)It is lost for ever b)It can be replicated form its alternative locations ... to read it. d)The Mapreduce job process runs ignoring the block and the data stored in it.

Last Answer : It can be replicated form its alternative locations to other live machines.

Description : The building blocks of Google’s App Engine architecture comprises of a) GFS, MapReduce, Chubby, BigTable b) GFS, Scheduler, Chubby, BigTable c) GFS, MapReduce, Chunk Server, BigTable d) Scheduler, MapReduce, Chubby, Chunk Server

Last Answer : GFS, MapReduce, Chubby, BigTable

Description : _______ in GAE provides coarse-grained locking service a) Bigtable b) Chubby c) MapReduce d) GFS

Last Answer : Chubby

Description : . ___________ was built primarily as the fundamental storage service for Google’s search engine. a) BFS b) HDFS c) GFS d) GAE

Last Answer : 200 TB

Description : ___________ was built primarily as the fundamental storage service for Google’s search engine. a) BFS b) HDFS c) GFS d) GAE

Last Answer : GFS

Description : Point out the correct statement. a) With PaaS, the goal is to create hosted scalable applications that are used in a Software as a Service model b) Applications built using PaaS tools need to ... , the data and transaction management conforms to the business rules you create d) All of the mentioned

Last Answer : All of the mentioned

Description : Point out the correct statement. a) With PaaS, the goal is to create hosted scalable applications that are used in a Software as a Service model b) Applications built using PaaS tools need to ... the data and transaction management conforms to the business rules you create d) All of the mentioned

Last Answer : All of the mentioned

Description : Above the file systems comes the ________ engine, which consists of one Job Tracker, to which client applications submit MapReduce jobs. a) MapReduce b) Google c) Functional programming d) Facebook

Last Answer : MapReduce

Description : GFS makes checksums on every _________ block in each chunk. a) 64 MB b) 64 KB c) 4MB d) 4KB

Last Answer : 64 KB

Description : GFS makes checksums on every _________ block in each chunk. a) 64 MB b) 64 KB c) 4MB d) 4KB

Last Answer : 64 KB

Description : _________ software library to write and run large user applications on vast data sets in business applications a) Apache Tomcat b) Hadoop c) Open Stack d) Open Nebula

Last Answer : Hadoop

Description : The gateway administrator provides the _________ template information when the infrastructure is set up. a) Cloud b) HDFS c) better d) VM

Last Answer : VM

Description : ______________ is one of the heterogeneous benefits of storage virtualization. A. Data migration. B. Scalable. C. Effective use of capacity. D. Lower total cost of ownership.

Last Answer : Data migration.

Description : Point out the wrong statement. a) A MapReduce job usually splits the input data-set into independent chunks which are processed by the map tasks in a completely parallel manner b) The ... implement the Mapper and Reducer interfaces toprovide the map and reduce methods d) None of the mentioned

Last Answer : None of the mentioned

Description : MapReduce engine has a master/slave architecture consisting of a single ______ as the master and a number of _______ as the slaves a) Name Node, Data Node b) JobTracker, TaskTrackers c) TaskTrackers, JobTracker d) Data Node, Name Node

Last Answer : JobTracker, TaskTrackers

Description : The client reading the data from HDFS filesystem in Hadoop a) gets the data from the namenode b) gets the block location from the datanode c) gets only the block locations form the namenode d) gets both the data and block location from the namenode

Last Answer : gets only the block locations form the namenode

Description : The main goal of HDFS High availability is a) Faster creation of the replicas of primary namenode. b) To reduce the cycle time required to bring back a new primary namenode after existing ... data loss due to failure of primary namenode. d) Prevent the primary namenode form becoming single point

Last Answer : To reduce the cycle time required to bring back a new primary namenode after existing primary fails.

Description : In HDFS, reliability of storing data was maintained by a) Block Operations b) Block Replication c) Block Storage d) All the above

Last Answer : Block Replication

Description : The Master of the HDFS Architecture is a) Node Manager b) Data Node c) Name Node d) None of the above

Last Answer : Name Node

Description : A _______ system uses components to assemble services that can be tailored for a specific purpose using standard parts. a) reliable b) scalable c) composable d) none of the mentioned

Last Answer : composable

Description : Point out the wrong statement. a) The massive scale of cloud computing systems was enabled by the popularization of the Internet b) Soft computing represents a real paradigm shift in the way in ... with a pay-as-you-go, infinitely scalable, universally available system d) All of the mentioned

Last Answer : Soft computing represents a real paradigm shift in the way in which systems are deployed

Description : Which of the following is not a characteristic of Private Cloud? a) Secured b) Central Controlc) Scalable d) Weak Service Level Agreements (SLAs)

Last Answer : Scalable

Description : Point out the wrong statement. a) The massive scale of cloud computing systems was enabled by the popularization of the Internet b) Soft computing represents a real paradigm shift in the way in ... with a pay-as-you-go, infinitely scalable, universally available system d) All of the mentioned

Last Answer : Soft computing represents a real paradigm shift in the way in which systems are deployed

Description : In Google File System (GFS) the file name, file index was maintained by a) GFS Chunk Server b) GFS Master c) Linux File System d) None of the above

Last Answer : GFS Master

Description : With the current quality ofcommodity servers, the single master can handle a cluster of more than _____ nodes in GFS, a) 1000 b) 500 c) 200 d) 120

Last Answer : 1000

Description : With the current quality ofcommodity servers, the single master can handle a cluster of more than _____ nodes in GFS, a) 1000 b) 500 c) 200 d) 120

Last Answer : 1000

Description : A component called the ____________ performs resource allocation and management on behalf of specific user applications. a) IGG manager b) GFS manager c) DVE manager d) DVA manager

Last Answer : DVE manager

Description : The Hadoop MapReduce framework spawns one map task for each __________ generated by the InputFormat for the job. a) OutputSplit b) InputSplit c) InputSplitStream d) All of the mentioned

Last Answer : ) InputSplit

Description : Which of the following is built on top of a Hadoop framework using the Elastic Compute Cloud? a) Amazon Elastic MapReduce b) Amazon Mechanical Turkc) Amazon DevPay d) Multi-Factor Authentication

Last Answer : Amazon Elastic MapReduce

Description : Which of the following is a means for accessing human researchers or consultants to help solve problems on a contractual or temporary basis? a) Amazon Elastic MapReduce b) Amazon Mechanical Turk c) Amazon DevPay d) Multi-Factor Authentication

Last Answer : Amazon Mechanical Turk

Description : Which of the following is a billing and account management service? a) Amazon Elastic MapReduce b) Amazon Mechanical Turk c) Amazon DevPay d) Multi-Factor Authentication

Last Answer : Amazon DevPay

Description : Which one of the following statements is true regarding pairs of a MapReduce job? a)A key class must implement Writable. b)A key class must implement WritableComparable. c)A value class must implement WritableComparable. d)A value class must extend WritableComparable.

Last Answer : A key class must implement WritableComparable.

Description : .A ________ node acts as the Slave and is responsible for executing a Task assigned to it by the JobTracker. a) MapReduce b) Mapper c) TaskTracker d) JobTracker

Last Answer : TaskTracker

Description : .A ________ node acts as the Slave and is responsible for executing a Task assigned to it by the JobTracker. a) MapReduce b) Mapper c) TaskTracker d) JobTracker

Last Answer : TaskTracker

Description : Network bottlenecks occur when ______ data sets must be transferred. a) large b) small c) big d) all of the mentioned

Last Answer : large

Description : What are the advantages of HDFS federation in Hadoop? a) Isolation b) Namespace scalability c) Improves throughput d) All of the above

Last Answer : All of the above

Description : The comman used to copy a directory form one node to another in HDFS is a) rcp b) dcp c) drcp d) distcp

Last Answer : distcp

Description : The hdfs command to create the copy of a file from a local system is a) CopyFromLocal b) copyfromlocal c) CopyLocal d) copyFromLocal

Last Answer : copyFromLocal

Description : The architecture for _______ is built on the concepts of shared-nothing and messaging-based information exchange. a) HDFS b) Chabby c) Nova d) BigTable

Last Answer : Nova

Description : The architecture for _______ is built on the concepts of shared-nothing and messaging-based information exchange. a) HDFS b) Chabby c) Nova d) BigTable

Last Answer : Nova

Description : To read a file in HDFS, a user sends an _______ request to the NameNode toget the location of file blocks. a) access b) open c) authentication d) create

Last Answer : open

Description : SaaS supports multiple users and provides a shared data model through _________ model. a) single-tenancy b) multi-tenancy c) multiple-instance d) all of the mentioned

Last Answer : multi-tenancy

Description : SaaS supports multiple users and provides a shared data model through _________ model. A. single-tenancy B. multi-tenancy C. multiple-instance D. None of the above

Last Answer : multi-tenancy