Which one of the following statements is false regarding the Distributed
Cache?
a) The Hadoop framework will ensure that any files in the Distributed
Cache are distributed to all map and reduce tasks.
b) The files in the cache can be text files, or they can be archive files like
zip and JAR files.
c) Disk I/O is avoided because data in the cache is stored in memory.
d) The Hadoop framework will copy the files in the Distributed Cache on to
the slave node before any tasks for the job are executed on that node.

1 Answer

Answer :

MapReduce tries to place the data and the compute as close as
possible

Related questions

Description : Point out the wrong statement. a) A MapReduce job usually splits the input data-set into independent chunks which are processed by the map tasks in a completely parallel manner b) The ... implement the Mapper and Reducer interfaces toprovide the map and reduce methods d) None of the mentioned

Last Answer : None of the mentioned

Description : The Hadoop MapReduce framework spawns one map task for each __________ generated by the InputFormat for the job. a) OutputSplit b) InputSplit c) InputSplitStream d) All of the mentioned

Last Answer : ) InputSplit

Description : In a Hadoop cluster, what is true for a HDFS block that is no longer available due to disk corruption or machine failure? a)It is lost for ever b)It can be replicated form its alternative locations ... to read it. d)The Mapreduce job process runs ignoring the block and the data stored in it.

Last Answer : It can be replicated form its alternative locations to other live machines.

Description : Point out the correct statement. a) Hadoop is an ideal environment for extracting and transforming small volumes of data b) Hadoop stores data in HDFS and supports data compression/decompression c) The ... less useful than a MapReduce job to solve graph and machine learning d) None of the mentioned

Last Answer : Hadoop stores data in HDFS and supports data compression/decompression

Description : Point out the correct statement. a) Hadoop is an ideal environment for extracting and transforming small volumes of data b) Hadoop stores data in HDFS and supports data compression/decompression c) The ... useful than a MapReduce job to solve graph and machine learning d) None of the mentioned

Last Answer : Hadoop stores data in HDFS and supports data compression/decompression

Description : Point out the correct statement. a) Hadoop is an ideal environment for extracting and transforming small volumes of data b) Hadoop stores data in HDFS and supports data compression/decompression c) The ... useful than a MapReduce job to solve graph and machine learning d) None of the mentioned

Last Answer : Hadoop stores data in HDFS and supports data compression/decompression

Description : The process of the JobTracker creates one map task for each computed input split by the user node and assigns the map tasks to the execution slots of the TaskTrackers is called as a) Task Execution b) Task Running Check c) Job Submission d) Task Assignment

Last Answer : Task Assignment

Description : In Hadoop, the files are stored in a) Directory b) DFS c) GFS d) HDFS

Last Answer : HDFS

Description : Point out the wrong statement. a) The Mapper outputs are sorted and then partitioned per Reducer b) The total number of partitions is the same as the number of reduce tasks for the job c) The intermediate, ... always stored in a simple (key-len, key, value-len, value) format d) None of the mentioned

Last Answer : None of the mentioned

Description : Which of the following is built on top of a Hadoop framework using the Elastic Compute Cloud? a) Amazon Elastic MapReduce b) Amazon Mechanical Turkc) Amazon DevPay d) Multi-Factor Authentication

Last Answer : Amazon Elastic MapReduce

Description : Which of the following genres does Hadoop produce? a) Distributed file system b) JAX-RS c) Java Message Service d) Relational Database Management System

Last Answer : Distributed file system

Description : What license is Hadoop distributed under? a) Apache License 2.0 b) Mozilla Public License c) Shareware d) Commercia

Last Answer : Apache License 2.0

Description : IBM and ________ have announced a major initiative to use Hadoop to support university courses in distributed computer programming. a) Google Latitude b) Android (operating system) c) Google Variations d) Google

Last Answer : Google

Description : MapReduce engine has a master/slave architecture consisting of a single ______ as the master and a number of _______ as the slaves a) Name Node, Data Node b) JobTracker, TaskTrackers c) TaskTrackers, JobTracker d) Data Node, Name Node

Last Answer : JobTracker, TaskTrackers

Description : Which of the following information about the UNIX file system is not correct? (A) Super block contains the number of i-nodes, the number of disk blocks, and the start of the list of free disk blocks. (B ... Each i-node is 256-bytes long. (D) All the files and directories are stored in data blocks. 

Last Answer : (C) Each i-node is 256-bytes long.

Description : What is the default size of distributed cache a) 8 GB b) 10 GB c) 12 GB d) 16 GB

Last Answer : 10 GB

Description : .A ________ node acts as the Slave and is responsible for executing a Task assigned to it by the JobTracker. a) MapReduce b) Mapper c) TaskTracker d) JobTracker

Last Answer : TaskTracker

Description : .A ________ node acts as the Slave and is responsible for executing a Task assigned to it by the JobTracker. a) MapReduce b) Mapper c) TaskTracker d) JobTracker

Last Answer : TaskTracker

Description : In Replica placement how the three copies of data were stored? a) stores tworeplicas in the same node, one replica on a different node in the same rack b) stores one replica in the same node, two ... d) stores one replica in the same node, two replicas on a different node in a different rack

Last Answer : stores one replica in the same node, one replica on a different node but in the same rack, one replica on a different node in a different rack

Description : The topmost layer of Hadoop is the _________ engine a) HDFS b) Cluster c) MapReduce d) Job Tracker

Last Answer : MapReduce

Description : What are all the modular architecture for OpenStack are: a) OpenStack Compute: For overseeing extensive systems of the virtual machine. b) OpenStack Object Storage: The storage framework that offers help for both object storage ... to virtual disk pictures. d) A & B e) B & C f) All of the above

Last Answer : All of the above

Description : What are all the modular architecture for OpenStack are: a) OpenStack Compute: For overseeing extensive systems of the virtual machine. b) OpenStack Object Storage: The storage framework that offers help for both object storage ... to virtual disk pictures. d) A & B e) B & C f) All of the above

Last Answer : All of the above

Description : Hadoop achieves reliability by replicating the data across multiple hosts and hence does not require ________ storage on hosts. a) RAID b) Standard RAID levels c) ZFS d) Operating system

Last Answer : RAID

Description : The client reading the data from HDFS filesystem in Hadoop a) gets the data from the namenode b) gets the block location from the datanode c) gets only the block locations form the namenode d) gets both the data and block location from the namenode

Last Answer : gets only the block locations form the namenode

Description : _________ software library to write and run large user applications on vast data sets in business applications a) Apache Tomcat b) Hadoop c) Open Stack d) Open Nebula

Last Answer : Hadoop

Description : The comman used to copy a directory form one node to another in HDFS is a) rcp b) dcp c) drcp d) distcp

Last Answer : distcp

Description : Point out the correct statement. a) Mapper maps input key/value pairs to a set of intermediate key/value pairs b) Applications typically implement the Mapper and Reducer interfaces to provide the map and ... methods c) Mapper and Reducer interfaces form the core of the job d) None of the mentioned

Last Answer : None of the mentioned

Description : The framework groups Reducer inputs by key in _________ stage. a) sort b) shuffle c) reduce d) none of the mentioned

Last Answer : sort

Description : Which of the following is a web service that manages the memory cache in the cloud? A. Amazon Elastic Cache B. Amazon RDS C. RDMS on EC2 D. Auto Scaling

Last Answer : Amazon Elastic Cache

Description : Which of the following is false? a) The more important the data, the greater the need for backing it up b) A backup is as useful as its associated restore strategy c) Storing the ... near to its original site is best strategy d) Automated backup and scheduling is preferred over manual operations

Last Answer : Storing the backup copy near to its original site is best strategy

Description : Which of the following is false? a) The more important the data, the greater the need for backing it up b) A backup is as useful as its associated restore strategy c) Storing the ... near to its original site is best strategy d) Automated backup and scheduling is preferred over manual operations

Last Answer : Storing the backup copy near to its original site is best strategy

Description : Which memory is used to copy instructions or data currently used by CPU: a. Main memory b. Secondary memory c. Cache memory d. None of these

Last Answer : c. Cache memory

Description : Copy of instruction in cache memory is known as: a Execution cache b. Data cache c. Instruction cache d. Allof these

Last Answer : c. Instruction cache

Description : Copy of data in cache memory is called. a. Datacache b. Execution cache c. Address cache d. Control cache

Last Answer : a. Datacache

Description : Which of the following platforms does Hadoop run on? a) Bare metal b) Debian c) Cross-platform d) Unix-like

Last Answer : Cross-platform

Description : In some cases, it’s also hard to map an object to XML (e.g., an object graph with cyclic references) because there’s actually an impedance mismatch between an object model and an XML model. a) True b) False

Last Answer : True

Description : How would you use java & App Engine to write to the file system? a. Disconnect from the internet and Google App Engine automatically writes to the disk. b. Edit the configuration file accordingly c. ... Invoke an instance of Google App Engine within the sandbox and it will then write to the disk.

Last Answer : You cannot write t the file system with the exception of resources files.

Description : How would you use java & App Engine to write to the file system? a. Disconnect from the internet and Google App Engine automatically writes to the disk. b. Edit the configuration file accordingly c. ... Invoke an instance of Google App Engine within the sandbox and it will then write to the disk.

Last Answer : You cannot write t the file system with the exception of resources files.

Description : Which of the following can be identified as cloud? a) Web Applicationsb) Intranet c) Hadoop d) All of the mentioned

Last Answer : Hadoop

Description : The Hadoop list includes the HBase database, the Apache Mahout ________ system, and matrix operations. a) Machine learning b) Pattern recognition c) Statistical classification d) Artificial intelligence

Last Answer : Machine learning

Description : What was Hadoop written in? a) Java (software platform) b) Perl c) Java (programming language) d) Lua (programming language)

Last Answer : Java (programming language)

Description : Sun also has the Hadoop Live CD ________ project, which allows running a fully functional Hadoop cluster using a live CD. a) OpenOffice.org b) OpenSolaris c) GNU d) Linux

Last Answer : OpenSolaris

Description : Which of the following can be identified as cloud? a) Web Applications b) Intranet c) Hadoop d) All of the mentioned

Last Answer : Hadoop

Description : During Safemode Hadoop cluster is in a) Read-only b) Write-only c) Read-Write d) None of the above

Last Answer : Read-only

Description : Which of the following command is used to enter Safemode a) hadoop dfsadmin –safemode get b) bin dfsadmin –safemode get c) hadoop dfsadmin –safemode enter d) None of the above

Last Answer : hadoop dfsadmin –safemode enter

Description : What are the advantages of HDFS federation in Hadoop? a) Isolation b) Namespace scalability c) Improves throughput d) All of the above

Last Answer : All of the above

Description : Under Hadoop High Availability, Fencing means a)Preventing a previously active namenode from start running again. b)Preventing the start of a failover in the event of network failure with the ... previously active namenode. d)Preventing a previously active namenode from writing to the edit log.

Last Answer : Preventing a previously active namenode from writing to the edit log.

Description : ________ is a utility which allows users to create and run jobs with any executables as the mapper and/or the reducer. a) Hadoop Strdata b) Hadoop Streaming c) Hadoop Stream d) None of the mentioned

Last Answer : Hadoop Streaming

Description : ________ is a utility which allows users to create and run jobs with any executables as the mapper and/or the reducer. a) Hadoop Strdata b) Hadoop Streaming c) Hadoop Stream d) None of the mentioned

Last Answer : Hadoop Streaming

Description : Hadoop was originally developed by a) Microsoft b) Google c) Yahoo d) IBM

Last Answer : Yahoo