During Safemode Hadoop cluster is in
a) Read-only
b) Write-only
c) Read-Write
d) None of the above

1 Answer

Answer :

Read-only

Related questions

Description : Which of the following command is used to enter Safemode a) hadoop dfsadmin –safemode get b) bin dfsadmin –safemode get c) hadoop dfsadmin –safemode enter d) None of the above

Last Answer : hadoop dfsadmin –safemode enter

Description : In a Hadoop cluster, what is true for a HDFS block that is no longer available due to disk corruption or machine failure? a)It is lost for ever b)It can be replicated form its alternative locations ... to read it. d)The Mapreduce job process runs ignoring the block and the data stored in it.

Last Answer : It can be replicated form its alternative locations to other live machines.

Description : Sun also has the Hadoop Live CD ________ project, which allows running a fully functional Hadoop cluster using a live CD. a) OpenOffice.org b) OpenSolaris c) GNU d) Linux

Last Answer : OpenSolaris

Description : The topmost layer of Hadoop is the _________ engine a) HDFS b) Cluster c) MapReduce d) Job Tracker

Last Answer : MapReduce

Description : _________ software library to write and run large user applications on vast data sets in business applications a) Apache Tomcat b) Hadoop c) Open Stack d) Open Nebula

Last Answer : Hadoop

Description : The client reading the data from HDFS filesystem in Hadoop a) gets the data from the namenode b) gets the block location from the datanode c) gets only the block locations form the namenode d) gets both the data and block location from the namenode

Last Answer : gets only the block locations form the namenode

Description : Which of the following can be identified as cloud? a) Web Applicationsb) Intranet c) Hadoop d) All of the mentioned

Last Answer : Hadoop

Description : The Hadoop MapReduce framework spawns one map task for each __________ generated by the InputFormat for the job. a) OutputSplit b) InputSplit c) InputSplitStream d) All of the mentioned

Last Answer : ) InputSplit

Description : The Hadoop list includes the HBase database, the Apache Mahout ________ system, and matrix operations. a) Machine learning b) Pattern recognition c) Statistical classification d) Artificial intelligence

Last Answer : Machine learning

Description : Hadoop achieves reliability by replicating the data across multiple hosts and hence does not require ________ storage on hosts. a) RAID b) Standard RAID levels c) ZFS d) Operating system

Last Answer : RAID

Description : Which of the following platforms does Hadoop run on? a) Bare metal b) Debian c) Cross-platform d) Unix-like

Last Answer : Cross-platform

Description : What was Hadoop written in? a) Java (software platform) b) Perl c) Java (programming language) d) Lua (programming language)

Last Answer : Java (programming language)

Description : Which of the following genres does Hadoop produce? a) Distributed file system b) JAX-RS c) Java Message Service d) Relational Database Management System

Last Answer : Distributed file system

Description : What license is Hadoop distributed under? a) Apache License 2.0 b) Mozilla Public License c) Shareware d) Commercia

Last Answer : Apache License 2.0

Description : Point out the correct statement. a) Hadoop is an ideal environment for extracting and transforming small volumes of data b) Hadoop stores data in HDFS and supports data compression/decompression c) The ... less useful than a MapReduce job to solve graph and machine learning d) None of the mentioned

Last Answer : Hadoop stores data in HDFS and supports data compression/decompression

Description : IBM and ________ have announced a major initiative to use Hadoop to support university courses in distributed computer programming. a) Google Latitude b) Android (operating system) c) Google Variations d) Google

Last Answer : Google

Description : Which of the following is built on top of a Hadoop framework using the Elastic Compute Cloud? a) Amazon Elastic MapReduce b) Amazon Mechanical Turkc) Amazon DevPay d) Multi-Factor Authentication

Last Answer : Amazon Elastic MapReduce

Description : Which of the following can be identified as cloud? a) Web Applications b) Intranet c) Hadoop d) All of the mentioned

Last Answer : Hadoop

Description : Point out the correct statement. a) Hadoop is an ideal environment for extracting and transforming small volumes of data b) Hadoop stores data in HDFS and supports data compression/decompression c) The ... useful than a MapReduce job to solve graph and machine learning d) None of the mentioned

Last Answer : Hadoop stores data in HDFS and supports data compression/decompression

Description : Point out the correct statement. a) Hadoop is an ideal environment for extracting and transforming small volumes of data b) Hadoop stores data in HDFS and supports data compression/decompression c) The ... useful than a MapReduce job to solve graph and machine learning d) None of the mentioned

Last Answer : Hadoop stores data in HDFS and supports data compression/decompression

Description : Which one of the following statements is false regarding the Distributed Cache? a) The Hadoop framework will ensure that any files in the Distributed Cache are distributed to all map and reduce tasks. b) ... Cache on to the slave node before any tasks for the job are executed on that node.

Last Answer : MapReduce tries to place the data and the compute as close as possible

Description : What are the advantages of HDFS federation in Hadoop? a) Isolation b) Namespace scalability c) Improves throughput d) All of the above

Last Answer : All of the above

Description : Under Hadoop High Availability, Fencing means a)Preventing a previously active namenode from start running again. b)Preventing the start of a failover in the event of network failure with the ... previously active namenode. d)Preventing a previously active namenode from writing to the edit log.

Last Answer : Preventing a previously active namenode from writing to the edit log.

Description : ________ is a utility which allows users to create and run jobs with any executables as the mapper and/or the reducer. a) Hadoop Strdata b) Hadoop Streaming c) Hadoop Stream d) None of the mentioned

Last Answer : Hadoop Streaming

Description : ________ is a utility which allows users to create and run jobs with any executables as the mapper and/or the reducer. a) Hadoop Strdata b) Hadoop Streaming c) Hadoop Stream d) None of the mentioned

Last Answer : Hadoop Streaming

Description : In Hadoop, the files are stored in a) Directory b) DFS c) GFS d) HDFS

Last Answer : HDFS

Description : Hadoop was originally developed by a) Microsoft b) Google c) Yahoo d) IBM

Last Answer : Yahoo

Description : Very large sustainable reading and writing bandwidth, mostly continuous accessing instead of random accessing. The programming interface is similar to that of the POSIX file system accessing interface. This is ... : Google File System c) HDFS: Hadoop Distributes File System d) None of the above

Last Answer : GFS: Google File System

Description : With the current quality ofcommodity servers, the single master can handle a cluster of more than _____ nodes in GFS, a) 1000 b) 500 c) 200 d) 120

Last Answer : 1000

Description : With the current quality ofcommodity servers, the single master can handle a cluster of more than _____ nodes in GFS, a) 1000 b) 500 c) 200 d) 120

Last Answer : 1000

Description : .In computing, ________ improves the distribution of workloads across multiple computing resources, such as computers, a computer cluster, network links, central processing units, or disk drives. a) Virtual machine b) Virtual computing c) Virtual cloud d) Load balancer

Last Answer : Load balancer

Description : The system is notified of a read or write operation by ___________ a) Appending an extra bit of the address b) Enabling the read or write bits of the devices c) Raising an appropriate interrupt signal d) Sending a special signal along the BUS

Last Answer : Sending a special signal along the BUS

Description : Google cloud SQL Features a) Lets your application read files from and write files to buckets in Google cloud storage b) A Fully -managed web service that allows you to create , ... large applications into logical components that can be share stateful services and communicate in secure fashion.

Last Answer : A Fully –managed web service that allows you to create , configure , and use relational databases that live in Google’s cloud

Description : .Which technology allows use of read/write web, blogs & interactive web application etc. ? a) Mobile Computing b) Grid Computing c) Cloud Computing d) Web 2.0

Last Answer : Web 2.0

Description : The disk drive component used to position read/write heads over a specific track I known as a. Acoustic couples b. Access arm c. Cluster d. All of the above

Last Answer : Access arm

Description : The disk drive component used to position read/write heads over a specific track I known as a. Acoustic couples b. Access arm c. Cluster d. All of the above

Last Answer : b. Access arm

Description : The disk drive component used to position read/write heads over a specific track I known as a. Acoustic couples b. Access arm c. Cluster d. All of the above

Last Answer : b. Access arm

Description : What are two common concerns in a cloud environment? (Choose two.) A. Inability to use proxy servers and load balancers. B. Not enough computing capacity during peak usage times. C. Illegal access ... same physical ethernet, they should not be able to read or modify each other's network traffic.

Last Answer : Illegal access to private data stored by applications in the cloud.If two guest instances are running on the same host or on different hosts connected to the same physical ethernet, they should not be able to read or modify each other's network traffic

Description : To read a file in HDFS, a user sends an _______ request to the NameNode toget the location of file blocks. a) access b) open c) authentication d) create

Last Answer : open

Description : On what concept the Hadoop framework works?

Last Answer : Learn Hadoop Online Training to build your big data analytics and data processing file system skills today. Become familiar with Hadoop cluster, Hadoop distributed file system, Hadoop map ... It adjusts processing load to available resources, reducing downtime.The Hadoop software library is dev

Description : Are you choosing between Spark and Hadoop for your development project?

Last Answer : Both Hadoop and Spark are open source projects by Apache Software Foundation and both are the flagship products in big data analytics. Hadoop has been leading the big data market for more than 5 ... Hadoop's installed base amounts to 50,000+ customers, while Spark boasts 10,000+ installations only.

Description : Best Hadoop online Training?

Last Answer : Answers is the place to go to get the answers you need and to ask the questions you want

Description : Which of the following allows a distributed ID system to write and enforce custom policy expressions? a) XACML b) XML c) SOAP d) None of the mentioned

Last Answer : XACML

Description : Which of the following allows a distributed ID system to write and enforce custom policy expressions? a) XACML b) XML c) SOAP d) None of the mentioned

Last Answer : XACML

Description : How would you use java & App Engine to write to the file system? a. Disconnect from the internet and Google App Engine automatically writes to the disk. b. Edit the configuration file accordingly c. ... Invoke an instance of Google App Engine within the sandbox and it will then write to the disk.

Last Answer : You cannot write t the file system with the exception of resources files.

Description : How would you use java & App Engine to write to the file system? a. Disconnect from the internet and Google App Engine automatically writes to the disk. b. Edit the configuration file accordingly c. ... Invoke an instance of Google App Engine within the sandbox and it will then write to the disk.

Last Answer : You cannot write t the file system with the exception of resources files.

Description : What are two common concerns in a cloud environment? A. Inability to use proxy servers and load balancers. B. Not enough computing capacity during peak usage times. C. Illegal access to private ... in the cloud. D. Running out of storage space and network bandwidth at peak usage times.

Last Answer : Illegal access to private data stored by applications in the cloud.

Description : Point out the correct statement. a) To determine whether your application will port successfully, you should perform a functionality mapping exercise b) Cloud computing supports some application features better ... directs traffic to the cloud during times of high traffic d) All of the mentioned

Last Answer : All of the mentioned

Description : Point out the correct statement. a) To determine whether your application will port successfully, you should perform a functionality mapping exercise b) Cloud computing supports some application features better ... directs traffic to the cloud during times of high traffic d) All of the mentioned

Last Answer : All of the mentioned

Description : Is it possible for a star cluster to change it metallicity during time by YSO's appearance? -Space

Last Answer : yes,Since all the stars in a cluster formed at the same time, they are all the same age. A very young cluster will have a HR diagram with a cluster of T-Tauri stars evolving ... The rest become gravitationally bound, meaning they will exist as collection orbiting each other forever. thank you