Very large sustainable reading and writing bandwidth, mostly continuous
accessing instead of random accessing. The programming interface is similar to
that of the POSIX file system accessing interface. This is the feature of
a) Amazon S3 and EBS
b) GFS: Google File System
c) HDFS: Hadoop Distributes File System
d) None of the above

1 Answer

Answer :

GFS: Google File System

Related questions

Description : . _____________ The open source clone of GFS. Written in Java. The programming interfaces are similar to POSIX but not identical. a) HDFS b) SAMIL c) SCIM d) DVE

Last Answer : HDFS

Description : In Hadoop, the files are stored in a) Directory b) DFS c) GFS d) HDFS

Last Answer : HDFS

Description : _________ is a web programming model for scalable data processing on large clusters over large data sets a) MapReduce b) HDFS c) GFS d) DFS

Last Answer : MapReduce

Description : The client reading the data from HDFS filesystem in Hadoop a) gets the data from the namenode b) gets the block location from the datanode c) gets only the block locations form the namenode d) gets both the data and block location from the namenode

Last Answer : gets only the block locations form the namenode

Description : Point out the correct statement. a) Eucalyptus and Rackspace both use Amazon EC2 and S3 services b) The RightScale user interface provides real-time measurements of individual server instances c) ... technology are highly configurable and can be run under batch control d) All of the mentioned

Last Answer : All of the mentioned

Description : . ___________ was built primarily as the fundamental storage service for Google’s search engine. a) BFS b) HDFS c) GFS d) GAE

Last Answer : 200 TB

Description : ___________ was built primarily as the fundamental storage service for Google’s search engine. a) BFS b) HDFS c) GFS d) GAE

Last Answer : GFS

Description : Point out the correct statement. a) Hadoop is an ideal environment for extracting and transforming small volumes of data b) Hadoop stores data in HDFS and supports data compression/decompression c) The ... less useful than a MapReduce job to solve graph and machine learning d) None of the mentioned

Last Answer : Hadoop stores data in HDFS and supports data compression/decompression

Description : Point out the correct statement. a) Hadoop is an ideal environment for extracting and transforming small volumes of data b) Hadoop stores data in HDFS and supports data compression/decompression c) The ... useful than a MapReduce job to solve graph and machine learning d) None of the mentioned

Last Answer : Hadoop stores data in HDFS and supports data compression/decompression

Description : Point out the correct statement. a) Hadoop is an ideal environment for extracting and transforming small volumes of data b) Hadoop stores data in HDFS and supports data compression/decompression c) The ... useful than a MapReduce job to solve graph and machine learning d) None of the mentioned

Last Answer : Hadoop stores data in HDFS and supports data compression/decompression

Description : What are the advantages of HDFS federation in Hadoop? a) Isolation b) Namespace scalability c) Improves throughput d) All of the above

Last Answer : All of the above

Description : In a Hadoop cluster, what is true for a HDFS block that is no longer available due to disk corruption or machine failure? a)It is lost for ever b)It can be replicated form its alternative locations ... to read it. d)The Mapreduce job process runs ignoring the block and the data stored in it.

Last Answer : It can be replicated form its alternative locations to other live machines.

Description : The topmost layer of Hadoop is the _________ engine a) HDFS b) Cluster c) MapReduce d) Job Tracker

Last Answer : MapReduce

Description : IBM and ________ have announced a major initiative to use Hadoop to support university courses in distributed computer programming. a) Google Latitude b) Android (operating system) c) Google Variations d) Google

Last Answer : Google

Description : In Google File System (GFS) the file name, file index was maintained by a) GFS Chunk Server b) GFS Master c) Linux File System d) None of the above

Last Answer : GFS Master

Description : Amazon EBS volumes can be maximized up to ? A. 1 TB B. 2 TB C. 3 TB D. 4 TB

Last Answer : 1 TB

Description : Amazon EBS currently supports up to ________ IOPS per volume A. 10 B. 100 C. 1000 D. 10000

Last Answer : 1000

Description : Point out the correct statement. a) Amazon Elastic Cloud is a system for creating virtual disks(volume) b) SimpleDB interoperates with both Amazon EC2 and Amazon S3 c) EC3 is an Analytics as a Service provider d) None of the mentioned

Last Answer : SimpleDB interoperates with both Amazon EC2 and Amazon S3

Description : Which of the following is built on top of a Hadoop framework using the Elastic Compute Cloud? a) Amazon Elastic MapReduce b) Amazon Mechanical Turkc) Amazon DevPay d) Multi-Factor Authentication

Last Answer : Amazon Elastic MapReduce

Description : Amazon S3 stores data as objects within resources called ? A. Auto Scaling B. EC2 Instances C. Buckets D. Amazon RDS

Last Answer : Buckets

Description : Point out the wrong statement. a) AWS S3 essentially lets you create your own cloud storage b) AWS created Availability Zones within regions, which are sets of systems that are isolated from ... adds redundancy to its IaaS systems by allowing EC2 virtual machine instances d) None of the mentioned

Last Answer : None of the mentioned

Description : Which of the following is a structured data store that supports indexing and data queries to both EC2 and S3? a) CloudWatch b) Amazon SimpleDB c) Amazon Cloudfront d) All of the mentioned

Last Answer : Amazon SimpleDB

Description : What was Hadoop written in? a) Java (software platform) b) Perl c) Java (programming language) d) Lua (programming language)

Last Answer : Java (programming language)

Description : Which of the following is a means for accessing human researchers or consultants to help solve problems on a contractual or temporary basis? a) Amazon Elastic MapReduce b) Amazon Mechanical Turk c) Amazon DevPay d) Multi-Factor Authentication

Last Answer : Amazon Mechanical Turk

Description : The hdfs command to create the copy of a file from a local system is a) CopyFromLocal b) copyfromlocal c) CopyLocal d) copyFromLocal

Last Answer : copyFromLocal

Description : Under Hadoop High Availability, Fencing means a)Preventing a previously active namenode from start running again. b)Preventing the start of a failover in the event of network failure with the ... previously active namenode. d)Preventing a previously active namenode from writing to the edit log.

Last Answer : Preventing a previously active namenode from writing to the edit log.

Description : Which of the following genres does Hadoop produce? a) Distributed file system b) JAX-RS c) Java Message Service d) Relational Database Management System

Last Answer : Distributed file system

Description : To read a file in HDFS, a user sends an _______ request to the NameNode toget the location of file blocks. a) access b) open c) authentication d) create

Last Answer : open

Description : Hadoop was originally developed by a) Microsoft b) Google c) Yahoo d) IBM

Last Answer : Yahoo

Description : Which of the following is an industry organization that develops industry system management standards for platform interoperability? a) DMTF b) DMS c) EBS d) All of the mentioned

Last Answer : DMTF

Description : Above the file systems comes the ________ engine, which consists of one Job Tracker, to which client applications submit MapReduce jobs. a) MapReduce b) Google c) Functional programming d) Facebook

Last Answer : MapReduce

Description : _________ software library to write and run large user applications on vast data sets in business applications a) Apache Tomcat b) Hadoop c) Open Stack d) Open Nebula

Last Answer : Hadoop

Description : Which of the following describes a message-passing taxonomy for a component- based architecture that provides services to clients upon demand? a) SOA b) EBS c) GEC d) All of the mentioned

Last Answer : SOA

Description : Which of the following describes a message-passing taxonomy for a component- based architecture that provides services to clients upon demand? a) SOA b) EBS c) GEC d) All of the mentioned

Last Answer : SOA

Description : What facet of cloud computing helps to guard against downtime and determines costs? A. Service-level agreements B. Application programming interfaces C. Virtual private networks D. Bandwidth fees

Last Answer : Service-level agreements

Description : The building blocks of Google’s App Engine architecture comprises of a) GFS, MapReduce, Chubby, BigTable b) GFS, Scheduler, Chubby, BigTable c) GFS, MapReduce, Chunk Server, BigTable d) Scheduler, MapReduce, Chubby, Chunk Server

Last Answer : GFS, MapReduce, Chubby, BigTable

Description : With the current quality ofcommodity servers, the single master can handle a cluster of more than _____ nodes in GFS, a) 1000 b) 500 c) 200 d) 120

Last Answer : 1000

Description : GFS makes checksums on every _________ block in each chunk. a) 64 MB b) 64 KB c) 4MB d) 4KB

Last Answer : 64 KB

Description : With the current quality ofcommodity servers, the single master can handle a cluster of more than _____ nodes in GFS, a) 1000 b) 500 c) 200 d) 120

Last Answer : 1000

Description : GFS makes checksums on every _________ block in each chunk. a) 64 MB b) 64 KB c) 4MB d) 4KB

Last Answer : 64 KB

Description : _______ in GAE provides coarse-grained locking service a) Bigtable b) Chubby c) MapReduce d) GFS

Last Answer : Chubby

Description : A component called the ____________ performs resource allocation and management on behalf of specific user applications. a) IGG manager b) GFS manager c) DVE manager d) DVA manager

Last Answer : DVE manager

Description : _________ serves as a PaaS vendor within Google App Engine system. a) Google b) Amazon c) Microsoft d) All of the mentioned

Last Answer : Google

Description : _________ serves as a PaaS vendor within Google App Engine system. a) Google b) Amazon c) Microsoft d) All of the mentioned

Last Answer : Google

Description : The Hadoop list includes the HBase database, the Apache Mahout ________ system, and matrix operations. a) Machine learning b) Pattern recognition c) Statistical classification d) Artificial intelligence

Last Answer : Machine learning

Description : Hadoop achieves reliability by replicating the data across multiple hosts and hence does not require ________ storage on hosts. a) RAID b) Standard RAID levels c) ZFS d) Operating system

Last Answer : RAID

Description : The comman used to copy a directory form one node to another in HDFS is a) rcp b) dcp c) drcp d) distcp

Last Answer : distcp

Description : The architecture for _______ is built on the concepts of shared-nothing and messaging-based information exchange. a) HDFS b) Chabby c) Nova d) BigTable

Last Answer : Nova

Description : The architecture for _______ is built on the concepts of shared-nothing and messaging-based information exchange. a) HDFS b) Chabby c) Nova d) BigTable

Last Answer : Nova

Description : The main goal of HDFS High availability is a) Faster creation of the replicas of primary namenode. b) To reduce the cycle time required to bring back a new primary namenode after existing ... data loss due to failure of primary namenode. d) Prevent the primary namenode form becoming single point

Last Answer : To reduce the cycle time required to bring back a new primary namenode after existing primary fails.