Practice More Questions From: Spark Lesson 1
Q:
Apache Spark was developed in order to provide solutions to shortcomings of another project, and eventually replace it. What is the name of this project?
Q:
Why is Hadoop MapReduce slow for iterative algorithms?
Q:
What is the most important feature of Apache Spark to speedup iterative algorithms?
Q:
Which other Hadoop project can Spark rely to provision and manage the cluster of nodes?
Q:
When Spark reads data out of HDFS, what is the process that interfaces directly with HDFS?
Q:
Under which circumstances is preferable to run Spark in Standalone mode instead of relying on YARN?
Subscribe
0 Comments