Databricks Certified Associate Developer for Apache Spark - Certified Associate Developer for Apache Spark Exam
Page: 1 / 42
Total 206 questions
Question #1 (Topic: Exam A)
Which of the following describes the Spark driver?
A. The Spark driver is responsible for performing all execution in all execution modes – it is the entire Spark application.
B. The Spare driver is fault tolerant – if it fails, it will recover the entire Spark application.
C. The Spark driver is the coarsest level of the Spark execution hierarchy – it is synonymous with the Spark application.
D. The Spark driver is the program space in which the Spark application’s main method runs coordinating the Spark entire application.
E. The Spark driver is horizontally scaled to increase overall processing throughput of a Spark application.
Answer: D
Question #2 (Topic: Exam A)
Which of the following describes the relationship between nodes and executors?
A. Executors and nodes are not related.
B. Anode is a processing engine running on an executor.
C. An executor is a processing engine running on a node.
D. There are always the same number of executors and nodes.
E. There are always more nodes than executors.
Answer: C
Question #3 (Topic: Exam A)
Which of the following will occur if there are more slots than there are tasks?
A. The Spark job will likely not run as efficiently as possible.
B. The Spark application will fail – there must be at least as many tasks as there are slots.
C. Some executors will shut down and allocate all slots on larger executors first.
D. More tasks will be automatically generated to ensure all slots are being used.
E. The Spark job will use just one single slot to perform all tasks.
Answer: A
Question #4 (Topic: Exam A)
Which of the following is the most granular level of the Spark execution hierarchy?
A. Task
B. Executor
C. Node
D. Job
E. Slot
Answer: A
Question #5 (Topic: Exam A)
Which of the following statements about Spark jobs is incorrect?
A. Jobs are broken down into stages.
B. There are multiple tasks within a single job when a DataFrame has more than one partition.
C. Jobs are collections of tasks that are divided up based on when an action is called.
D. There is no way to monitor the progress of a job.
E. Jobs are collections of tasks that are divided based on when language variables are defined.
Answer: D