Welcome to ExamTopics
ExamTopics Logo
- Expert Verified, Online, Free.
exam questions

Exam Certified Data Engineer Professional All Questions

View all questions & answers for the Certified Data Engineer Professional exam

Exam Certified Data Engineer Professional topic 1 question 90 discussion

Actual exam question from Databricks's Certified Data Engineer Professional
Question #: 90
Topic #: 1
[All Certified Data Engineer Professional Questions]

Which statement regarding Spark configuration on the Databricks platform is true?

  • A. The Databricks REST API can be used to modify the Spark configuration properties for an interactive cluster without interrupting jobs currently running on the cluster.
  • B. Spark configurations set within a notebook will affect all SparkSessions attached to the same interactive cluster.
  • C. Spark configuration properties can only be set for an interactive cluster by creating a global init script.
  • D. Spark configuration properties set for an interactive cluster with the Clusters UI will impact all notebooks attached to that cluster.
  • E. When the same Spark configuration property is set for an interactive cluster and a notebook attached to that cluster, the notebook setting will always be ignored.
Show Suggested Answer Hide Answer
Suggested Answer: D 🗳️

Comments

Chosen Answer:
This is a voting comment (?). It is better to Upvote an existing comment if you don't have anything to add.
Switch to a voting comment New
hamzaKhribi
Highly Voted 11 months, 4 weeks ago
Selected Answer: D
I tried it myself, setting a spark conf on the cluster ui, will impact all notebooks attached to that cluster, for example i set the number of shuffle partitions to 4, and in every notebook when i inspect the number of partitions i find 4.
upvoted 7 times
...
Curious76
Most Recent 9 months ago
Selected Answer: B
A. Incorrect: Modifying configurations through the Databricks REST API while jobs are running can lead to unexpected behavior or disruption. It's generally not recommended. C. Incorrect: While global init scripts can be used, it's not the only way. Configurations can also be set within notebooks. D. Incorrect: Configurations set through the Clusters UI apply to the entire cluster, but they might not necessarily override configurations set within notebooks attached to the cluster. E. Incorrect: Notebook configurations can take precedence over cluster-level configurations for the same property, offering finer-grained control at the notebook level.
upvoted 1 times
...
vctrhugo
9 months, 3 weeks ago
Selected Answer: D
These settings are applied at the cluster level and affect all SparkSessions on the cluster.
upvoted 3 times
...
alexvno
11 months, 1 week ago
Selected Answer: D
A wrong, cluster will restart -> D
upvoted 4 times
...
petrv
12 months ago
Selected Answer: A
In Databricks, you can use the Databricks REST API to modify Spark configuration properties for an interactive cluster without interrupting currently running jobs. This allows you to dynamically adjust Spark configurations to optimize performance or meet specific requirements without the need to restart the cluster.
upvoted 1 times
alexvno
11 months, 1 week ago
wrong, cluster will restart
upvoted 1 times
...
petrv
12 months ago
If you update the configuration of a cluster using the Databricks REST API or the Clusters UI while the cluster is in a RUNNING state, the cluster will be restarted to apply the new configuration. However, Databricks typically handles this situation in a way that minimizes disruption to running jobs.
upvoted 2 times
...
...
Community vote distribution
A (35%)
C (25%)
B (20%)
Other
Most Voted
A voting comment increases the vote count for the chosen answer by one.

Upvoting a comment with a selected answer will also increase the vote count towards that answer by one. So if you see a comment that you already agree with, you can upvote it instead of posting a new comment.

SaveCancel
Loading ...