Black Friday Special Sale - 70% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: best70

Page: 1 / 4
Total 40 questions
Exam Code: Databricks-Certified-Associate-Developer-for-Apache-Spark-3.5                Update: Nov 30, 2025
Exam Name: Databricks Certified Associate Developer for Apache Spark 3.5 – Python

Databricks Databricks Certified Associate Developer for Apache Spark 3.5 – Python Databricks-Certified-Associate-Developer-for-Apache-Spark-3.5 Exam Dumps: Updated Questions & Answers (November 2025)

Question # 1

Which Spark configuration controls the number of tasks that can run in parallel on the executor?

Options:

A.

spark.executor.cores

B.

spark.task.maxFailures

C.

spark.driver.cores

D.

spark.executor.memory

Question # 2

40 of 55.

A developer wants to refactor older Spark code to take advantage of built-in functions introduced in Spark 3.5.

The original code:

from pyspark.sql import functions as F

min_price = 110.50

result_df = prices_df.filter(F.col("price") > min_price).agg(F.count("*"))

Which code block should the developer use to refactor the code?

A.

result_df = prices_df.filter(F.col("price") > F.lit(min_price)).agg(F.count("*"))

B.

result_df = prices_df.where(F.lit("price") > min_price).groupBy().count()

C.

result_df = prices_df.withColumn("valid_price", when(col("price") > F.lit(min_price), True))

D.

result_df = prices_df.filter(F.lit(min_price) > F.col("price")).count()

Question # 3

49 of 55.

In the code block below, aggDF contains aggregations on a streaming DataFrame:

aggDF.writeStream \

.format("console") \

.outputMode("???") \

.start()

Which output mode at line 3 ensures that the entire result table is written to the console during each trigger execution?

A.

AGGREGATE

B.

COMPLETE

C.

REPLACE

D.

APPEND

Question # 4

54 of 55.

What is the benefit of Adaptive Query Execution (AQE)?

A.

It allows Spark to optimize the query plan before execution but does not adapt during runtime.

B.

It automatically distributes tasks across nodes in the clusters and does not perform runtime adjustments to the query plan.

C.

It optimizes query execution by parallelizing tasks and does not adjust strategies based on runtime metrics like data skew.

D.

It enables the adjustment of the query plan during runtime, handling skewed data, optimizing join strategies, and improving overall query performance.

Question # 5

55 of 55.

An application architect has been investigating Spark Connect as a way to modernize existing Spark applications running in their organization.

Which requirement blocks the adoption of Spark Connect in this organization?

A.

Debuggability: the ability to perform interactive debugging directly from the application code

B.

Upgradability: the ability to upgrade the Spark applications independently from the Spark driver itself

C.

Complete Spark API support: the ability to migrate all existing code to Spark Connect without modification, including the RDD APIs

D.

Stability: isolation of application code and dependencies from each other and the Spark driver

Question # 6

What is a feature of Spark Connect?

A.

It supports DataStreamReader, DataStreamWriter, StreamingQuery, and Streaming APIs

B.

Supports DataFrame, Functions, Column, SparkContext PySpark APIs

C.

It supports only PySpark applications

D.

It has built-in authentication

Question # 7

A data engineer is running a Spark job to process a dataset of 1 TB stored in distributed storage. The cluster has 10 nodes, each with 16 CPUs. Spark UI shows:

Low number of Active Tasks

Many tasks complete in milliseconds

Fewer tasks than available CPUs

Which approach should be used to adjust the partitioning for optimal resource allocation?

A.

Set the number of partitions equal to the total number of CPUs in the cluster

B.

Set the number of partitions to a fixed value, such as 200

C.

Set the number of partitions equal to the number of nodes in the cluster

D.

Set the number of partitions by dividing the dataset size (1 TB) by a reasonable partition size, such as 128 MB

Question # 8

45 of 55.

Which feature of Spark Connect should be considered when designing an application that plans to enable remote interaction with a Spark cluster?

A.

It is primarily used for data ingestion into Spark from external sources.

B.

It provides a way to run Spark applications remotely in any programming language.

C.

It can be used to interact with any remote cluster using the REST API.

D.

It allows for remote execution of Spark jobs.

Question # 9

5 of 55.

What is the relationship between jobs, stages, and tasks during execution in Apache Spark?

A.

A job contains multiple tasks, and each task contains multiple stages.

B.

A stage contains multiple jobs, and each job contains multiple tasks.

C.

A stage contains multiple tasks, and each task contains multiple jobs.

D.

A job contains multiple stages, and each stage contains multiple tasks.

Question # 10

17 of 55.

A data engineer has noticed that upgrading the Spark version in their applications from Spark 3.0 to Spark 3.5 has improved the runtime of some scheduled Spark applications.

Looking further, the data engineer realizes that Adaptive Query Execution (AQE) is now enabled.

Which operation should AQE be implementing to automatically improve the Spark application performance?

A.

Dynamically switching join strategies

B.

Collecting persistent table statistics and storing them in the metastore for future use

C.

Improving the performance of single-stage Spark jobs

D.

Optimizing the layout of Delta files on disk

Page: 1 / 4
Total 40 questions

Most Popular Certification Exams

Payment

       

Contact us

dumpscollection live chat

Site Secure

mcafee secure

TESTED 30 Nov 2025