Free Practice Test

Free Certified Associate Developer Practice Exam – 2025 Updated

Get Ready Smarter for the Certified Associate Developer Exam with Our Free and Trusted Certified Associate Developer Exam Questions – 2025 Updated.

At Cert Empire, we are dedicated to providing the latest and most accurate exam questions for students preparing for the Databricks Certified Associate Developer Exam. To support better preparation, we’ve made parts of our Certified Associate Developer exam resources free for everyone. You can practice as much as you want with Free Certified Associate Developer Practice Test.

Databricks Certified Associate Developer for Apache Spark Free Exam Questions

Disclaimer

Please keep a note that the demo questions are not frequently updated. You may as well find them in open communities around the web. However, this demo is only to depict what sort of questions you may find in our original files.

Nonetheless, the premium exam dumps files are frequently updated and are based on the latest exam syllabus and real exam questions.

1 / 60

Which of the following code blocks creates a new 6-column DataFrame by appending the rows of the 6-column DataFrame yesterdayTransactionsDf to the rows of the 6-column DataFrame todayTransactionsDf, ignoring that both DataFrames have different column names?

2 / 60

Which of the following code blocks concatenates rows of DataFrames transactionsDf and transactionsNewDf, omitting any duplicates?

3 / 60

The code block shown below should return an exact copy of DataFrame transactionsDf that does not include rows in which values in column storeId have the value 25. Choose the answer that correctly fills the blanks in the code block to accomplish this.

4 / 60

Which of the following statements about stages is correct?

5 / 60

The code block displayed below contains an error. The code block should write DataFrame transactionsDf as a parquet file to location filePath after partitioning it on column storeId. Find the error.
Code block:
transactionsDf.write.partitionOn("storeId").parquet(filePath)

6 / 60

Which of the following describes properties of a shuffle?

7 / 60

Which of the following code blocks returns all unique values across all values in columns value and productId in DataFrame transactionsDf in a one-column DataFrame?

8 / 60

Which of the following code blocks stores DataFrame itemsDf in executor memory and, if insufficient memory is available, serializes it and saves it to disk?

9 / 60

Which of the following code blocks generally causes a great amount of network traffic?

10 / 60

Which of the following describes a narrow transformation?

11 / 60

Which of the following statements about reducing out-of-memory errors is incorrect?

12 / 60

The code block displayed below contains an error. The code block should produce a DataFrame with color as the only column and three rows with color values of red, blue, and green,
respectively. Find the error.
Code block:
1. spark.createDataFrame([("red",), ("blue",), ("green",)], "color")
Instead of calling spark.createDataFrame, just DataFrame should be called.

13 / 60

Which of the following statements about the differences between actions and transformations is correct?

14 / 60

Which of the following code blocks returns a DataFrame containing a column dayOfYear, an integer representation of the day of the year from column openDate from DataFrame storesDF?
Note that column openDate is of type integer and represents a date in the UNIX epoch format – the number of seconds since midnight on January 1st, 1970.
A sample of storesDF is displayed below:

databricks certified associate developer for apache spark exam demo question

15 / 60

Which of the following Spark properties is used to configure whether DataFrame partitions that do not meet a minimum size threshold are automatically coalesced into larger partitions during a shuffle?

16 / 60

The code block shown below contains an error. The code block is intended to return a new 12-partition DataFrame from the 8-partition DataFrame storesDF by inducing a shuffle. Identify the error.
Code block:
storesDF.coalesce(12)

17 / 60

Which of the following operations can be used to return a new DataFrame from DataFrame storesDF without inducing a shuffle?

18 / 60

The code block shown below contains an error. The code block is intended to create a Python UDF assessPerformanceUDF() using the integer-returning Python function assessPerformance() and apply it to column customerSatisfaction in DataFrame storesDF. Identify the error.
Code block:
assessPerformanceUDF – udf(assessPerformance)
storesDF.withColumn("result", assessPerformanceUDF(col("customerSatisfaction")))

19 / 60

The code block shown below contains an error. The code block is intended to print the schema of DataFrame storesDF. Identify the error.
Code block:
storesDF.printSchema

20 / 60

Which of the following code blocks returns a 15 percent sample of rows from DataFrame storesDF without replacement?

21 / 60

The code block shown below contains an error. The code block is intended to return a new DataFrame with the mean of column sqft from DataFrame storesDF in column sqftMean. Identify the error.
Code block:
storesDF.agg(mean("sqft").alias("sqftMean"))

22 / 60

Which of the following operations returns a GroupedData object?

23 / 60

The code block shown contains an error. The code block is intended to return a new DataFrame where column sqft from DataFrame storesDF has had its missing values replaced with the value 30,000. Identify the error.
A sample of DataFrame storesDF is displayed below:

databricks certified associate developer for apache spark exam demo question

 

 

 

 

 

Code block:
storesDF.na.fill(30000, col("sqft"))

24 / 60

Which of the following code blocks returns a new DataFrame with column storeDescription where the pattern "Description: " has been removed from the beginning of column storeDescription in DataFrame storesDF?
A sample of DataFrame storesDF is below:

databricks certified associate developer for apache spark exam demo question

25 / 60

Which of the following code blocks returns a DataFrame where column storeCategory from DataFrame storesDF is split at the underscore character into column storeValueCategory and column storeSizeCategory?
A sample of DataFrame storesDF is displayed below:

databricks certified associate developer for apache spark exam demo question

26 / 60

Which of the following code blocks returns a new DataFrame from DataFrame storesDF where column storeId is of the type string?

27 / 60

Which of the following operations can be used to create a DataFrame with a subset of columns from DataFrame storesDF that are specified by name?

28 / 60

Which of the following statements about Spark DataFrames is incorrect?

29 / 60

Which of the following object types cannot be contained within a column of a Spark DataFrame?

30 / 60

A Spark application has a 128 GB DataFrame A and a 1 GB DataFrame B. If a broadcast join were to be performed on these two DataFrames, which of the following describes which DataFrame should be broadcasted and why?

31 / 60

Which of the following cluster configurations is most likely to experience an out-of-memory error in response to data skew in a single partition?

databricks certified associate developer for apache spark exam demo question

 

 

 

 

 

 

 

 

Note: each configuration has roughly the same compute power using 100 GB of RAM and 200 cores.

32 / 60

Which of the following statements about Spark’s stability is incorrect?

33 / 60

Which of the following DataFrame operations is classified as an action?

34 / 60

Which of the following is the most complete description of lazy evaluation?

35 / 60

Which of the following operations is most likely to result in a shuffle?

36 / 60

Which of the following describes the relationship between nodes and executors?

37 / 60

Which of the following is the most granular level of the Spark execution hierarchy?

38 / 60

The code block shown below contains an error. The code block is intended to return a DataFrame containing a column openDateString, a string representation of Java’s SimpleDateFormat. Identify the error.
Note that column openDate is of type integer and represents a date in the UNIX epoch format – the number of seconds since midnight on January 1st, 1970.
An example of Java’s SimpleDateFormat is "Sunday, Dec 4, 2008 1:05 PM".
A sample of storesDF is displayed below:

databricks certified associate developer for apache spark exam demo question

 

 

 

 

 

Code block:
storesDF.withColumn("openDateString", from_unixtime(col("openDate"), "EEE, MMM d, yyyy h:mm a", TimestampType()))

39 / 60

The code block shown below contains an error. The code block is intended to cache DataFrame storesDF only in Spark’s memory and then return the number of rows in the cached DataFrame. Identify the error.
Code block:
storesDF.cache().count()

40 / 60

The code block shown below contains an error. The code block is intended to use SQL to return a new DataFrame containing column storeId and column managerName from a table created from DataFrame storesDF. Identify the error.
Code block:
storesDF.createOrReplaceTempView("stores")
storesDF.sql("SELECT storeId, managerName FROM stores")

41 / 60

Which of the following code blocks fails to return a DataFrame reverse sorted alphabetically based on column division?

42 / 60

Which of the following code blocks returns all the rows from DataFrame storesDF?

43 / 60

Which of the following code blocks applies the function assessPerformance() to each row of DataFrame storesDF?

44 / 60

Which of the following code blocks returns a collection of summary statistics for all columns in DataFrame storesDF?

45 / 60

Which of the following code blocks will most quickly return an approximation for the number of distinct values in column division in DataFrame storesDF?

46 / 60

Which of the following operations can be used to return the number of rows in a DataFrame?

47 / 60

Which of the following code blocks returns a new DataFrame where column productCategories only has one word per row, resulting in a DataFrame with many more rows than DataFrame storesDF?
A sample of storesDF is displayed below:

databricks certified associate developer for apache spark exam demo question

48 / 60

Which of the following code blocks returns a new DataFrame where column division from DataFrame storesDF has been replaced and renamed to column state and column managerName from DataFrame storesDF has been replaced and renamed to column managerFullName?

49 / 60

Which of the following operations fails to return a DataFrame with no duplicate rows?

50 / 60

Which of the following code blocks returns a DataFrame containing only the rows from DataFrame storesDF where the value in column sqft is less than or equal to 25,000 OR the value in column customerSatisfaction is greater than or equal to 30?

51 / 60

Which of the following code blocks returns a new DataFrame with a new column employeesPerSqft that is the quotient of column numberOfEmployees and column sqft, both of which are from DataFrame storesDF? Note that column employeesPerSqft is not in the original DataFrame storesDF.

52 / 60

Which of the following operations can be used to create a new DataFrame that has 12 partitions from an original DataFrame df that has 8 partitions?

53 / 60

The code block shown below contains an error. The code block is intended to return a DataFrame containing all columns from DataFrame storesDF except for column sqft and column customerSatisfaction. Identify the error.
Code block:
storesDF.drop(sqft, customerSatisfaction)

54 / 60

Which of the following describes the difference between cluster and client execution modes?

55 / 60

Of the following situations, in which will it be most advantageous to store DataFrame df at the MEMORY_AND_DISK storage level rather than the MEMORY_ONLY storage level?

56 / 60

The default value of spark.sql.shuffle.partitions is 200. Which of the following describes what that means?

57 / 60

Which of the following DataFrame operations is classified as a wide transformation?

58 / 60

Which of the following describes the Spark driver?

59 / 60

Which of the following will occur if there are more slots than there are tasks?

60 / 60

Which of the following statements about Spark jobs is incorrect?

Your score is

The average score is 7%

Shopping Cart
Scroll to Top

FLASH OFFER

Days
Hours
Minutes
Seconds

avail $6 DISCOUNT on YOUR PURCHASE