exam questions

Exam Certified Associate Developer for Apache Spark All Questions

View all questions & answers for the Certified Associate Developer for Apache Spark exam

Exam Certified Associate Developer for Apache Spark topic 1 question 115 discussion

Which of the following code blocks returns a DataFrame containing only the rows from DataFrame storesDF where the value in column sqft is less than or equal to 25,000 AND the value in column customerSatisfaction is greater than or equal to 30?

  • A. storesDF.filter(col("sqft") <= 25000 and col("customerSatisfaction") >= 30)
  • B. storesDF.filter(col("sqft") <= 25000 or col("customerSatisfaction") >= 30)
  • C. storesDF.filter(sqft) <= 25000 and customerSatisfaction >= 30)
  • D. storesDF.filter(col("sqft") <= 25000 & col("customerSatisfaction") >= 30)
  • E. storesDF.filter(sqft <= 25000) & customerSatisfaction >= 30)
Show Suggested Answer Hide Answer
Suggested Answer: D 🗳️

Comments

Chosen Answer:
This is a voting comment (?). It is better to Upvote an existing comment if you don't have anything to add.
Switch to a voting comment New
Souvik_79
1 week, 4 days ago
Selected Answer: D
& is used as "and" in pyspark.
upvoted 1 times
...
Jgo1986
5 months ago
The most similar is D, And and OR are not valid statements for filtering in pySpark
upvoted 1 times
...
gaco
5 months, 2 weeks ago
in pyspark, all wrong as the conditions inside the filter should be wrapped inside parentesis. should be: D. storesDF.filter((col("sqft") <= 25000) & (col("customerSatisfaction") >= 30))
upvoted 4 times
...
65bd33e
5 months, 3 weeks ago
Selected Answer: D
D is the answer
upvoted 1 times
...
deadbeef38
7 months, 2 weeks ago
Selected Answer: A
A is right
upvoted 1 times
Jgo1986
5 months ago
no its not, ...
upvoted 1 times
...
...
Sowwy1
10 months, 1 week ago
It's D: https://sparkbyexamples.com/spark/spark-and-or-not-operators/ PySpark Logical operations use the bitwise operators: & for and | for or ~ for not
upvoted 2 times
...
sionita
1 year, 2 months ago
The answer should be E. In case of multiple conditions spark requires () such as: df.filter( (cond1) & (cond2) )
upvoted 1 times
...
MSH_6
1 year, 6 months ago
Selected Answer: A
A is the right answer.
upvoted 1 times
newusername
1 year, 2 months ago
No, you do not use and but & in Pyspark D is correct
upvoted 2 times
...
...
Community vote distribution
A (35%)
C (25%)
B (20%)
Other
Most Voted
A voting comment increases the vote count for the chosen answer by one.

Upvoting a comment with a selected answer will also increase the vote count towards that answer by one. So if you see a comment that you already agree with, you can upvote it instead of posting a new comment.

SaveCancel
Loading ...
exam
Someone Bought Contributor Access for:
SY0-701
London, 1 minute ago