exam questions

Exam DP-203 All Questions

View all questions & answers for the DP-203 exam

Exam DP-203 topic 2 question 21 discussion

Actual exam question from Microsoft's DP-203
Question #: 21
Topic #: 2
[All DP-203 Questions]

DRAG DROP -
You have an Azure Data Lake Storage Gen2 account that contains a JSON file for customers. The file contains two attributes named FirstName and LastName.
You need to copy the data from the JSON file to an Azure Synapse Analytics table by using Azure Databricks. A new column must be created that concatenates the FirstName and LastName values.
You create the following components:
✑ A destination table in Azure Synapse
✑ An Azure Blob storage container
✑ A service principal
Which five actions should you perform in sequence next in is Databricks notebook? To answer, move the appropriate actions from the list of actions to the answer area and arrange them in the correct order.
Select and Place:

Show Suggested Answer Hide Answer
Suggested Answer:
Step 1: Mount the Data Lake Storage onto DBFS
Begin with creating a file system in the Azure Data Lake Storage Gen2 account.
Step 2: Read the file into a data frame.
You can load the json files as a data frame in Azure Databricks.
Step 3: Perform transformations on the data frame.
Step 4: Specify a temporary folder to stage the data
Specify a temporary folder to use while moving data between Azure Databricks and Azure Synapse.
Step 5: Write the results to a table in Azure Synapse.
You upload the transformed data frame into Azure Synapse. You use the Azure Synapse connector for Azure Databricks to directly upload a dataframe as a table in a Azure Synapse.
Reference:
https://docs.microsoft.com/en-us/azure/azure-databricks/databricks-extract-load-sql-data-warehouse

Comments

Chosen Answer:
This is a voting comment (?). It is better to Upvote an existing comment if you don't have anything to add.
Switch to a voting comment New
sagga
Highly Voted 3 years, 5 months ago
I think the correct order is: 1) mount onto DBFS 2) read into data frame 3) transform data frame 4) specify temporary folder 5) write to table in SQL data warehouse About temporary folder, there is a note explain this: https://docs.microsoft.com/en-us/azure/databricks/scenarios/databricks-extract-load-sql-data-warehouse#load-data-into-azure-synapse Discussions about this question: https://www.examtopics.com/discussions/microsoft/view/11653-exam-dp-200-topic-2-question-30-discussion/
upvoted 172 times
satyamkishoresingh
3 years, 1 month ago
This order is absolutely correct.
upvoted 3 times
...
snna4
2 years, 9 months ago
OMG... the 5th step should be "Write the results to a table in Azure synapse". Who are those people "liked" this answer? Guys, just read the task.
upvoted 11 times
dev2dev
2 years, 9 months ago
hehe, those who understand sql dw = azure synapse :D
upvoted 36 times
kkk5566
1 year, 1 month ago
it is the incorrect answer
upvoted 1 times
...
...
...
GameLift
3 years, 1 month ago
I agree, although, why do we need a temporary folder? We already have storage blob as temporary storage?
upvoted 2 times
gaganmay26
2 years, 4 months ago
Databricks uses polybase to write to Synapse and thus we need to stage the file is required.
upvoted 2 times
...
...
andylop04
3 years, 3 months ago
Today I received this question in my exam. Only appeared the 5 options of this response. I only had to order, not choice. This solutions is the correct. Thanks sagga.
upvoted 41 times
KingIlo
2 years, 11 months ago
Correct also received the only five options. Also see: https://www.examtopics.com/discussions/microsoft/view/11653-exam-dp-200-topic-2-question-30-discussion/
upvoted 4 times
...
...
...
Miris
Highly Voted 3 years, 4 months ago
1) mount the data onto DBFS 2) Read the file into a data frame 3) Perform transformations on the file 4) Specify a temporary folder to stage the data 5) Write the results to a table in Azure synapse
upvoted 22 times
Tickxit
1 year, 11 months ago
transformations on dataframe, not on the file.
upvoted 7 times
...
...
Momoanwar
Most Recent 10 months, 1 week ago
Answer are correct, chatgpt say : To accomplish the task in an Azure Databricks notebook, the logical sequence of actions would be: 1. **Mount the Data Lake Storage onto DBFS**: This allows access to the JSON file stored in Azure Data Lake Storage using the Databricks File System. 2. **Read the file into a data frame**: Use Spark to read the JSON file into a DataFrame for processing. 3. **Perform transformations on the data frame**: Apply transformations to concatenate the FirstName and LastName fields to create a new column. 4. **Specify a temporary folder to stage the data**: Before writing the data to Azure Synapse, it is a common practice to stage it in a temporary folder. 5. **Write the results to a table in Azure Synapse**: Finally, write the transformed DataFrame to the destination table in Azure Synapse Analytics. These steps would ensure the JSON file data is properly transformed and loaded into Azure Synapse Analytics for further use.
upvoted 1 times
...
EliteAllen
11 months ago
Just remember the initials first: M.R.P.S.W then go to the details.
upvoted 1 times
...
bakamon
1 year, 4 months ago
1. Mount the data lake storage onto DBFS. 2. Read the file into a data frame. 3. Perform transformations on the data frame. 4. Specify a temporary folder to stage the data. 5. Write the results to a table in Azure Synapse. This will allow you to read the data from the JSON file into a data frame, perform the necessary transformations to concatenate the FirstName and LastName values, and then write the results to a table in Azure Synapse.
upvoted 3 times
...
Deeksha1234
2 years, 2 months ago
answer is correct, explained by the reference link in the given solution
upvoted 1 times
...
carloalbe
2 years, 5 months ago
I don not see the reason why "specify temporary folder" can not be both before or after the "read and transformation phase"
upvoted 3 times
Davico93
2 years, 4 months ago
I want to know the reason too!
upvoted 1 times
...
...
Egocentric
2 years, 6 months ago
given answer is correct, after reading and rereading stand with the given answer
upvoted 1 times
...
Sandip4u
2 years, 9 months ago
I think the correct order is: 1) mount onto DBFS 2) read into data frame 3) transform data frame 4) specify temporary folder 5) write to table in SQL data warehouse
upvoted 3 times
...
Canary_2021
2 years, 10 months ago
Here is my answer. 1) Create a service principal - Not sure why this step is not a choice in this question. I don't thing need to mount onto DBFS, but you do need to assign permission to allow databricks talk with Data Lake and read file. 2) Read the file into data frame 3) Perform transformations on the data frame Data have been read into data from, so should transform data from data frame, not data file. 4) Specify temporary folder to stage the data 5) Write the results to a table in Azure Synapse I reviewed this online document. No any place mentioned that the data frame needs to be dropped. https://docs.microsoft.com/en-us/azure/databricks/scenarios/databricks-extract-load-sql-data-warehouse
upvoted 2 times
Gina8008
2 years, 7 months ago
you do not need to create a service principal, this is already exist
upvoted 1 times
...
...
Sayour
2 years, 10 months ago
There Is A Contradiction Between Answers On The Drag & Drop And The Answers In The Steps Listing, And I Think The Correct Ones Are That In The Listing And Not The Drag & Drop.
upvoted 1 times
...
VJPR
2 years, 10 months ago
1) Mount the data onto DBFS 2) Read the file into a data frame 3) Perform transformations 4) Specify a temporary folder to stage the data 5) Write the results to a table in Azure synapse
upvoted 1 times
...
[Removed]
3 years, 1 month ago
The given answer is correct, after read the link provided carefully several times. There's already a service principal. With that, it's no need to mount. You do need to drop the dataframe as the last step.
upvoted 1 times
GameLift
3 years, 1 month ago
Service Principal has nothing to do with DataBricks.
upvoted 4 times
nefarious_smalls
2 years, 5 months ago
Actually you can assign a service principal to any data bricks account and use OAuth to connect with its tenant id app secret, and app id. You can then mount the data lake to databricks.
upvoted 1 times
...
...
...
labasmuse
3 years, 5 months ago
Correct solution: Read the file into a data frame Perform transformations on the file Specify a temporary folder to stage the data Write the results to a table in Azure synapse Drop the data frame
upvoted 5 times
hello2tomoki
2 years, 6 months ago
Step 1: Read the file into a data frame. You can load the json files as a data frame in Azure Databricks. Step 2: Perform transformations on the data frame. Step 3:Specify a temporary folder to stage the data Specify a temporary folder to use while moving data between Azure Databricks and Azure SQL Data Warehouse. Step 4: Write the results to a table in Azure synapse Step 5: Drop the data frame - Clean up resources. https://www.examtopics.com/discussions/microsoft/view/11653-exam-dp-200-topic-2-question-30-discussion/
upvoted 3 times
kkk5566
1 year, 1 month ago
Correct
upvoted 1 times
...
...
Wisenut
3 years, 5 months ago
I believe you perform transformation on the data frame and not on the file
upvoted 6 times
...
ThiruthuvaRajan
3 years, 4 months ago
you should not perform transformation on the file. You need not to drop the dataframe. sagga options are correct
upvoted 3 times
...
...
Community vote distribution
A (35%)
C (25%)
B (20%)
Other
Most Voted
A voting comment increases the vote count for the chosen answer by one.

Upvoting a comment with a selected answer will also increase the vote count towards that answer by one. So if you see a comment that you already agree with, you can upvote it instead of posting a new comment.

SaveCancel
Loading ...
exam
Someone Bought Contributor Access for:
SY0-701
London, 1 minute ago