Weekend Special Limited Time 65% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: 65pass65

Good News !!! DP-203 Data Engineering on Microsoft Azure is now Stable and With Pass Result

DP-203 Practice Exam Questions and Answers

Data Engineering on Microsoft Azure

Last Update 1 hour ago
Total Questions : 333

Data Engineering on Microsoft Azure is stable now with all latest exam questions are added 1 hour ago. Incorporating DP-203 practice exam questions into your study plan is more than just a preparation strategy.

DP-203 exam questions often include scenarios and problem-solving exercises that mirror real-world challenges. Working through DP-203 dumps allows you to practice pacing yourself, ensuring that you can complete all Data Engineering on Microsoft Azure practice test within the allotted time frame.

DP-203 PDF

DP-203 PDF (Printable)
$46.55
$132.99

DP-203 Testing Engine

DP-203 PDF (Printable)
$50.75
$144.99

DP-203 PDF + Testing Engine

DP-203 PDF (Printable)
$63.7
$181.99
Question # 1

You have an Azure subscription that contains an Azure Data Lake Storage account. The storage account contains a data lake named DataLake1.

You plan to use an Azure data factory to ingest data from a folder in DataLake1, transform the data, and land the data in another folder.

You need to ensure that the data factory can read and write data from any folder in the DataLake1 file system. The solution must meet the following requirements:

  • Minimize the risk of unauthorized user access.
  • Use the principle of least privilege.
  • Minimize maintenance effort.

How should you configure access to the storage account for the data factory? To answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.

Question # 1

Options:

Discussion 0
Question # 2

You need to implement a Type 3 slowly changing dimension (SCD) for product category data in an Azure Synapse Analytics dedicated SQL pool.

You have a table that was created by using the following Transact-SQL statement.

Question # 2

Which two columns should you add to the table? Each correct answer presents part of the solution.

NOTE: Each correct selection is worth one point.

Options:

A.  

[EffectiveScarcDate] [datetime] NOT NULL,

B.  

[CurrentProduccCacegory] [nvarchar] (100) NOT NULL,

C.  

[EffectiveEndDace] [dacecime] NULL,

D.  

[ProductCategory] [nvarchar] (100) NOT NULL,

E.  

[OriginalProduccCacegory] [nvarchar] (100) NOT NULL,

Discussion 0
Question # 3

You have a partitioned table in an Azure Synapse Analytics dedicated SQL pool.

You need to design queries to maximize the benefits of partition elimination.

What should you include in the Transact-SQL queries?

Options:

A.  

JOIN

B.  

WHERE

C.  

DISTINCT

D.  

GROUP BY

Discussion 0
Question # 4

You are creating an Apache Spark job in Azure Databricks that will ingest JSON-formatted data.

You need to convert a nested JSON string into a DataFrame that will contain multiple rows.

Which Spark SQL function should you use?

Options:

A.  

explode

B.  

filter

C.  

coalesce

D.  

extract

Discussion 0
Question # 5

You configure monitoring for a Microsoft Azure SQL Data Warehouse implementation. The implementation uses PolyBase to load data from comma-separated value (CSV) files stored in Azure Data Lake Gen 2 using an external table.

Files with an invalid schema cause errors to occur.

You need to monitor for an invalid schema error.

For which error should you monitor?

Options:

A.  

EXTERNAL TABLE access failed due to internal error: 'Java exception raised on call to HdfsBridge_Connect: Error

[com.microsoft.polybase.client.KerberosSecureLogin] occurred while accessing

external files.'

B.  

EXTERNAL TABLE access failed due to internal error: 'Java exception raised on call to HdfsBridge_Connect: Error [No FileSystem for scheme: wasbs] occurred while accessing external file.'

C.  

Cannot execute the query "Remote Query" against OLE DB provider "SQLNCLI11": for linked server "(null)", Query aborted- the maximum reject threshold (o

rows) was reached while regarding from an external source: 1 rows rejected out of total 1 rows processed.

D.  

EXTERNAL TABLE access failed due to internal error: 'Java exception raised on call to HdfsBridge_Connect: Error [Unable to instantiate LoginClass] occurred

while accessing external files.'

Discussion 0
Question # 6

Note: This question is part of a series of questions that present the same scenario. Each question in the series contains a unique solution that might meet the stated goals. Some question sets might have more than one correct solution, while others might not have a correct solution.

After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not appear in the review screen.

You have an Azure Data Lake Storage account that contains a staging zone.

You need to design a daily process to ingest incremental data from the staging zone, transform the data by executing an R script, and then insert the transformed data into a data warehouse in Azure Synapse Analytics.

Solution: You schedule an Azure Databricks job that executes an R notebook, and then inserts the data into the data warehouse.

Does this meet the goal?

Options:

A.  

Yes

B.  

No

Discussion 0
Question # 7

You are designing an application that will store petabytes of medical imaging data

When the data is first created, the data will be accessed frequently during the first week. After one month, the data must be accessible within 30 seconds, but files will be accessed infrequently. After one year, the data will be accessed infrequently but must be accessible within five minutes.

You need to select a storage strategy for the data. The solution must minimize costs.

Which storage tier should you use for each time frame? To answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.

Question # 7

Options:

Discussion 0
Question # 8

You are building an Azure Data Factory solution to process data received from Azure Event Hubs, and then ingested into an Azure Data Lake Storage Gen2 container.

The data will be ingested every five minutes from devices into JSON files. The files have the following naming pattern.

/{deviceType}/in/{YYYY}/{MM}/{DD}/{HH}/{deviceID}_{YYYY}{MM}{DD}HH}{mm}.json

You need to prepare the data for batch data processing so that there is one dataset per hour per deviceType. The solution must minimize read times.

How should you configure the sink for the copy activity? To answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.

Question # 8

Options:

Discussion 0
Question # 9

You have an Azure Synapse Analytics job that uses Scala.

You need to view the status of the job.

What should you do?

Options:

A.  

From Azure Monitor, run a Kusto query against the AzureDiagnostics table.

B.  

From Azure Monitor, run a Kusto query against the SparkLogying1 Event.CL table.

C.  

From Synapse Studio, select the workspace. From Monitor, select Apache Sparks applications.

D.  

From Synapse Studio, select the workspace. From Monitor, select SQL requests.

Discussion 0
Question # 10

You have an Azure data factory.

You need to examine the pipeline failures from the last 180 flays.

What should you use?

Options:

A.  

the Activity tog blade for the Data Factory resource

B.  

Azure Data Factory activity runs in Azure Monitor

C.  

Pipeline runs in the Azure Data Factory user experience

D.  

the Resource health blade for the Data Factory resource

Discussion 0
Get DP-203 dumps and pass your exam in 24 hours!

Free Exams Sample Questions