Free DP-203 Exam Braindumps

Pass your Data Engineering on Microsoft Azure exam with these free Questions and Answers

Page 10 of 61
QUESTION 41

- (Exam Topic 1)
You need to design a data storage structure for the product sales transactions. The solution must meet the sales transaction dataset requirements.
What should you include in the solution? To answer, select the appropriate options in the answer area. NOTE: Each correct selection is worth one point.
DP-203 dumps exhibit
Solution:
Graphical user interface, text, application, chat or text message Description automatically generated
Box 1: Hash Scenario:
Ensure that queries joining and filtering sales transaction records based on product ID complete as quickly as possible.
A hash distributed table can deliver the highest query performance for joins and aggregations on large tables. Box 2: Set the distribution column to the sales date.
Scenario: Partition data that contains sales transaction records. Partitions must be designed to provide efficient loads by month. Boundary values must belong to the partition on the right.
Reference:
https://rajanieshkaushikk.com/2020/09/09/how-to-choose-right-data-distribution-strategy-for-azure-synapse/

Does this meet the goal?

  1. A. Yes
  2. B. No

Correct Answer: A

QUESTION 42

- (Exam Topic 3)
You have an Azure Synapse Analytics dedicated SQL pool named Pool1 and a database named DB1. DB1 contains a fact table named Table1.
You need to identify the extent of the data skew in Table1. What should you do in Synapse Studio?

  1. A. Connect to the built-in pool and run dbcc pdw_showspaceused.
  2. B. Connect to the built-in pool and run dbcc checkalloc.
  3. C. Connect to Pool1 and query sys.dm_pdw_node_scacus.
  4. D. Connect to Pool1 and query sys.dm_pdw_nodes_db_partition_scacs.

Correct Answer: A
A quick way to check for data skew is to use DBCC PDW_SHOWSPACEUSED. The following SQL code returns the number of table rows that are stored in each of the 60 distributions. For balanced performance, the rows in your distributed table should be spread evenly across all the distributions.
DBCC PDW_SHOWSPACEUSED('dbo.FactInternetSales'); Reference:
https://docs.microsoft.com/en-us/azure/synapse-analytics/sql-data-warehouse/sql-data-warehouse-tables-distribu

QUESTION 43

- (Exam Topic 3)
You have two fact tables named Flight and Weather. Queries targeting the tables will be based on the join between the following columns.
DP-203 dumps exhibit
You need to recommend a solution that maximizes query performance. What should you include in the recommendation?

  1. A. In the tables use a hash distribution of ArrivalDateTime and ReportDateTime.
  2. B. In the tables use a hash distribution of ArrivaIAirportID and AirportlD.
  3. C. In each table, create an identity column.
  4. D. In each table, create a column as a composite of the other two columns in the table.

Correct Answer: B
Hash-distribution improves query performance on large fact tables.

QUESTION 44

- (Exam Topic 3)
A company plans to use Platform-as-a-Service (PaaS) to create the new data pipeline process. The process must meet the following requirements:
Ingest:
DP-203 dumps exhibitAccess multiple data sources.
DP-203 dumps exhibit Provide the ability to orchestrate workflow.
DP-203 dumps exhibit Provide the capability to run SQL Server Integration Services packages. Store:
DP-203 dumps exhibit Optimize storage for big data workloads.
DP-203 dumps exhibit Provide encryption of data at rest.
DP-203 dumps exhibit Operate with no size limits. Prepare and Train:
DP-203 dumps exhibit Provide a fully-managed and interactive workspace for exploration and visualization.
DP-203 dumps exhibit Provide the ability to program in R, SQL, Python, Scala, and Java.
DP-203 dumps exhibit Provide seamless user authentication with Azure Active Directory. Model & Serve:
DP-203 dumps exhibit Implement native columnar storage.
DP-203 dumps exhibit Support for the SQL language
DP-203 dumps exhibit Provide support for structured streaming. You need to build the data integration pipeline.
Which technologies should you use? To answer, select the appropriate options in the answer area.
NOTE: Each correct selection is worth one point.
DP-203 dumps exhibit
Solution:
Graphical user interface, application, table, email Description automatically generated

Does this meet the goal?

  1. A. Yes
  2. B. No

Correct Answer: A

QUESTION 45

- (Exam Topic 3)
You need to create an Azure Data Factory pipeline to process data for the following three departments at your company: Ecommerce, retail, and wholesale. The solution must ensure that data can also be processed for the entire company.
How should you complete the Data Factory data flowscript? To answer, drag the appropriate values to the correct targets. Each value may be used once, more than once, or not at all. You may need to drag the split bar between panes or scroll to view content.
NOTE: Each correct selection is worth one point.
DP-203 dumps exhibit
Solution:
The conditional split transformation routes data rows to different streams based on matching conditions. The conditional split transformation is similar to a CASE decision structure in a programming language. The transformation evaluates expressions, and based on the results, directs the data row to the specified stream.
Box 1: dept=='ecommerce', dept=='retail', dept=='wholesale'
First we put the condition. The order must match the stream labeling we define in Box 3. Syntax:
split(

disjoint: {true | false}
) ~> @(stream1, stream2, ..., )
Box 2: discount : false
disjoint is false because the data goes to the first matching condition. All remaining rows matching the third condition go to output stream all.
Box 3: ecommerce, retail, wholesale, all Label the streams
Reference:
https://docs.microsoft.com/en-us/azure/data-factory/data-flow-conditional-split

Does this meet the goal?

  1. A. Yes
  2. B. No

Correct Answer: A

Page 10 of 61

Post your Comments and Discuss Microsoft DP-203 exam with other Community members: