microsoft dp-200 online test

Implementing an Azure Data Solution (beta)

What students need to know about the dp-200 exam

  • Total 372 Questions & Answers
  • Has case studies
  • Passing score: 700

Skills Measured:

  • mplement data storage solutions (40-45%)
  • Manage and develop data processing (25-30%)
  • Monitor and optimize data solutions (30-35%)

Question 1 Topic 1, Case Study 1Case Study Question View Case

DRAG DROP
You need to ensure that phone-based polling data can be analyzed in the PollingData database.
Which three actions should you perform in sequence? To answer, move the appropriate actions from the list of actions to the
answer are and arrange them in the correct order.
Select and Place:

Answer:

Explanation:
Scenario:
All deployments must be performed by using Azure DevOps. Deployments must use templates used in multiple
environments No credentials or secrets should be used during deployments

Discussions

Question 2 Topic 1, Case Study 1Case Study Question View Case

DRAG DROP
You need to provision the polling data storage account.
How should you configure the storage account? To answer, drag the appropriate Configuration Value to the correct Setting.
Each Configuration Value may be used once, more than once, or not at all. You may need to drag the split bar between
panes or scroll to view content.
NOTE: Each correct selection is worth one point.
Select and Place:

Answer:

Explanation:
Account type: StorageV2
You must create new storage accounts as type StorageV2 (general-purpose V2) to take advantage of Data Lake Storage
Gen2 features.
Scenario: Polling data is stored in one of the two locations:
An on-premises Microsoft SQL Server 2019 database named PollingData Azure Data Lake Gen 2


Data in Data Lake is queried by using PolyBase
Replication type: RA-GRS
Scenario: All services and processes must be resilient to a regional Azure outage.
Geo-redundant storage (GRS) is designed to provide at least 99.99999999999999% (16 9's) durability of objects over a
given year by replicating your data to a secondary region that is hundreds of miles away from the primary region. If your
storage account has GRS enabled, then your data is durable even in the case of a complete regional outage or a disaster in
which the primary region isn't recoverable.
If you opt for GRS, you have two related options to choose from:
GRS replicates your data to another data center in a secondary region, but that data is available to be read only if

Microsoft initiates a failover from the primary to secondary region.
Read-access geo-redundant storage (RA-GRS) is based on GRS. RA-GRS replicates your data to another data center in

a secondary region, and also provides you with the option to read from the secondary region. With RA-GRS, you can read
from the secondary region regardless of whether Microsoft initiates a failover from the primary to secondary region.
References:
https://docs.microsoft.com/bs-cyrl-ba/azure/storage/blobs/data-lake-storage-quickstart-create-account
https://docs.microsoft.com/en-us/azure/storage/common/storage-redundancy-grs
Implement data storage solutions

Discussions

Question 3 Topic 2, Case Study 2Case Study Question View Case

HOTSPOT
You need to build a solution to collect the telemetry data for Race Central.
What should you use? To answer, select the appropriate options in the answer area.
NOTE: Each correct selection is worth one point.
Hot Area:

Answer:

Explanation:
API: Table
Azure Cosmos DB provides native support for wire protocol-compatible APIs for popular databases. These include
MongoDB, Apache Cassandra, Gremlin, and Azure Table storage.
Scenario: The telemetry data must migrate toward a solution that is native to Azure.
Consistency level: Strong
Use the strongest consistency Strong to minimize convergence time.
Scenario: The data must be written to the Azure datacenter closest to each race and must converge in the least amount of
time.
Reference: https://docs.microsoft.com/en-us/azure/cosmos-db/consistency-levels

Discussions

Question 4 Topic 2, Case Study 2Case Study Question View Case

On which data store should you configure TDE to meet the technical requirements?

  • A. Cosmos DB
  • B. Azure Synapse Analytics
  • C. Azure SQL Database
Answer:

B

Explanation:
Scenario: Transparent data encryption (TDE) must be enabled on all data stores, whenever possible. The database for
Mechanical Workflow must be moved to Azure Synapse Analytics.
Incorrect Answers:
A: Cosmos DB does not support TDE.

Discussions

Question 5 Topic 2, Case Study 2Case Study Question View Case

HOTSPOT
You are building the data store solution for Mechanical Workflow.
How should you configure Table1? To answer, select the appropriate options in the answer area.
NOTE: Each correct selection is worth one point.
Hot Area:

Answer:

Explanation:
Table Type: Hash distributed.
Hash-distributed tables improve query performance on large fact tables.
Index type: Clusted columnstore
Scenario:
Mechanical Workflow has a named Table1 that is 1 TB. Large aggregations are performed on a single column of Table 1.
References: https://docs.microsoft.com/en-us/azure/sql-data-warehouse/sql-data-warehouse-tables-distribute

Discussions

Question 6 Topic 2, Case Study 2Case Study Question View Case

HOTSPOT
Which masking functions should you implement for each column to meet the data masking requirements? To answer, select
the appropriate options in the answer area.
NOTE: Each correct selection is worth one point.
Hot Area:

Answer:

Explanation:
Box 1: Credit Card
The Credit Card Masking method exposes the last four digits of the designated fields and adds a constant string as a prefix
in the form of a credit card.
Example: XXXX-XXXX-XXXX-1234
Only show the last four digits of the values in a column named SuspensionSprings.

Box 2: Default
Default uses a zero value for numeric data types (bigint, bit, decimal, int, money, numeric, smallint, smallmoney, tinyint, float,
real).
Only show a zero value for the values in a column named ShockOilWeight.

Scenario:
The company identifies the following data masking requirements for the Race Central data that will be stored in SQL
Database:
Only show a zero value for the values in a column named ShockOilWeight.

Only show the last four digits of the values in a column named SuspensionSprings.

Reference: https://docs.microsoft.com/en-us/azure/azure-sql/database/dynamic-data-masking-overview

Discussions

Question 7 Topic 2, Case Study 2Case Study Question View Case

HOTSPOT
Which masking functions should you implement for each column to meet the data masking requirements? To answer, select
the appropriate options in the answer area.
NOTE: Each correct selection is worth one point.
Hot Area:

Answer:

Explanation:
Box 1: Custom text/string: A masking method, which exposes the first and/or last characters and adds a custom padding
string in the middle.
Only show the last four digits of the values in a column named SuspensionSprings.
Box 2: Default
Default uses a zero value for numeric data types (bigint, bit, decimal, int, money, numeric, smallint, smallmoney, tinyint, float,
real).
Scenario: Only show a zero value for the values in a column named ShockOilWeight.
Scenario:
The company identifies the following data masking requirements for the Race Central data that will be stored in SQL
Database:
Only show a zero value for the values in a column named ShockOilWeight.

Only show the last four digits of the values in a column named SuspensionSprings.

Reference: https://docs.microsoft.com/en-us/azure/azure-sql/database/dynamic-data-masking-overview
Implement data storage solutions

Discussions

Question 8 Topic 3, Case Study 3Case Study Question View Case

You need to configure a disaster recovery solution for SALESDB to meet the technical requirements.
What should you configure in the backup policy?

  • A. weekly long-term retention backups that are retained for three weeks
  • B. failover groups
  • C. a point-in-time restore
  • D. geo-replication
Answer:

C

Explanation:
Scenario: SALESDB must be restorable to any given minute within the past three weeks.
The Azure SQL Database service protects all databases with an automated backup system. These backups are retained for
7 days for Basic, 35 days for Standard and 35 days for Premium. Point-in-time restore is a self-service capability, allowing
customers to restore a Basic, Standard or Premium database from these backups to any point within the retention period.
References: https://azure.microsoft.com/en-us/blog/azure-sql-database-point-in-time-restore/

Discussions

Question 9 Topic 3, Case Study 3Case Study Question View Case

You need to implement event processing by using Stream Analytics to produce consistent JSON documents.
Which three actions should you perform? Each correct answer presents part of the solution.
NOTE: Each correct selection is worth one point.

  • A. Define an output to Cosmos DB.
  • B. Define a query that contains a JavaScript user-defined aggregates (UDA) function.
  • C. Define a reference input.
  • D. Define a transformation query.
  • E. Define an output to Azure Data Lake Storage Gen2.
  • F. Define a stream input.
Answer:

D E F

Explanation:
DOCDB stored documents that connect to the sales data in SALESDB. The documents are stored in two different JSON

formats based on the sales channel.
The sales data, including the documents in JSON format, must be gathered as it arrives and analyzed online by using

Azure Stream Analytics. The analytic process will perform aggregations that must be done continuously, without gaps, and
without overlapping.
As they arrive, all the sales documents in JSON format must be transformed into one consistent format.

Discussions

Question 10 Topic 4, Case Study 4Case Study Question View Case

You need to ensure that phone-based poling data can be analyzed in the PollingData database.
How should you configure Azure Data Factory?

  • A. Use a tumbling schedule trigger
  • B. Use an event-based trigger
  • C. Use a schedule trigger
  • D. Use manual execution
Answer:

C

Explanation:
When creating a schedule trigger, you specify a schedule (start date, recurrence, end date etc.) for the trigger, and associate
with a Data Factory pipeline.
Scenario:
All data migration processes must use Azure Data Factory
All data migrations must run automatically during non-business hours References:
https://docs.microsoft.com/en-us/azure/data-factory/how-to-create-schedule-trigger

Discussions
To page 2