Winter Sale Special Limited Time 65% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: v4s65

DP-600 Exam Dumps - Implementing Analytics Solutions Using Microsoft Fabric

Question # 4

You need to create a DAX measure to calculate the average overall satisfaction score.

How should you complete the DAX code? To answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.

Full Access
Question # 5

You have a Fabric tenant that contains a lakehouse.

You are using a Fabric notebook to save a large DataFrame by using the following code.

For each of the following statements, select Yes if the statement is true. Otherwise, select No.

NOTE: Each correct selection is worth one point.

Full Access
Question # 6

You have a Fabric tenant that contains a semantic model. The model contains data about retail stores.

You need to write a DAX query that will be executed by using the XMLA endpoint. The query must return the total amount of sales from the same period last year.

How should you complete the DAX expression? To answer, select the appropriate options in the answer area. NOTE: Each correct selection is worth one point.

Full Access
Question # 7

You have a Fabric tenant that contains a warehouse.

You are designing a star schema model that will contain a customer dimension. The customer dimension table will be a Type 2 slowly changing dimension (SCD).

You need to recommend which columns to add to the table. The columns must NOT already exist in the source.

Which three types of columns should you recommend? Each correct answer presents part of the solution.

NOTE: Each correct answer is worth one point.

A.

an effective end date and time

B.

a foreign key

C.

a surrogate key

D.

a natural key

E.

an effective start date and time

Full Access
Question # 8

You have a Fabric tenant that contains customer churn data stored as Parquet files in OneLake. The data contains details about customer demographics and product usage.

You create a Fabric notebook to read the data into a Spark DataFrame. You then create column charts in the notebook that show the distribution of retained customers as compared to lost customers based on geography, the number of products purchased, age. and customer tenure.

Which type of analytics are you performing?

A.

prescriptive

B.

diagnostic

C.

descriptive

D.

predictive

Full Access
Question # 9

You have a Fabric tenant.

You are creating a Fabric Data Factory pipeline.

You have a stored procedure that returns the number of active customers and their average sales for the current month.

You need to add an activity that will execute the stored procedure in a warehouse. The returned values must be available to the downstream activities of the pipeline.

Which type of activity should you add?

A.

Script

B.

Copy data

C.

Switch

D.

Append variable

Full Access
Question # 10

You have a Fabric tenant that contains a new semantic model in OneLake.

You use a Fabric notebook to read the data into a Spark DataFrame.

You need to evaluate the data to calculate the min, max, mean, and standard deviation values for all the string and numeric columns.

Solution: You use the following PySpark expression:

df .sumary ()

Does this meet the goal?

A.

Yes

B.

No

Full Access
Question # 11

You have a Fabric tenant tha1 contains a takehouse named Lakehouse1. Lakehouse1 contains a Delta table named Customer.

When you query Customer, you discover that the query is slow to execute. You suspect that maintenance was NOT performed on the table.

You need to identify whether maintenance tasks were performed on Customer.

Solution: You run the following Spark SQL statement:

REFRESH TABLE customer

Does this meet the goal?

A.

Yes

B.

No

Full Access
Question # 12

You have a Fabric tenant tha1 contains a takehouse named Lakehouse1. Lakehouse1 contains a Delta table named Customer.

When you query Customer, you discover that the query is slow to execute. You suspect that maintenance was NOT performed on the table.

You need to identify whether maintenance tasks were performed on Customer.

Solution: You run the following Spark SQL statement:

EXPLAIN TABLE customer

Does this meet the goal?

A.

Yes

B.

No

Full Access
Question # 13

You to need assign permissions for the data store in the AnalyticsPOC workspace. The solution must meet the security requirements.

Which additional permissions should you assign when you share the data store? To answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.

Full Access
Question # 14

You need to design a semantic model for the customer satisfaction report.

Which data source authentication method and mode should you use? To answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.

Full Access
Question # 15

You need to resolve the issue with the pricing group classification.

How should you complete the T-SQL statement? To answer, select the appropriate options in the answer area.

NOTE: Each correct selection is worth one point.

Full Access
Question # 16

You need to ensure the data loading activities in the AnalyticsPOC workspace are executed in the appropriate sequence. The solution must meet the technical requirements.

What should you do?

A.

Create a pipeline that has dependencies between activities and schedule the pipeline.

B.

Create and schedule a Spark job definition.

C.

Create a dataflow that has multiple steps and schedule the dataflow.

D.

Create and schedule a Spark notebook.

Full Access
Question # 17

What should you recommend using to ingest the customer data into the data store in the AnatyticsPOC workspace?

A.

a stored procedure

B.

a pipeline that contains a KQL activity

C.

a Spark notebook

D.

a dataflow

Full Access
Question # 18

You need to implement the date dimension in the data store. The solution must meet the technical requirements.

What are two ways to achieve the goal? Each correct answer presents a complete solution.

NOTE: Each correct selection is worth one point.

A.

Populate the date dimension table by using a dataflow.

B.

Populate the date dimension table by using a Stored procedure activity in a pipeline.

C.

Populate the date dimension view by using T-SQL.

D.

Populate the date dimension table by using a Copy activity in a pipeline.

Full Access
Question # 19

Which type of data store should you recommend in the AnalyticsPOC workspace?

A.

a data lake

B.

a warehouse

C.

a lakehouse

D.

an external Hive metaStore

Full Access
Question # 20

You need to ensure that Contoso can use version control to meet the data analytics requirements and the general requirements. What should you do?

A.

Store all the semantic models and reports in Data Lake Gen2 storage.

B.

Modify the settings of the Research workspaces to use a GitHub repository.

C.

Store all the semantic models and reports in Microsoft OneDrive.

D.

Modify the settings of the Research division workspaces to use an Azure Repos repository.

Full Access