Topic 1: Contoso, Ltd
You need to create the product dimension.
How should you complete the Apache Spark SQL code? To answer, select the appropriate
options in the answer area.
NOTE: Each correct selection is worth one point.
You need to populate the MAR1 data in the bronze layer.
Which two types of activities should you include in the pipeline? Each correct answer
presents part of the solution.
NOTE: Each correct selection is worth one point.
A. ForEach
B. Copy data
C. WebHook
D. Stored procedure
You need to ensure that the data analysts can access the gold layer lakehouse.
What should you do?
A. Add the DataAnalyst group to the Viewer role for WorkspaceA.
B. Share the lakehouse with the DataAnalysts group and grant the Build reports on the default semantic model permission.
C. Share the lakehouse with the DataAnalysts group and grant the Read all SQL Endpoint data permission.
D. Share the lakehouse with the DataAnalysts group and grant the Read all Apache Spark permission.
Explanation:
Data Analysts' Access Requirements must only have read access to the Delta tables in the
gold layer and not have access to the bronze and silver layers.
The gold layer data is typically queried via SQL Endpoints. Granting the Read all SQL
Endpoint data permission allows data analysts to query the data using familiar SQL-based
tools while restricting access to the underlying files.
You need to ensure that usage of the data in the Amazon S3 bucket meets the technical
requirements.
What should you do?
A. Create a workspace identity and enable high concurrency for the notebooks.
B. Create a shortcut and ensure that caching is disabled for the workspace.
C. Create a workspace identity and use the identity in a data pipeline.
D. Create a shortcut and ensure that caching is enabled for the workspace.
Explanation:
To ensure that the usage of the data in the Amazon S3 bucket meets the technical
requirements, we must address two key points:
Minimize egress costs associated with cross-cloud data access: Using a shortcut ensures
that Fabric does not replicate the data from the S3 bucket into the lakehouse but rather
provides direct access to the data in its original location. This minimizes cross-cloud data
transfer and avoids additional egress costs.
Prevent saving a copy of the raw data in the lakehouses: Disabling caching ensures that
the raw data is not copied or persisted in the Fabric workspace. The data is accessed ondemand
directly from the Amazon S3 bucket.
You need to recommend a method to populate the POS1 data to the lakehouse medallion
layers.
What should you recommend for each layer? To answer, select the appropriate options in
the answer area.
NOTE: Each correct selection is worth one point.
You need to ensure that the authors can see only their respective sales data.
How should you complete the statement? To answer, drag the appropriate values the
correct targets. Each value may be used once, more than once, or not at all. You may need
to drag the split bar between panes or scroll to view content
NOTE: Each correct selection is worth one point.
You need to resolve the sales data issue. The solution must minimize the amount of data transferred.
What should you do?
A. Spilt the dataflow into two dataflows.
B. Configure scheduled refresh for the dataflow.
C. Configure incremental refresh for the dataflow. Set Store rows from the past to 1 Month.
D. Configure incremental refresh for the dataflow. Set Refresh rows from the past to 1 Year.
E. Configure incremental refresh for the dataflow. Set Refresh rows from the past to 1 Month.
Explanation:
The sales data issue can be resolved by configuring incremental refresh for the dataflow.
Incremental refresh allows for only the new or changed data to be processed, minimizing
the amount of data transferred and improving performance.
The solution specifies that data older than one month never changes, so setting the refresh
period to 1 Month is appropriate. This ensures that only the most recent month of data will
be refreshed, reducing unnecessary data transfers.
You need to implement the solution for the book reviews.
Which should you do?
A. Create a Dataflow Gen2 dataflow.
B. Create a shortcut.
C. Enable external data sharing.
D. Create a data pipeline.
Explanation: The requirement specifies that Litware plans to make the book reviews available in the lakehouse without making a copy of the data. In this case, creating a shortcut in Fabric is the most appropriate solution. A shortcut is a reference to the external data, and it allows Litware to access the book reviews stored in Amazon S3 without duplicating the data into the lakehouse.
HOTSPOT
You need to troubleshoot the ad-hoc query issue.
How should you complete the statement? To answer, select the appropriate options in the
answer area.
NOTE: Each correct selection is worth one point.
What should you do to optimize the query experience for the business users?
A. Enable V-Order.
B. Create and update statistics.
C. Run the VACUUM command.
D. Introduce primary keys.
You have an Azure event hub. Each event contains the following fields:
BikepointID
Street
Neighbourhood
Latitude
Longitude
No_Bikes
No_Empty_Docks
You need to ingest the events. The solution must only retain events that have a
Neighbourhood value of Chelsea, and then store the retained events in a Fabric lakehouse.
What should you use?
A. a KQL queryset
B. an eventstream
C. a streaming dataset
D. Apache Spark Structured Streaming
Explanation:
An eventstream is the best solution for ingesting data from Azure Event Hub into Fabric,
while applying filtering logic such as retaining only the events that have a Neighbourhood
value of "Chelsea." Eventstreams in Microsoft Fabric are designed for handling real-time
data streams and can apply transformation logic directly on incoming events. In this case,
the eventstream can filter events based on the Neighbourhood field before storing the
retained events in a Fabric lakehouse.
Eventstreams are well-suited for stream processing, such as this case where you need to
filter out only specific data (events with a Neighbourhood of "Chelsea") before storing it in
the lakehouse.
You are building a data orchestration pattern by using a Fabric data pipeline named
Dynamic Data Copy as shown in the exhibit. (Click the Exhibit tab.)
Dynamic Data Copy does NOT use parametrization.
You need to configure the ForEach activity to receive the list of tables to be copied.
How should you complete the pipeline expression? To answer, select the appropriate
options in the answer area.
NOTE: Each correct selection is worth one point.
Page 1 out of 8 Pages |