MICROSOFT CERTIFICATION

DP-600 Fabric Analytics Engineer Associate Practice Exam

Exam Number: 3117 | Last updated 16-Apr-26 | 788+ questions across 4 vendor-aligned objectives

The DP-600 Fabric Analytics Engineer Associate certification validates the skills of analytics engineers who design, build, and manage enterprise-scale analytics solutions using Microsoft Fabric. This exam measures your ability to work with Microsoft Fabric, Power BI, OneLake, Lakehouse, Data Warehouse, Dataflow Gen2, Notebooks, demonstrating both conceptual understanding and practical implementation skills required in today’s enterprise environments.

The heaviest exam domains include Prepare and Serve Data (40–45%), Implement and Manage Semantic Models (20–25%), and Explore and Analyze Data (20–25%). These areas collectively represent the majority of exam content and require focused preparation across their respective subtopics.

Additional domains tested include Plan, Implement, and Manage a Solution for Data Analytics (10–15%). Together, these areas round out the full exam blueprint and ensure candidates possess well-rounded expertise across the certification scope.

 Prepare and Serve Data carries the heaviest weight at 40–45%. Focus on Lakehouse architecture, Delta table optimization, and Dataflow Gen2 transformations. Know Direct Lake mode and when to use Import vs. DirectQuery.

Every answer links to the source. Each explanation below includes a hyperlink to the exact Microsoft documentation page the question was derived from. PowerKram is the only practice platform with source-verified explanations. Learn about our methodology →

372

practice exam users

93.9%

satisfied users

90%

passed the exam

4.1/5

quality rating

Test your DP-600 Fabric Analytics Engineer Associate knowledge

10 of 788+ questions

Question #1 - Prepare and Serve Data

A retail analytics team ingests daily point-of-sale data from an on-premises SQL Server into a Fabric Lakehouse for transformation and reporting.

Which Fabric component should the team use for this recurring ingestion?

A) Azure Data Factory linked service only
B) Dataflow Gen2 with an on-premises data gateway
C) Power BI Desktop import mode
D) Manual CSV upload to OneLake

 

Correct answers: B – Explanation:
Dataflow Gen2 with a gateway connects to on-premises sources, applies Power Query transformations, and loads directly into the Lakehouse on a schedule. Manual upload is not automated. Power BI Desktop imports into a model, not a Lakehouse. ADF alone requires additional Fabric integration steps. Source: Check Source

A retail analytics team ingests daily point-of-sale data from an on-premises SQL Server into a Fabric Lakehouse for transformation and reporting.

Which Fabric component should the team use for this recurring ingestion?

A) Manual CSV file export uploaded to OneLake storage through the browser-based file manager
B) Dataflow Gen2 configured with an on-premises data gateway connection to the SQL source
C) Azure Data Factory linked service deployed independently outside the Fabric workspace
D) Power BI Desktop using import mode to pull SQL data directly into a semantic model

 

Correct answers: B – Explanation:
Dataflow Gen2 with a gateway connects to on-premises sources, applies Power Query transformations, and loads directly into the Lakehouse on a configurable schedule. Manual CSV upload is not automated and requires human intervention each day. Power BI Desktop imports data into a semantic model, not into a Lakehouse Delta table. ADF deployed outside Fabric requires additional integration steps to land data in a Lakehouse. Source: Check Source

A data engineer needs to improve read performance on a 500-million-row Delta table that analysts filter by region and order date.

Which optimization should the engineer apply?

A) Partition the table by the high-cardinality primary key column for maximum distribution
B) Apply V-Order optimization and Z-Order the table on the region and order_date columns
C) Convert the Delta table to flat CSV format for simpler file-based access by analysts
D) Remove all file-level statistics from the Delta log to reduce metadata overhead

 

Correct answers: B – Explanation:
V-Order improves columnar read speed in the Fabric engine, while Z-Order co-locates data by frequently filtered columns to minimize the number of files scanned. CSV format lacks columnar optimization, predicate pushdown, and Delta transaction support. Removing statistics degrades the query planner’s ability to skip irrelevant files. High-cardinality partitioning creates excessive small files that degrade both metadata and scan performance. Source: Check Source

A logistics company implements a medallion architecture. Raw JSON telemetry lands in Bronze. The team needs PySpark to clean and deduplicate data into Silver.

Which Fabric item should orchestrate this transformation on a nightly schedule?

A) A Power BI datamart with scheduled refresh configured against the Bronze layer tables
B) A KQL Queryset running scheduled Kusto queries against the real-time intelligence database
C) A Fabric Notebook executed by a Data Pipeline activity on a recurring nightly schedule
D) A Dataflow Gen2 visual editor using only Power Query M for all transformation logic

 

Correct answers: C – Explanation:
Notebooks execute PySpark transformations with full programmatic control and can be orchestrated by Data Pipelines on a configurable schedule. Datamarts target self-service SQL analytics scenarios, not Spark-based ETL pipelines. KQL Querysets analyze streaming and real-time intelligence data, not batch medallion ETL. Dataflow Gen2 uses Power Query M which is less suited for complex PySpark deduplication logic requiring custom code. Source: Check Source

The Gold layer of a Lakehouse must feed Power BI dashboards with the lowest possible query latency while avoiding data duplication.

Which connectivity mode should the semantic model use?

A) Live Connection mode connecting to an externally hosted SQL Server Analysis Services instance
B) DirectQuery mode sending live SQL queries to the SQL analytics endpoint for each visual
C) Import mode copying all Gold layer data into the Power BI model with scheduled daily refresh
D) Direct Lake mode reading Gold Delta tables through the endpoint without importing or copying

 

Correct answers: D – Explanation:
Direct Lake reads Gold Delta tables directly via the SQL analytics endpoint without duplicating data into the model, providing near-import performance with real-time freshness. Import mode copies data, creating duplication and refresh latency. DirectQuery sends live queries but with higher per-query latency than Direct Lake. Live Connection references external SSAS, not native Fabric Lakehouse tables. Source: Check Source

A BI developer creates a semantic model needing year-over-year revenue growth that adjusts dynamically with any date slicer selection.

Which DAX function pattern should the developer use?

A) CALCULATE combined with SAMEPERIODLASTYEAR to shift the date filter context dynamically
B) CONCATENATEX building a formatted text string combining year labels with revenue figures
C) A simple SUM measure divided by a manually entered hard-coded prior-year revenue constant
D) RANKX ranking revenue values across year dimension members by descending magnitude

 

Correct answers: A – Explanation:
CALCULATE with SAMEPERIODLASTYEAR shifts the date filter context to the prior year, enabling dynamic YoY comparison that responds to any slicer selection. Hard-coded constants break when filters change or new data arrives. RANKX assigns ordinal positions rather than computing period comparisons. CONCATENATEX builds display text strings but cannot perform the arithmetic comparison needed for growth calculation. Source: Check Source

A finance team needs row-level security so regional managers see only their own region’s data in Power BI reports.

Which configuration should the BI developer implement?

A) Define RLS roles with DAX filter expressions on the Region table and assign manager accounts
B) Publish separate workspaces per region each containing a full duplicate of the semantic model
C) Set Power BI report-level page filters configured to each manager’s region before sharing
D) Create completely separate Power BI reports for each region with filtered underlying data

 

Correct answers: A – Explanation:
RLS roles apply DAX filters at query execution time so each manager sees only their assigned region from a single shared model. Separate reports multiply maintenance effort for each model change. Page-level filters can be cleared or modified by report viewers with edit permissions. Separate workspaces duplicate the entire model, increasing storage costs and complicating governance. Source: Check Source

A star schema with Sales facts and Product, Date, Store, Customer dimensions has slow queries due to complex relationship traversal.

What modeling improvement should the developer make?

A) Ensure all relationships are single-direction one-to-many from dimensions to the fact table
B) Flatten all dimension tables into the fact table creating a single wide denormalized table
C) Add bidirectional cross-filtering on every relationship to maximize query path flexibility
D) Remove the date dimension table and use calculated columns on the fact table for time logic

 

Correct answers: A – Explanation:
Single-direction one-to-many from dimensions to facts is the star schema pattern that the VertiPaq engine optimizes best for compression and query performance. Flattening loses dimensional modeling benefits and inflates model size. Bidirectional filtering introduces ambiguity in filter propagation and degrades performance. Removing the date dimension eliminates time intelligence functions like YTD, QTD, and period comparisons. Source: Check Source

A marketing analyst needs to detect spending outliers across 200 campaigns without building a formal ML model.

Which Fabric tool should the analyst use for this ad-hoc exploration?

A) A full Azure Machine Learning pipeline with registered experiments and model training
B) A paginated report with fixed layouts displaying campaign spending summary tables
C) A Fabric Notebook with interactive PySpark and matplotlib visualizations for exploration
D) The SQL analytics endpoint running only a single SELECT TOP 10 query for sampling

 

Correct answers: C – Explanation:
Fabric Notebooks enable interactive exploration with PySpark, pandas, and visualization libraries, ideal for iterative ad-hoc outlier analysis. A full ML pipeline is heavyweight for exploratory analysis without model training. Paginated reports display static formatted data without interactive exploration. A single TOP 10 query cannot identify outliers across 200 campaigns effectively. Source: Check Source

An HR team needs to explore employee attrition data using natural-language questions directly inside a Power BI report.

Which Power BI feature enables this?

A) A decomposition tree visual drilling through predetermined hierarchical dimension levels
B) An export to Excel button allowing users to pivot and filter data in external spreadsheets
C) Custom SQL queries typed into the Power BI Desktop query editor for each analysis question
D) The Q&A visual configured with a well-defined linguistic schema for natural-language queries

 

Correct answers: D – Explanation:
The Q&A visual lets users type natural-language questions and receive instant auto-generated visualizations powered by the semantic model and linguistic schema. Custom SQL in the query editor is developer-oriented and not accessible to business users. Decomposition trees explore fixed hierarchies but cannot handle arbitrary natural-language questions. Excel export loses interactive exploration and removes users from the report experience. Source: Check Source

Get 788+ more questions with source-linked explanations

Every answer traces to the exact Microsoft documentation page — so you learn from the source, not just memorize answers.

Exam mode & learn mode · Score by objective · Updated 16-Apr-26

Learn more...

What the DP-600 Fabric Analytics Engineer Associate exam measures

  • Plan, Implement, and Manage a Solution for Data Analytics (10–15%) — Evaluate your ability to implement and manage tasks within this domain, including real-world job skills and scenario-based problem solving.
  • Prepare and Serve Data (40–45%) — Evaluate your ability to implement and manage tasks within this domain, including real-world job skills and scenario-based problem solving.
  • Implement and Manage Semantic Models (20–25%) — Evaluate your ability to implement and manage tasks within this domain, including real-world job skills and scenario-based problem solving.
  • Explore and Analyze Data (20–25%) — Evaluate your ability to implement and manage tasks within this domain, including real-world job skills and scenario-based problem solving.

  • Review the official exam guide to understand every objective and domain weight before you begin studying
  • Complete the relevant Microsoft Learn learning path to build a structured foundation across all exam topics
  • Get hands-on practice in an Azure free-tier sandbox or trial environment to reinforce what you have studied with real configurations
  • Apply your knowledge through real-world project experience — whether at work, in volunteer roles, or contributing to open-source initiatives
  • Master one objective at a time, starting with the highest-weighted domain to maximize your score potential early
  • Use PowerKram learn mode to study by individual objective and review detailed explanations for every question
  • Switch to PowerKram exam mode to simulate the real test experience with randomized questions and timed conditions

Earning this certification can open doors to several in-demand roles:

Microsoft provides comprehensive free training to prepare for the DP-600 Fabric Analytics Engineer Associate exam. Start with the official Microsoft Learn learning path for structured, self-paced modules covering every exam domain. Review the exam study guide for the complete skills outline and recent updates.

Related certifications to explore

Related reading from our Learning Hub