Get the FREE Ultimate OpenClaw Setup Guide →

fabric-lakehouse-views-perf-remediate

npx machina-cli add skill PatrickGallucci/fabric-skills/fabric-lakehouse-views-perf-remediate --openclaw
Files (1)
SKILL.md
13.6 KB

Fabric Materialized Lake Views Performance remediate

Diagnose and resolve performance issues with materialized lake views (MLVs) in Microsoft Fabric lakehouses. This skill covers refresh optimization, Spark job diagnostics, data quality constraint tuning, and lineage execution remediate.

When to Use This Skill

  • MLV refresh runs are taking longer than expected
  • Incremental refresh is falling back to full refresh unexpectedly
  • MLV lineage execution shows Failed or Skipped nodes
  • Spark jobs for MLV refresh are failing with errors
  • "Delta table not found" errors during MLV creation or refresh
  • Data quality constraints causing unexpected pipeline failures
  • Need to enable or verify optimal refresh configuration
  • Custom Spark environment tuning for MLV workloads
  • Monitoring and interpreting MLV run history

Prerequisites

  • Microsoft Fabric workspace with Lakehouse items
  • Schema-enabled lakehouse (recommended for MLV support)
  • Fabric notebook for executing Spark SQL commands
  • Workspace Admin or Contributor role for scheduling and monitoring
  • Access to Monitor Hub for viewing MLV run details

Quick Diagnostics Checklist

Run through these checks in order when remediate MLV performance:

StepCheckAction
1Identify refresh modeVerify optimal refresh toggle is enabled in lineage view
2Check CDF statusConfirm delta.enableChangeDataFeed=true on ALL source tables
3Review query patternsEnsure only supported SQL constructs are used (see supported expressions)
4Inspect run historyOpen lineage view dropdown to see last 25 runs and their states
5Check node failuresClick failed nodes in lineage to view error messages
6Review Spark logsFollow Detailed Logs link to Monitor Hub for Spark error logs
7Validate data qualityCheck if FAIL constraints are causing "delta table not found" errors
8Assess source dataDetermine if source has updates/deletes (forces full refresh)

Step-by-Step Workflows

Workflow 1: Diagnose Slow Refresh

  1. Determine current refresh strategy - Navigate to Manage materialized lake views in the lakehouse ribbon. Check if Optimal refresh toggle is ON.
  2. Verify change data feed - Run the diagnostic script to check CDF status on all source tables. See scripts/check-cdf-status.sql.
  3. Check for unsupported expressions - Review the MLV definition for constructs that force full refresh. See Supported Expressions.
  4. Inspect source data patterns - If source tables have UPDATE or DELETE operations, Fabric always performs full refresh regardless of CDF status.
  5. Review partition strategy - Consider adding PARTITIONED BY to large MLVs to improve refresh parallelism.
  6. Attach custom environment - Configure a custom Spark environment with optimized compute for heavy workloads. See references/custom-environment-guide.md.

Workflow 2: Resolve Failed Refresh Runs

  1. Open lineage view - Navigate to Manage materialized lake views, select the failed run from the dropdown (last 25 runs available).
  2. Identify failed node - Click the failed node in the lineage graph. Review the error message in the right-side panel.
  3. Access Spark logs - Click the Detailed Logs link to navigate to Monitor Hub. Review Apache Spark error logs.
  4. Common failure patterns - See references/common-failure-patterns.md for resolution steps.
  5. Retry or recreate - For transient Spark failures, retry the run. For persistent errors, drop and recreate the MLV.

Workflow 3: Enable Optimal Refresh

  1. Enable CDF on all source tables:
ALTER TABLE bronze.customers SET TBLPROPERTIES (delta.enableChangeDataFeed = true);
ALTER TABLE bronze.orders SET TBLPROPERTIES (delta.enableChangeDataFeed = true);
  1. Verify CDF is enabled:
DESCRIBE DETAIL bronze.customers;
-- Check properties column for delta.enableChangeDataFeed=true
  1. Create MLV with CDF property:
CREATE OR REPLACE MATERIALIZED LAKE VIEW silver.customer_orders
TBLPROPERTIES (delta.enableChangeDataFeed=true)
AS
SELECT 
    c.customerID,
    c.customerName,
    c.region,
    o.orderDate,
    o.orderAmount
FROM bronze.customers c INNER JOIN bronze.orders o
ON c.customerID = o.customerID;
  1. Enable optimal refresh toggle - Navigate to Manage materialized lake views and verify the Optimal refresh toggle is ON (enabled by default).

Supported Expressions for Incremental Refresh

MLVs using only these constructs qualify for incremental refresh:

SQL ConstructNotes
SELECTOnly deterministic built-in functions. Non-deterministic and window functions force full refresh.
FROMStandard table references
WHEREOnly deterministic built-in functions
INNER JOINSupported for incremental
WITH (CTE)Common table expressions supported
UNION ALLSupported
CONSTRAINT ... CHECKOnly deterministic built-in functions in constraint conditions

Unsupported constructs that force full refresh:

  • LEFT JOIN, RIGHT JOIN, FULL OUTER JOIN
  • Window functions (ROW_NUMBER, RANK, LAG, LEAD, etc.)
  • Non-deterministic functions (current_timestamp(), rand(), etc.)
  • Subqueries in SELECT or WHERE
  • GROUP BY with HAVING
  • DISTINCT
  • User-defined functions (UDFs)

Key Spark SQL Reference

List All MLVs

SHOW MATERIALIZED LAKE VIEWS IN silver;

View MLV Definition

SHOW CREATE MATERIALIZED LAKE VIEW silver.customer_orders;

Force Full Refresh

REFRESH MATERIALIZED LAKE VIEW silver.customer_orders FULL;

Drop and Recreate

DROP MATERIALIZED LAKE VIEW silver.customer_orders;

CREATE OR REPLACE MATERIALIZED LAKE VIEW silver.customer_orders
TBLPROPERTIES (delta.enableChangeDataFeed=true)
AS
SELECT ...;

Known Issues and Limitations

IssueImpactWorkaround
FAIL constraint + "delta table not found"MLV creation or refresh failsRecreate MLV using DROP action instead of FAIL
Schema names with ALL CAPSMLV creation failsUse lowercase or mixed-case schema names
Session-level Spark propertiesNot applied during scheduled refreshSet properties in custom environment instead
Delta time travel in MLV definitionNot supportedRemove VERSION AS OF or TIMESTAMP AS OF from queries
DML statements on MLVsNot supportedMLVs are read-only; modify source tables instead
UDFs in SELECTNot supportedUse built-in Spark SQL functions
Temporary views as MLV sourceNot supportedReference base tables or other MLVs directly
Cross-lakehouse lineageNot supportedKeep all related MLVs within same lakehouse
Updating data quality constraintsNot supportedDrop and recreate the MLV with new constraints
LIKE/regex in constraint conditionsNot supportedUse simple comparison operators in constraints
Service principal authenticationNot supported for MLV APIsUse Microsoft Entra user authentication
Single schedule per lineageUI instability if exceededMaintain only one active schedule per lineage
South Central US regionFeature not availableUse a workspace in a different region

Run States Reference

StateMeaningAction
CompletedAll nodes executed successfullyNo action needed
FailedOne or more nodes failed; child nodes skippedReview failed node error, check Spark logs
SkippedPrevious run still in progressWait for current run to complete, or cancel it
In ProgressRun started, not yet terminalMonitor progress in lineage view
CanceledManually canceled from Monitor HubRe-trigger if needed

remediate

SymptomLikely CauseResolution
Refresh always full, never incrementalCDF not enabled on source tablesRun ALTER TABLE ... SET TBLPROPERTIES (delta.enableChangeDataFeed = true) on ALL sources
Refresh always full despite CDF enabledUnsupported SQL constructs in MLVReview definition for window functions, LEFT JOIN, non-deterministic functions
Refresh always full despite CDF and supported SQLSource has UPDATE/DELETE operationsIncremental only supports append-only; redesign ETL to be append-only or accept full refresh
"Delta table not found" on createFAIL constraint issueRecreate MLV without FAIL; use DROP action instead
Node shows Skipped stateParent node failedFix the parent node failure first
Schedule not runningPrevious run still in progressCancel the stuck run from Monitor Hub, or wait
Environment not accessibleUser lacks access to selected environmentSelect an accessible environment from dropdown
Deleted environment errorAssociated environment was removedChoose a new environment in lineage settings
MLV names changed to lowercaseExpected behaviorMLV names are case-insensitive and stored as lowercase
Workspace names with spaces cause errorsBacktick syntax requiredUse backtick notation:`My Workspace`.lakehouse.schema.view_name

References

Source

git clone https://github.com/PatrickGallucci/fabric-skills/blob/main/skills/fabric-lakehouse-views-perf-remediate/SKILL.mdView on GitHub

Overview

Diagnose and resolve performance issues with Microsoft Fabric lakehouse materialized lake views (MLVs). This skill covers slow refreshes, incremental vs full refresh behavior, Spark job failures, and data quality constraints, guiding you to optimize refresh configuration and monitor runs in Monitor Hub.

How This Skill Works

Follow a structured diagnostic flow: verify the refresh configuration and optimal toggle, check Change Data Feed (CDF) status on all source tables, review MLV definitions for unsupported expressions, and inspect Spark logs in Monitor Hub. Apply targeted fixes such as enabling CDF, adjusting partitioning (PARTITIONED BY), and tuning a custom Spark environment to improve incremental refresh eligibility and overall performance.

When to Use It

  • MLV refresh runs are noticeably slow
  • Incremental refresh falls back to a full refresh
  • MLV lineage shows Failed or Skipped nodes
  • Spark jobs for MLV refresh fail with errors
  • Delta table not found errors during MLV creation or refresh

Quick Start

  1. Step 1: Open Manage materialized lake views and verify the Optimal refresh toggle is ON
  2. Step 2: Run scripts/check-cdf-status.sql to confirm CDF is enabled on all source tables
  3. Step 3: Review MLV definitions for unsupported expressions and consult Monitor Hub Spark logs for errors

Best Practices

  • Verify the Optimal refresh toggle is ON in the lineage view
  • Ensure delta.enableChangeDataFeed=true on all source tables
  • Review MLV definitions for unsupported expressions that trigger full refresh
  • Consult Monitor Hub spark logs for error details
  • Apply PARTITIONED BY to large MLVs and tune the Spark environment for workloads

Example Use Cases

  • Diagnosing a slow MLV refresh by analyzing Monitor Hub run history and enabling CDF
  • Fixing incremental refresh by removing unsupported expressions in the MLV
  • Resolving 'delta table not found' by correcting CDF settings and data quality constraints
  • Diagnosing Spark job failures and applying a custom Spark environment for MLV workloads
  • Adding PARTITIONED BY to large MLVs to improve parallelism and refresh duration

Frequently Asked Questions

Add this skill to your agents
Sponsor this space

Reach thousands of developers