Microsoft Fabric Deployment Pipelines

What are Deployment Pipelines? Pipeline Structure Stage Comparison in UI Item Pairing & Status Selective Deployment Deployment Rules Lakehouse & Deployment Pipelines in Fabric Deployment pipelines move Lakehouse metadata, including shortcuts, but do not copy data or table schemas. After deployment, shortcuts still point to the original source, and the Lakehouse will be empty unless … Read more

cdm_Archive_to_STG

Overview The cdm_Archive_to_STG notebook is a critical component in the cdm_today and cdm_Archive pipeline. Its primary role is to create a staging table that represents a point-in-time snapshot of source data, which is then used by the cdm__Archive_upsert notebook to accurately update the main dimension table. This staging layer acts as a buffer between raw … Read more

myOSH_To_Bronze

Overview The myOSH_To_Bronze notebook is designed to handle multiple MyOSH API endpoints dynamically, meaning it can process different endpoints such as records or users without hardcoding logic. For most endpoints (like users), the process is straightforward: it performs a regular API call and writes the response directly to JSON in the source container and to … Read more

My_osh_Archive

Overview The myosh_Archive notebook is designed to maintain a complete and accurate archive of records from the MyOSH API. It ensures that no data is lost by identifying gaps in the current API response and backfilling missing records, then merging everything into a single, consolidated archive stored in the Azure Data Lake container. Why Do … Read more

cdm__Archive_upsert

Overview The cdm__Archive_upsert notebook is designed to help keep accurate historical records of data. It implements a technique called Slowly Changing Dimension Type 2 (SCD2), which means instead of overwriting old information, it creates a timeline of changes so you can see what was true at any point in time. This method ensures that previous … Read more

Refresh SQL Endpoint Metadata

When working with Microsoft Fabric, one common challenge is keeping the SQL endpoint in sync with the Lakehouse endpoint after a pipeline run. Fabric provides two endpoints for every Lakehouse: Why does the SQL endpoint sometimes lag? The SQL endpoint doesn’t automatically refresh its metadata the moment new data lands in the Lakehouse. Instead, a … Read more

Running a Stored Procedure for Multiple Dates in Power BI

Purpose:To run the stored procedure get.myArchive from Power BI for a series of previous Sundays (for example, the last six Sundays), and optionally combine or parameterise the results. 🧭 Background The get.myArchive procedure is designed to retrieve snapshot data for a specified date.By default, Power BI passes a single date parameter such as: However, when … Read more

Power BI guide: managing the current views & a simple pivot pattern

This note explains how to turn the current views (wide–from–tall) into a report‑friendly table using Power Query (recommended) or DAX. It covers two common layouts: It also shows how to extract a human‑readable Description from the JSON in VALUE_METADATA. Source schema (tall, meta-driven) Expected columns: Assumption: Each (OBJECT_SEQ, [META_TYPE], ATTRIBUTE) appears at most once for … Read more

Look in the box first

Should We Reuse Standard Dynamics 365 / Dataverse Tables for Our In‑House Apps? 1) Why We’re Talking About This We’re starting an important conversation: when we build apps internally, should we lean on the standard tables already in Dynamics 365 / Dataverse (like Account, Contact, Opportunity, Project), or should we keep spinning up new SharePoint … Read more

🌟 Transparency Meets Simplicity – Updates to the Data Dictionary in myBMT

Hey everyone 👋 We’ve been working hard behind the scenes on something that we think will really help you. 🔍 What’s New? We’ve significantly expanded the Attribute Lineage and Data Dictionary tooling in myBMT. This means: It’s a fully transparent journey from requirement ➡ mapping ➡ code — all in one place. 😅 One Small … Read more