
Within the two earlier posts of the Incremental Refresh in Energy BI sequence, now we have realized what incremental refresh is, find out how to implement it, and greatest practices on find out how to safely publish the semantic mannequin modifications to Microsoft Cloth (aka Energy BI Service). This publish focuses on a few extra greatest practices in implementing incremental refresh on giant semantic fashions in Energy BI.
Notice
Since Might 2023 that Microsoft introduced Microsoft Cloth for the primary time, Energy BI is part of Microsoft Cloth. Therefore, we use the time period Microsoft Cloth all through this publish to discuss with Energy BI or Energy BI Service.
Implementing incremental refresh on Energy BI is often easy if we rigorously observe the implementation steps. Nevertheless in some real-world situations, following the implementation steps just isn’t sufficient. In several elements of my newest e book, Professional Information Modeling with Energy BI, 2’nd Version, I emphasis the truth that understanding enterprise necessities is the important thing to each single improvement venture and knowledge modelling is not any totally different. Let me clarify it extra within the context of incremental knowledge refresh implementation.
Let’s say we adopted all of the required implementation steps and we additionally adopted the deployment greatest practices and every little thing runs fairly good in our improvement surroundings; the primary knowledge refresh takes longer, we we anticipated, all of the partitions are additionally created and every little thing appears superb. So, we deploy the answer to manufacturing surroundings and refresh the semantic mannequin. Our manufacturing knowledge supply has considerably bigger knowledge than the event knowledge supply. So the info refresh takes means too lengthy. We wait a few hours and depart it to run in a single day. The subsequent day we discover out that the primary refresh failed. A few of the potentialities that lead the primary knowledge refresh to fail are Timeout, Out of sources, or Out of reminiscence errors. This may occur no matter your licensing plan, even on Energy BI Premium capacities.
One other situation chances are you’ll face often occurs throughout improvement. Many improvement groups attempt to preserve their improvement knowledge supply’s measurement as shut as attainable to their manufacturing knowledge supply. And… NO, I’m NOT suggesting utilizing the manufacturing knowledge supply for improvement. Anyway, chances are you’ll be tempted to take action. You set one month’s value of information utilizing the RangeStart and RangeEnd parameters simply to search out out that the info supply truly has tons of of hundreds of thousands of rows in a month. Now, your PBIX file in your native machine is means too giant so you can not even reserve it in your native machine.
This publish supplies some greatest practices. A few of the practices this publish focuses on require implementation. To maintain this publish at an optimum size, I save the implementations for future posts. With that in thoughts, let’s start.
To this point, now we have scratched the floor of some frequent challenges that we could face if we don’t take note of the necessities and the scale of the info being loaded into the info mannequin. The excellent news is that this publish explores a few good practices to ensure smoother and extra managed implementation avoiding the info refresh points as a lot as attainable. Certainly, there would possibly nonetheless be instances the place we observe all greatest practices and we nonetheless face challenges.
Notice
Whereas implementing incremental refresh is out there in Energy BI Professional semantic fashions, however the restrictions on parallelism and lack of XMLA endpoint could be a deal breaker in lots of situations. So most of the methods and greatest practices mentioned on this publish require a premium semantic mannequin backed by both Premium Per Person (PPU), Energy BI Capability (P/A/EM) or Cloth Capability.
The subsequent few sections clarify some greatest practices to mitigate the dangers of dealing with tough challenges down the street.
Observe 1: Examine the info supply when it comes to its complexity and measurement
This one is simple; not likely. It’s essential to know what sort of beast we’re coping with. When you have entry to the pre-production knowledge supply or to the manufacturing, it’s good to understand how a lot knowledge shall be loaded into the semantic mannequin. Let’s say the supply desk accommodates 400 million rows of information for the previous 2 years. A fast math means that on common we could have greater than 16 million rows monthly. Whereas these are simply hypothetical numbers, you will have even bigger knowledge sources. So having some knowledge supply measurement and progress estimation is all the time useful for taking the following steps extra totally.
Observe 2: Preserve the date vary between the RangeStart and RangeEnd small
Persevering with from the earlier follow, if we cope with pretty giant knowledge sources, then ready for hundreds of thousands of rows to be loaded into the info mannequin at improvement time doesn’t make an excessive amount of sense. So relying on the numbers you get from the earlier level, choose a date vary that’s sufficiently small to allow you to simply proceed together with your improvement with no need to attend a very long time to load the info into the mannequin with each single change within the Energy Question layer. Keep in mind, the date vary chosen between the RangeStart and RangeEnd does NOT have an effect on the creation of the partition on Microsoft Cloth after publishing. So there wouldn’t be any points when you selected the values of the RangeStart and RangeEnd to be on the identical day and even at the very same time. One essential level to recollect is that we can not change the values of the RangeStart and RangeEnd parameters after publishing the mannequin to Microsoft Cloth.
Observe 3: Be aware of variety of parallelism
As talked about earlier than, one of many frequent challenges arises after the semantic mannequin is printed to Microsoft Cloth and is refreshed for the primary time. It isn’t unusual to refresh giant semantic fashions that the primary refresh will get timeout and fails. There are a few potentialities inflicting the failure. Earlier than we dig deeper, let’s take a second to remind ourselves of what actually occurs behind the scenes on Microsoft Cloth when a semantic mannequin containing a desk with incremental refresh configuration refreshes for the primary time. To your reference, this publish explains every little thing in additional element.
What occurs in Microsoft Cloth to semantic fashions containing tables with incremental refresh configuration?
Once we publish a semantic mannequin from Energy BI Desktop to Microsoft Cloth, every desk within the printed semantic mannequin has a single partition. That partition accommodates all rows of the desk which can be additionally current within the knowledge mannequin on Energy BI Desktop. When the primary refresh operates, Microsoft Cloth creates knowledge partitions, categorised as incremental and historic partitions, and optionally a real-time DirectQuery partition based mostly on the incremental refresh coverage configuration. When the real-time DirectQuery partition is configured, the desk is a Hybrid desk. I’ll talk about Hybrid tables in a future publish.
Microsoft Cloth begins loading the info from the info supply into the semantic mannequin in parallel jobs. We will management the parallelism from the Energy BI Desktop, from Choices -> CURRENT FILE -> Information Load -> Parallel loading of tables. This configuration controls the variety of tables or partitions that shall be processed in parallel jobs. This configuration impacts the parallelism of the present file on Energy BI Desktop whereas loading the info into the native knowledge mannequin. It additionally influences the parallelism of the semantic mannequin after publishing it to Microsoft Cloth.

Because the previous picture exhibits, I elevated the Most variety of concurrent jobs to 12.
The next picture exhibits refreshing the semantic mannequin with 12 concurrent jobs on a Premium workspace on Microsoft:

The default is 6 concurrent jobs, which means that after we refresh the mannequin in Energy BI Desktop or after publishing it to Microsoft Cloth, the refresh course of picks 6 tables, or 6 partitions to run in parallel.
The next picture exhibits refreshing the semantic mannequin with the default concurrent jobs on a Premium workspace on Microsoft:

Tip
I used the Analyse my Refresh instrument to visualise my semantic mannequin refreshes. A giant shout out to the legendary Phil Seamark for creating such a tremendous instrument. Learn extra about find out how to use the instrument on Phil’s weblog.
We will additionally change the Most variety of concurrent jobs from third-party instruments corresponding to Tabular Editor; due to the superb Daniel Otykier for creating this glorious instrument. Tabular Editor makes use of the SSAS Tabular mannequin property referred to as MaxParallelism which is proven as Max Parallelism Per Refresh on the instrument (take a look at the under picture from Tabular Editor 3).

Whereas loading the info in parallel would possibly enhance the efficiency, relying on the info quantity being loaded into every partition, the concurrent question limitations on the info supply, and the useful resource availability in your capability, there may be nonetheless a danger of getting timeouts. In order a lot as growing the Most variety of concurrent jobs is tempting, it’s suggested to alter it with care. It’s also worthwhile to say that the behaviour of Energy BI Desktop in refreshing the info is totally different from Microsoft Cloth’s semantic mannequin knowledge refresh exercise. Due to this fact, whereas altering the Most variety of concurrent jobs could affect the engine on Microsoft Cloth’s semantic mannequin, it doesn’t assure of getting higher efficiency. I encourage you to learn Chris Webb’s weblog on this matter.
Observe 4: Think about making use of incremental insurance policies with out partition refresh on premium semantic fashions
When working with giant premium semantic fashions, implementing incremental refresh insurance policies is a key technique to handle and optimise knowledge refreshes effectively. Nevertheless, there could be situations the place we have to apply incremental refresh insurance policies to our semantic mannequin with out instantly refreshing the info throughout the partitions. This follow is especially helpful to manage the heavy lifting of the preliminary knowledge refresh. By doing so, we be sure that our mannequin is prepared and aligned with our incremental refresh technique, with out triggering a time-consuming and resource-intensive knowledge load.
There are a few methods to realize this. The best means is to make use of Tabular Editor to use the incremental coverage which means that each one partitions are created however they aren’t processed. The next picture exhibits the previous course of:

The opposite technique that some builders would possibly discover helpful, particularly if you’re not allowed to make use of third-party instruments corresponding to Tabular Editor is so as to add a brand new question parameter within the Energy Question Editor on Energy BI Desktop to manage the info refreshes. This technique ensures that the primary refresh of the semantic mannequin after publishing it to Microsoft Cloth could be fairly quick with out utilizing any third-party instruments. Which means that Microsoft Cloth creates and refreshes (aka processes) the partitions, however since there isn’t a knowledge to load, the processing could be fairly fast.
The implementation of this method is straightforward; we outline a brand new question parameter. We then use this new parameter to filter out all knowledge from the desk containing incremental refresh. In fact, we would like this filter to fold so your entire question on the Energy Question facet is absolutely foldable. So after we publish the semantic mannequin to Microsoft Cloth, we apply the preliminary refresh. For the reason that new question parameter is accessible through the semantic mannequin’s settings on Microsoft Cloth, we modify its worth after the preliminary knowledge refresh to load the info when the following knowledge refresh takes place.
It is very important notice that altering the parameter’s worth after the preliminary knowledge refresh is not going to populate the historic Vary. It signifies that when the following refresh occurs, Microsoft Cloth assumes that the historic partitions are already refreshed and ignores them. Due to this fact, after the preliminary refresh the historic partitions stay empty, however the incremental partitions shall be populated. To refresh the historic partitions we have to manually refresh them through XMLA endpoints which will be performed utilizing SSMS or Tabular Editor.
Explaining the implementation of this technique makes this weblog very lengthy so I reserve it for a separate publish. Keep tuned if you’re serious about studying find out how to implement this method.
Observe 5: Validate your partitioning technique earlier than implementation
Partitioning technique refers to planning how the info goes to be divided into partitions to match the enterprise necessities. For instance, let’s say we have to analyse the info for 10 years. As knowledge quantity to be loaded right into a desk is giant, it doesn’t make sense to truncate the desk and absolutely refresh it each night time. Through the discovery workshops, you came upon that the info modifications every day and it’s extremely unlikely for the info to alter as much as 7 days.
Within the previous state of affairs, the historic vary is 10 years and the incremental vary is 7 days. As there aren’t any indications of any real-time knowledge change necessities, there isn’t a have to preserve the incremental vary in DirectQuery mode which turns our desk right into a hybrid desk.
The incremental coverage for this state of affairs ought to appear to be the next picture:

So after publishing the semantic mannequin to Microsoft Cloth and the primary refresh, the engine solely refreshes the final 7 partitions on the following refreshes as proven within the following picture:

Deciding on the incremental coverage is a strategic determination. An inaccurate understanding of the enterprise necessities results in an inaccurate partitioning technique, therefore inefficient incremental refresh which may have some critical unwanted effects down the street. That is a kind of instances that may result in erasing the present partitions, creating new partitions, and refreshing them for the primary time. As you’ll be able to see, a easy mistake in our partitioning technique will result in incorrect implementation that results in a change within the partitioning coverage which suggests a full knowledge load shall be required.
Whereas understanding the enterprise necessities in the course of the discovery workshops is significant, everyone knows that the enterprise necessities evolve occasionally; and actually, the tempo of the modifications is usually fairly excessive.
For instance, what occurs if a brand new enterprise requirement comes up involving real-time knowledge processing for the incremental vary aka hybrid desk? Whereas it would sound to be a easy change within the incremental refresh configuration, in actuality, it isn’t that straightforward. To clarify extra, to get one of the best out of a hybrid desk implementation, we must always flip the storage mode of all of the related dimensions to the hybrid desk into Twin mode. However that isn’t a easy course of both if the present dimensions’ storage modes are already set to Import. We can not swap the storage mode of the tables from Import to both Twin or DirectQuery modes. Which means that now we have to take away and add these tables once more which in real-world situations just isn’t that straightforward. As talked about earlier than I’ll write one other publish about hybrid tables sooner or later, so chances are you’ll contemplate subscribing to my weblog to get notified on all new posts.
Observe 6: Think about using the Detect knowledge modifications for extra environment friendly knowledge refreshes
Let’s clarify this part utilizing our earlier instance the place we configured the incremental refresh to archive 10 years of information and incrementally refresh 7 days of information. This implies Energy BI is configured to solely refresh a subset of the info, particularly the info from the final 7 days, moderately than your entire semantic mannequin. The default refreshing mechanism in Energy BI for tables with incremental refresh configuration is to maintain all of the historic partitions intact, truncate the incremental partitions, and reload them. Nevertheless in situations coping with giant semantic fashions, the incremental partitions may very well be pretty giant, so the default truncation and cargo of the incremental partitions wouldn’t be an optimum strategy. Right here is the place the Detect knowledge modifications characteristic can assist. Configuring this characteristic within the incremental coverage requires an additional DateTime column, corresponding to LastUpdated, within the knowledge supply which is utilized by Energy BI to first detect the info modifications, then solely refresh the precise partitions which have modified because the earlier refresh as a substitute of truncating and reloading all incremental partitions. Due to this fact, the refreshes probably course of smaller quantities of information utilising fewer sources in comparison with common incremental refresh configuration. The column used for detecting knowledge modifications have to be totally different from the one used to partition the info with the _RangeStart and RangeEnd parameters. Energy BI makes use of the utmost worth of the column used for outlining the Detect knowledge modifications characteristic to establish the modifications from the earlier refresh and solely refreshes the modified partitions and shops it within the refreshBookmark property of the partitions throughout the incremental vary.
Whereas the Detect knowledge modifications can enhance the info refresh efficiency, we are able to improve it even additional. One attainable enhancement could be to keep away from importing the LastUpdated column into the semantic mannequin which is more likely to be a high-cardinality column. One possibility is to create a brand new question throughout the Energy Question Editor in Energy BI Desktop to establish the utmost date throughout the date vary filtered by the RangeStart and RangeEnd parameters. We then use this question within the pollingExpression property of our refresh coverage. This may be performed in varied methods corresponding to operating TMSL scripts through XMLA endpoint* or utilizing Tabular Editor. I may even clarify this technique in additional element in a future publish, so keep tuned.
This publish of the Incremental Refresh in Energy BI sequence delved into some greatest practices for implementing incremental refresh methods, significantly for big semantic fashions, and underscored the significance of aligning these methods with enterprise necessities and knowledge complexities. We’ve navigated by frequent challenges and supplied sensible greatest practices to mitigate dangers, enhance efficiency, and guarantee smoother knowledge refresh processes. I’ve a few extra blogs from this sequence in my pipeline so keep tuned for these and subscribe to my weblog to get notified after I publish a brand new publish. I hope you loved studying this lengthy weblog and discover it useful.
As all the time, be happy to depart your feedback and ask questions, observe me on LinkedIn and @_SoheilBakhshi on X (previously Twitter).
Associated