In my view dataflows definitely have added valueâŚ
Dataflows leverage Power Query, so you can start using them quickly.
If your company doesnât have Data Architects and/or a Data Warehouse for example, you could use dataflows as a central location to prep and have your data ready for consumption.
** develop and design just once
** schedule automatic refresh
** becomes the single source of truth
** with one artifact to maintain (should that need ever arise)
In report development youâll never have to query directly against your database (performance).
Speeds up Model development because all building blocks (Tables) are ready for use.
Dataflows can also be used to store data from legacy systems:
** setup once
** refresh once
** enjoy forever
Dataflows can be used in other tools and services, theyâre not exclusive to Power BI.
Dataflows can be mapped to CDS, the Common Data Service wich is a set of a standardized data schemas and a metadata system that allows consistency of data and its meaning across applications and business processes.
Dataflows run in the cloud. No additional infrastructure required.
Incremental refresh is available for dataflows
Just to name a few
.
New article and more food for thoughtâŚ
.
and just with everything else in the âpowerâ stack itâs in constant development.
Regarding the mapping to CDS, it has some value but it would be better that the powerplatform creates an accessible solution to hook up to display names used in solutions. All the mapping does is convert logical names to display names but those are generally also altered in solutions.
Another weird thing about dataflows. You basically have to create them in the powerbi service, but you can only create reports from them in the desktop? You canât really make them in power bi desktop because you will need to copy all the queries still manually to the service. Makes absolutely zero sense.
And one more annoyance ⌠itâs quite slow to setup, so would still recommend to still create everything in desktop and then copy/paste into the dataflow in the service. Just done a few larger ones (48 entities) on a premium capacity tennant ⌠get used to wait for each change of a query step. Also the validations of the query steps could push the limits of your patience as sometimes, they only pop-up much later (or sometimes they are already resolved but they just stay there as they werenât picked up on change).
But the concept is really good to enable self-service reporting at enterprise level.
PS: also ⌠donât use hashtags in query names. Unlike in desktop, it breaks the solution.
Why is that weird?
If you had a Datawarehouse youâd design and store your data there. Now through Power Query Online in the Power BI Service you are storing queries in Azure Data Lake Storage Gen2. Makes sense to me.
.
Well that depends⌠if you have a datamodel based on dataflows ready for use in the Power BI Service you can create a report there. However if that is not the case and you have to design a datamodel then youâll have to use Desktop, the actual restriction here is datamodel design not the data(flow) source.
.
I do agree it can be slow to set up. Like you most of the time I set up my queries in desktop and copy them to the Power BI Service.
Weird because:
For report creation based on dataflow: you need desktop (you would assume you can just a select a dataflow as source).
For dataflow creation, you need online (you would assume you can just publish as dataflow).
=> one needs the other yet you cannot do it on the same âplatformâ.
So regarding the that depends (not sure how you quote):
When you have set up a dataflow in the service, you still canât use it for report creation in the service. You first need to setup a report in desktop using that dataflow and then publish it to the service. After which you have a dataset which can only then use further in the service.
True self-service would be: the dataflow has been created in the service, you can now start creating dataset/reports from that in the service.
Think thatâs a misunderstanding for report creation you need a datamodel or a shared dataset. However one of the data sources for that model can be a dataflowâŚ
No argument there!
.
RADACAD did an article on dataflow vs dataset, you can find that here.
. To quote, just highlight that section and the option to âQuoteâ will become visible. The quote will be entered in the edit widow at the position of your cursor.
dataflow is awesome. I can run multiple thread with different refresh interval. Product refresh maybe good enough for once a day, but onhand/sales transaction may be refreshed every several hours/day
Just wanted to give you all a heads up on this post by Matthew Roche, see key points below.
Thereâs a new Power Automate (Flow) connector that will enable any authorized user to take action based on dataflow refresh completion, including:
Sending a notification â so you can keep additional users and groups informed
Trigger a dataflow or dataset refresh â so you donât need to maintain separate refresh schedules or live with unnecessary processing delays
Trigger a Power Automate flow â so you can do pretty much whatever you want, because Power Automate has a ton of capabilities this blog post wonât even attempt to mention
You can also use the connector to refresh dataflows and monitor dataflow refresh status.
.
Read the full post by Matthew Roche here on his blog.
or check out the official Power BI introduction post here.