We are testing ingestion of Power BI. We are seein...
# integrate-powerbi-datahub
f
We are testing ingestion of Power BI. We are seeing an issue where all ingested datasets are being placed into one random workspace. Have tried deleting all content and ingesting again several times, same problem. Has anyone else experienced this? We are running 0.11.0, Fairly sure this is related to the two settings: extract_datasets_to_containers: true extract_workspaces_to_containers: true so it happens when both are set to true, the last one being default, but makes no difference if it is explicitly set or not. Setting extract_workspaces_to_containers to false throws an error on my system, but that could/would be an okay way to fix it if it worked. (the error is 'Mapper' object has no attribute 'workspace_key'). Only ingesting workspaces works as tables are placed correctly, the issue is when enabling the dataset_to_container which we really need as the Power BI dataset is the analytical asset we want (would be fine if the tables could be excluded actually). Has anyone else successfully ingested both workspaces and datasets? Could be a Python version issue perhaps, so if you have please let me know what versions you are running.
no one has seen this behaviour?
Does anyone have this working e.g. datasets placed correctly in workspaces, if so can you please list version of Python and Datahub as well as the config/yaml file for setting up ingestion?
a
We have the same settings and the exact same behavior, but have not found the issue yet.
f
Thank god 🙂 was starting to worry that I was the only one.....I have tried most/all combinations and always with the same result....hope someone can shed light on this....
a
Yeah, same here. The only hunch I have is: it has to do with setting a variable to a workspace id and then when an exception occurs the variable doesn't get updated. So it stays pointing to the wrong workspace, logic recovers from the exception and continues sending datasets to the wrong workspace.
f
Makes sense Nikola, think we have exceptions for all ingested workspaces more or less, so that could make good sense. Hope someone from the development team or someone else can shed light on this. For us this 'bug' means we cannot use Power BI data which in terms severely limits the use of Datahub in our specific case.
Seems strange no one else beside us are seeing this (or seeing it as an issue). Can some one also using Datahub for ingesting Power BI giver some feedback?
We upgraded the CLI to 0.12.1.3 and now it seems to work correctly. Have ingested several times and datasets/semantic models are correctly placed in their respective workspace. Next issue is now that not all models get their schema extracted, haven't found any clue yet as to what the issue is, but lets just say that Power BI ingestion is still very much a work in progress -)
a
Thanks for the tip, we will upgrade as well and give it a try too.
f
short update, seems the missing schemas might be due to some datasets are test versions of production sets, seems like when both gets ingested one loses the schema.
a
Confirming that 0.12.1 solved the problem for us too. Thanks for the fix, @fast-barista-4910!