https://datahubproject.io logo
#integrate-tableau-datahub
Title
# integrate-tableau-datahub
g

gifted-diamond-19544

09/27/2022, 1:29 PM
Hello. I am currently facing a problem in my Tableau ingestion pipeline, where I am getting the NODE_LIMIT_EXCEEDED warning. I am currently making the ingestion from the UI. I have set the
page_size
to 1, as instructed in the docs, however I am still getting the error. So what I did was, instead of trying to ingest all the Tableau projects on the same pipeline, I create various pipeline with just a subset of the projects, and scheduled them with a few minutes offset. This seems to be working, however it is kinda of cumbersome. I think it would be great to add an option to the Tableau ingestion recipe that specifies a time interval between the extraction of each Tableau project. I have tried this using the Python emiter (basically I put a sleep statement between the extraction of each project), and this solved the problem. However, since I am not using the UI, I don’t see an easy way to achieve this. Does anyone have any solution for this problem, when making the ingestion vie the UI? Thank you!
cc @chilly-knife-8692
m

modern-artist-55754

09/27/2022, 1:56 PM
The latest version of datahub cli has some fixes for that, you may want to try it.
g

gifted-diamond-19544

09/27/2022, 2:12 PM
Hey @modern-artist-55754, thank you! How do I update the cli version from the UI? (What I mean is, when We started the ingestion process from the UI, Datahub automatically installed the dependencies. How do we control which versions are installed?)
m

modern-artist-55754

09/27/2022, 2:38 PM
You can try logging into the pod running the datahub action image and see what it is. I am not sure the action image has the latest datahub yer
g

gifted-diamond-19544

09/27/2022, 2:56 PM
m

modern-artist-55754

09/27/2022, 11:29 PM
try to run with
0.8.45
g

gifted-diamond-19544

09/28/2022, 6:28 AM
Yup, seems that it solved it. Thank you!
3 Views