Hi everyone, I need help ingesting dbt in my Data...
# ingestion
b
Hi everyone, I need help ingesting dbt in my Datahub. My team deployed it in Kubernetes (GCP GKE) and now I don’t know where I should store manifest.json and catalog.json. I’ve tried in a bucket on Google Cloud Storage and passed the URI as an argument, but it didn’t work. Now I am wondering if I should have stored them in a bucket inside GKE (If there is one, sorry I know almost nothing about Kubernetes yet).
I see another person had the same problem in the past, but couldn’t find a solution: https://datahubspace.slack.com/archives/CV2KB471C/p1656928845742369
m
@best-sunset-26241 we don’t yet have support for gcs as a source for the dbt artifact files, this would be a very welcome contribution to the source if you are interested!
m
@best-sunset-26241 maybe you can mount the gcs bucket to file system of the pod running the ingestion using something like FUSE
b
Hi, @mammoth-bear-12532, thanks for your answer. Is this fully operational in other clouds (Heroku, AWS)?
@modern-artist-55754, is this FUSE you are referring to?
m
@best-sunset-26241 https://github.com/GoogleCloudPlatform/gcsfuse this is the one
You might need to update the action image
b
Thank you, @modern-artist-55754, I’ll try this approach.
m
Thanks @modern-artist-55754 this looks like a promising approach until we add native gcs support
b
@mammoth-bear-12532, if the cloud were AWS, this won’t be a problem, am I correct?
m
correct
more importantly, the files just need to be on S3
so if your company has an S3 bucket where you could drop your dbt artifacts, it would work