nice-autumn-10105
01/12/2022, 3:13 PMmodern-monitor-81461
01/12/2022, 3:30 PMtable_name
provided by Superset, but that name is the name of the virtual dataset, which doesn't exist in DataHub. I think there should be a connection to the physical dataset like the following:
Dashboard -> Chart -> Dataset (virtual) -> Dataset (physical)
Right now, the virtual dataset refers to void, so I can't tell which physical dataset is being used for the query. And the query should be a property of the virtual dataset in my opinion. Am I missing something in my deployment, or there is actually a gap in DataHub's Superset source?shy-parrot-64120
01/12/2022, 9:05 PMkafka-connect
ingestion lib
when processing jdbc source connector set for postgres source via url jdbc:<postgresql://host>:port/db
its datasets are being ingested of source_platform=postgresql
rather than postgres
(as for postgresql ingestor)
this causes entities mismatch
is there a way to handle this?eager-gpu-17565
01/13/2022, 9:04 AMquaint-branch-37931
01/13/2022, 9:12 AMminiature-television-17996
01/13/2022, 10:00 AMmany-pilot-7340
01/13/2022, 5:10 PMgentle-florist-49869
01/13/2022, 5:22 PMbest-planet-6756
01/13/2022, 5:55 PM'db.table': ["Profiling exception (pymysql.err.OperationalError) (1046, 'No database selected')\n"
'[SQL: CREATE TEMPORARY TABLE ge_temp_b7535ba0 AS SELECT * \n'
'FROM db.table \n'
' LIMIT 100]\n'
'(Background on this error at: <http://sqlalche.me/e/13/e3q8)>']
Please note that in our ingestion recipe we do not specify a db name since there are multiple dbs at this source. When not profiling and only ingesting it works perfectly and ingests all of the dbs at the source. Looking to find a fix that will allow us to ingested with limited profiling and not have to specify the db as we want it to profile all of the dbs.important-machine-62199
01/14/2022, 5:33 AMglamorous-carpet-98686
01/14/2022, 5:50 AMmysterious-nail-70388
01/14/2022, 6:32 AMmysterious-nail-70388
01/14/2022, 6:40 AMboundless-student-48844
01/14/2022, 7:24 AMred-pizza-28006
01/14/2022, 8:19 AMmysterious-nail-70388
01/14/2022, 9:02 AMfew-air-56117
01/14/2022, 10:05 AMclever-australia-61035
01/17/2022, 7:56 AMlemon-hydrogen-83671
01/17/2022, 4:21 PMCorpUserInfo
object that's in all the examples or do we start using CorpUserProperties
as specified here: https://datahubproject.io/docs/graphql/objects/#corpuserproperties?agreeable-thailand-43234
01/17/2022, 8:18 PMdb in lake formation
, then a crawler to populate it, then plug Datahub into the s3 bucket...is this the best-cost effective way to do it? or would you rather to have an endpoint in an EC2 instance?mysterious-nail-70388
01/18/2022, 3:35 AMmysterious-nail-70388
01/18/2022, 5:51 AMbest-television-56567
01/18/2022, 8:18 AMcolossal-easter-99672
01/18/2022, 8:39 AMhandsome-belgium-11927
01/18/2022, 10:56 AMbreezy-controller-54597
01/19/2022, 4:16 AM${MSSQL_PASSWORD}
).
https://datahubproject.io/docs/metadata-ingestion#handling-sensitive-information-in-recipes
Where can I define this variable?
I created a .env file in the same directory as the recipe yml file with reference to docker-compose, but it didn't work.careful-engine-38533
01/19/2022, 7:20 AMacoustic-quill-54426
01/19/2022, 3:36 PMMLFeatures
and MLFeatureTables
. I had an issue with the MLFeatureProperties
field dataType
, which is optional, but once successfully ingested was raising exceptions here. I just added a default UNKNOWN
dataType, but I guess it should be either mandatory or nil-checked. WDYT? I can drop an issue later 👍
I have a question about MLFeatureTableProperties
aspect customProperties
field, which is not rendered in the UI. In other entities as datasets they are always rendered. Am I missing something? Many thanks 🙌damp-queen-61493
01/19/2022, 4:53 PM[2022-01-19, 16:43:56 UTC] {local_task_job.py:154} INFO - Task exited with return code Negsignal.SIGKILL
I'm using the inline approach like thisred-pizza-28006
01/19/2022, 6:10 PM