calm-river-44367
10/27/2021, 8:54 AMbland-teacher-17190
10/27/2021, 1:21 PMblue-holiday-20644
10/27/2021, 1:39 PMcurved-jordan-15657
10/27/2021, 3:38 PMvictorious-dream-46349
10/27/2021, 4:27 PMterraform provider
for datahub (and yes, we will opensource once it is matured). Something like below will add dataset for bigquery with schemas defined.
resource "datahub_dataset" "test" {
platform = "bigquery"
name = "testdataset"
origin = "DEV"
owner = "dinesh"
schema_name = "test"
field {
field_path = "name"
native_datatype = "String()"
recursive = false
nullable = true
}
field {
field_path = "address"
native_datatype = "String()"
recursive = true
nullable = false
}
tags = ["tag1", "tag2"]
upstreams = [
datahub_dataset.gcs.id,
"urn:li:dataset:(urn:li:dataPlatform:gcs,testgcs,DEV)"
]
}
Question: For this terraform provider, we are using rest-api based results. Will it scale ? Should we use graphql based apis ?calm-morning-92759
10/27/2021, 4:42 PMacceptable-vr-75043
10/27/2021, 10:31 PMplatform
(snowflake or mysql), Â name
 (db.schema.table for snowflake, db.table for mysql) and fabric
which doesn't include the snowflake/mysql cluster name. (https://datahubproject.io/docs/what/urn/)agreeable-hamburger-38305
10/27/2021, 11:34 PM>> helm install datahub datahub/datahub --set-string datahub-ingestion-cron.enabled=true --dry-run
>> dependencies.go:49: Warning: Condition path 'datahub-ingestion-cron.enabled' for chart datahub-ingestion-cron returned non-bool value
rhythmic-sundown-12093
10/28/2021, 8:34 AMred-pizza-28006
10/28/2021, 9:17 AMdamp-ambulance-34232
10/28/2021, 11:26 AMred-pizza-28006
10/28/2021, 2:37 PM/.well-known/openid-configuration
but what would be the URL?square-painting-93399
10/28/2021, 4:50 PMacceptable-greece-56919
10/28/2021, 5:19 PMdamp-ambulance-34232
10/29/2021, 3:59 AMFile "/usr/local/lib/python3.6/dist-packages/datahub/entrypoints.py", line 91, in main
sys.exit(datahub(standalone_mode=False, **kwargs))
File "/usr/lib/python3/dist-packages/click/core.py", line 722, in __call__
return self.main(*args, **kwargs)
File "/usr/lib/python3/dist-packages/click/core.py", line 697, in main
rv = self.invoke(ctx)
File "/usr/lib/python3/dist-packages/click/core.py", line 1066, in invoke
return _process_result(sub_ctx.command.invoke(sub_ctx))
File "/usr/lib/python3/dist-packages/click/core.py", line 1066, in invoke
return _process_result(sub_ctx.command.invoke(sub_ctx))
File "/usr/lib/python3/dist-packages/click/core.py", line 895, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/usr/lib/python3/dist-packages/click/core.py", line 535, in invoke
return callback(*args, **kwargs)
File "/usr/local/lib/python3.6/dist-packages/datahub/cli/ingest_cli.py", line 58, in run
pipeline.run()
File "/usr/local/lib/python3.6/dist-packages/datahub/ingestion/run/pipeline.py", line 125, in run
for wu in self.source.get_workunits():
File "/usr/local/lib/python3.6/dist-packages/datahub/ingestion/source/superset.py", line 339, in get_workunits
yield from self.emit_dashboard_mces()
File "/usr/local/lib/python3.6/dist-packages/datahub/ingestion/source/superset.py", line 250, in emit_dashboard_mces
dashboard_data
File "/usr/local/lib/python3.6/dist-packages/datahub/ingestion/source/superset.py", line 213, in construct_dashboard_from_api_data
position_data = json.loads(raw_position_data)
File "/usr/lib/python3.6/json/__init__.py", line 348, in loads
'not {!r}'.format(s.__class__.__name__))
TypeError: the JSON object must be str, bytes or bytearray, not 'NoneType'
handsome-belgium-11927
10/29/2021, 8:19 AMtall-controller-60779
10/29/2021, 11:32 AMnice-planet-17111
10/29/2021, 12:51 PMSQL_Profiling
.. Does anyone know where can i find related logs or does anyone know what queries are behind it? đbetter-orange-49102
10/29/2021, 1:05 PMdamp-ambulance-34232
11/01/2021, 8:05 AMdazzling-notebook-2883
11/01/2021, 9:50 AMdamp-minister-31834
11/01/2021, 11:04 AMred-pizza-28006
11/01/2021, 3:54 PM{
"size": 112,
"totalSize": 112,
"done": true,
"queryLocator": null,
"entityTypeName": "FieldDefinition",
"records": [
{
"attributes": {
"type": "FieldDefinition",
"url": "/services/data/v53.0/tooling/sobjects/FieldDefinition/MessagingSession.Id"
},
"DataType": "Lookup()",
"Description": null
},
{
"attributes": {
"type": "FieldDefinition",
"url": "/services/data/v53.0/tooling/sobjects/FieldDefinition/MessagingSession.Owner"
},
"DataType": "Lookup(User,Group)",
"Description": null
},
{
"attributes": {
"type": "FieldDefinition",
"url": "/services/data/v53.0/tooling/sobjects/FieldDefinition/MessagingSession.IsDeleted"
},
"DataType": "Checkbox",
"Description": null
},
{
"attributes": {
"type": "FieldDefinition",
"url": "/services/data/v53.0/tooling/sobjects/FieldDefinition/MessagingSession.Name"
},
"DataType": "Auto Number",
"Description": null
}
]
}
As you can see the DataType is not similar to what we have in other languages. Would datahub be able to handle this if I manually ingest this by putting this in a ingestable file?curved-jordan-15657
11/01/2021, 5:03 PMimportant-camera-38424
11/01/2021, 7:46 PMvictorious-dream-46349
11/02/2021, 11:03 AMwitty-butcher-82399
11/02/2021, 5:31 PMdamp-minister-31834
11/03/2021, 3:26 AMdatahub ingest -c hive_to_rest.yml
. But when my data updated in hive, I need to run the command again to update the metadata in datahub? If I want continuous update once ingested, what should I do?sparse-planet-56664
11/03/2021, 8:36 AMacceptable-eye-63357
11/03/2021, 8:38 AM