silly-translator-73123
11/15/2021, 7:02 AMorange-flag-48535
11/15/2021, 11:24 AMagreeable-hamburger-38305
11/16/2021, 2:14 AMAsk me about
and Teams
on their profile page? Thanks!creamy-library-6587
11/16/2021, 7:02 AMhundreds-finland-95017
11/16/2021, 6:02 PMmammoth-pencil-22596
11/16/2021, 6:35 PMaloof-airline-3441
11/17/2021, 12:53 PMnutritious-bird-77396
11/17/2021, 4:30 PMswift-lion-29806
11/17/2021, 6:53 PMswift-lion-29806
11/17/2021, 6:55 PMagreeable-hamburger-38305
11/17/2021, 11:36 PMcreamy-library-6587
11/19/2021, 6:43 AMswift-lion-29806
11/19/2021, 1:55 PMdatahub docker nuke
removes the datahub instance itself. Wondering if there is any command to just clean metadata entries?nutritious-bird-77396
11/21/2021, 3:02 PMdatahub-frontend
and datahub-gms
After that we ingested metadata from a local Redshift cluster using datahub ingestion recipe (source as redshift and sink as datahub-kafka
)
I was under the impression I might have to run MCE process to pickup the messages from Kafka to push to GMS but that wasn't necessary, I could already see the messages in frontend and Datastore (postgres)..
Could someone explain this?creamy-library-6587
11/22/2021, 4:48 PMmelodic-oil-59434
11/23/2021, 1:06 PMdazzling-appointment-34954
11/23/2021, 5:30 PMboundless-scientist-520
11/24/2021, 8:25 PMbrief-cricket-98290
11/25/2021, 11:28 AMagreeable-river-32119
11/29/2021, 1:09 AMbrave-forest-5974
11/29/2021, 8:53 AM{
"error": {
"root_cause": [],
"type": "search_phase_execution_exception",
"reason": "",
"phase": "fetch",
"grouped": true,
"failed_shards": [],
"caused_by": {
"type": "too_many_buckets_exception",
"reason": "Trying to create too many buckets. Must be less than or equal to: [65535] but was [65536]. This limit can be set by changing the [search.max_buckets] cluster level setting.",
"max_buckets": 65535
}
},
"status": 503
}
found this related thread in troubleshootingchilly-analyst-45780
11/29/2021, 2:02 PMaloof-london-98698
11/29/2021, 2:51 PMrough-garage-43684
11/29/2021, 7:06 PMrich-greece-17287
11/30/2021, 4:46 AMenough-london-69397
11/30/2021, 4:46 PMid_ + 'table_name'
, so the account
table has PK id_account
. If I ingest this metadata through Datahub, can it smartly tell that id_account
is a PK? On the image below it does not show anything like that:acceptable-honey-21072
12/04/2021, 7:43 AMbig-coat-53708
12/05/2021, 12:18 PM30k
tables in our Trino/Hive metastore environment, which ends up running 3 million
queries on neo4j. It takes about 90min for every sync, I’m still exploring whether DataHub could do better than this.
There’s one thing that I would like to clarify on the push model. I understand that it provides an interface for triggering the ingestion. But for every ingestion, it still does a full pull
, right? It would still extract every single metadata from the source, am I correct?
In my understanding, whenever I get an event from metastore, I could trigger an ingestion through Kafka or REST API. The ingestion basically pulls everything from the source and dump it into the sink according to the recipe. Please correct me if I’m wrong. Thanks 🙏dazzling-appointment-34954
12/08/2021, 11:13 AMbrash-carpenter-51184
12/08/2021, 4:19 PM