witty-keyboard-20400
11/03/2021, 12:12 PMconnect_uri: "<mongodb://dbuser1>:xyz123#$$!us@192.168.1.100:27017/mongodbName?authSource=mongodbName"
The above mentioned username and password are correct, which I verified using Robo 3T client.
When I try to ingest the metadata using the command
datahub ingest -c ./mongodb.yml
It fails with the following auth error:
all_credentials = {'mongodbName': MongoCredential ('SCRAM-SHA-1', 'mongodbName', 'dbuser1', 'xyz123#14911!us', None, <pymongo.auth._Cache object at 0x7ff8d7b91d00>, )
...
credentials = MongoCredential ('SCRAM-SHA-1', 'mongodbName', 'dbuser1', 'xyz123#14911!us', None, <pymongo.auth._Cache object at 0x7ff8d7b91d00>, )
...
...
OperationFailure: Authentication failed., full error: {'ok': 0.0, 'errmsg': 'Authentication failed.', 'code': 18, 'codeName': 'AuthenticationFailed', 'operationTime': Timestamp(1635941323, 2), '$clusterTime': {'clusterTime': Timestamp(1635941323, 2), 'signature': {'hash': b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', 'keyId': 0}}}
Notice that the password provided by me in the yml file is:
xyz123#$$!us
..while the password in the log statements are shown as:
xyz123#14911!us
Why is this happening?
Is this a bug? Or am I missing something here? @big-carpet-38439 @miniature-tiger-96062careful-insurance-60247
11/03/2021, 6:32 PMcareful-insurance-60247
11/03/2021, 7:35 PMdamp-minister-31834
11/04/2021, 1:40 AMtableA
is the upstream of tableB
. When I ingest hive resource to datahub, the two tables are not familiar with each other. Is there a way to get the lineage automatically while ingestging without calling lineage_emitter_kafka.py
or lineage_emitter_rest.py
artificially.damp-minister-31834
11/04/2021, 6:18 AMorange-flag-48535
11/04/2021, 6:23 AMorange-flag-48535
11/04/2021, 11:05 AMplain-farmer-27314
11/04/2021, 5:07 PM"master not discovered yet, this node has not previously joined a bootstrapped (v7+) cluster, and this node must discover master-eligible nodes [elasticsearch-master-0, elasticsearch-master-1, elasticsearch-master-2] to bootstrap a cluster: have discovered
Did I miss a step here?echoing-dress-35614
11/04/2021, 6:19 PMdamp-ambulance-34232
11/05/2021, 2:18 AMabundant-flag-19546
11/05/2021, 6:01 AMpolite-flower-25924
11/05/2021, 7:52 AMorange-flag-48535
11/05/2021, 9:36 AMorange-flag-48535
11/05/2021, 10:01 AMrhythmic-sundown-12093
11/08/2021, 3:16 AMeager-answer-71364
11/08/2021, 6:56 AMrough-zoo-50278
11/08/2021, 7:25 AM07:21:46.830 [qtp544724190-12] ERROR c.l.m.filter.RestliLoggingFilter - java.lang.RuntimeException: java.lang.reflect.InvocationTargetException
07:21:47.228 [qtp544724190-10] INFO c.l.m.filter.RestliLoggingFilter - POST /entities?action=ingest - ingest - 500 - 1ms
is there any additional config needed to make ingestion work?
EDIT: Seems to fail here:
: Invalid URN Parameter: 'No enum constant com.linkedin.common.FabricType.dev: urn:li:dataset:(urn:li:dataPlatform:postgres,...,...)
EDIT: It was wrong env (which didn't exist)damp-ambulance-34232
11/08/2021, 9:04 AMfreezing-teacher-87574
11/08/2021, 10:33 AMfeast
module.
source:
type: feast
config:
core_url: localhost:6565 # default
env: "PROD" # Optional, default is "PROD"
use_local_build: False # Whether to build Feast ingestion image locally, default is False
Thanksdamp-ambulance-34232
11/08/2021, 10:40 AMlittle-france-72098
11/08/2021, 10:43 AMFile "/usr/local/lib/python3.10/site-packages/datahub/ingestion/run/pipeline.py", line 141, in run
for wu in self.source.get_workunits():
File "/usr/local/lib/python3.10/site-packages/datahub/ingestion/source/kafka.py", line 84, in get_workunits
mce = self._extract_record(t)
File "/usr/local/lib/python3.10/site-packages/datahub/ingestion/source/kafka.py", line 115, in _extract_record
fields = schema_util.avro_schema_to_mce_fields(schema.schema_str)
File "/usr/local/lib/python3.10/site-packages/datahub/ingestion/extractor/schema_util.py", line 443, in avro_schema_to_mce_fields
return list(
File "/usr/local/lib/python3.10/site-packages/datahub/ingestion/extractor/schema_util.py", line 427, in to_mce_fields
yield from converter._to_mce_fields(avro_schema)
File "/usr/local/lib/python3.10/site-packages/datahub/ingestion/extractor/schema_util.py", line 408, in _to_mce_fields
yield from self._avro_type_to_mce_converter_map[type(avro_schema)](avro_schema)
File "/usr/local/lib/python3.10/site-packages/datahub/ingestion/extractor/schema_util.py", line 393, in _gen_from_non_field_nested_schemas
yield from self._to_mce_fields(sub_schema)
File "/usr/local/lib/python3.10/site-packages/datahub/ingestion/extractor/schema_util.py", line 408, in _to_mce_fields
yield from self._avro_type_to_mce_converter_map[type(avro_schema)](avro_schema)
File "/usr/local/lib/python3.10/site-packages/datahub/ingestion/extractor/schema_util.py", line 328, in _gen_nested_schema_from_field
yield from self._to_mce_fields(sub_schema)
File "/usr/local/lib/python3.10/site-packages/datahub/ingestion/extractor/schema_util.py", line 408, in _to_mce_fields
yield from self._avro_type_to_mce_converter_map[type(avro_schema)](avro_schema)
KeyError: <class 'avro.schema.UUIDSchema'>
The schemas in question indeed have fields with the logicalType uuid and this case doesn't seem to be handled in schema_util.brief-lizard-77958
11/08/2021, 3:05 PMagreeable-hamburger-38305
11/09/2021, 1:28 AMQueries
tab. Since the default timeframe for ingesting usage (I am working with BigQuery) is the past day, if the database is not frequently used, all the Queries
tab has is the ingestion queries, which is not very useful. I was wondering if there is a way to modify the default timeframe (instead of hard-coding `start_time`and end_time
) to like 5 days. Also wondering if there is a way to ignore the queries ran by the ingestion joborange-flag-48535
11/09/2021, 10:27 AMsquare-activity-64562
11/09/2021, 4:53 PMdazzling-appointment-34954
11/10/2021, 10:07 AMorange-flag-48535
11/10/2021, 11:11 AMorange-flag-48535
11/10/2021, 11:35 AMbrave-forest-5974
11/10/2021, 3:05 PMnumerous-guitar-35145
11/10/2021, 3:09 PM