rich-state-73859
02/06/2023, 10:04 PMcom.google.protobuf.InvalidProtocolBufferException$InvalidWireTypeException: Protocol message tag had invalid wire type.
when importing scalapb
in my protobuf. I added the following file option.
option (scalapb.options) = {
import: "path.to.package._"
};
green-lion-58215
02/06/2023, 11:30 PMrefined-energy-76018
02/06/2023, 11:44 PMbitter-evening-61050
02/07/2023, 5:58 AMbitter-evening-61050
02/07/2023, 6:51 AMhigh-toothbrush-90528
02/07/2023, 9:11 AMChangeTypeClass.DELETE
instead (But I saw it's not supported yet).chilly-potato-57465
02/07/2023, 12:40 PMlively-dusk-19162
02/07/2023, 2:46 PMkind-kite-29761
02/07/2023, 4:00 PM>> datahub ingest -c S3.yml
[2023-02-07 13:12:16,262] INFO {datahub.cli.ingest_cli:165} - DataHub CLI version: 0.9.6.4
[2023-02-07 13:12:16,317] INFO {datahub.ingestion.run.pipeline:179} - Sink configured successfully. DataHubRestEmitter: configured to talk to <http://a770b2e6a6b9d4997bc43a67978e8c9f-1534401114.us-east-1.elb.amazonaws.com:9002/api/gms> with token: eyJh**********Tdmk
[2023-02-07 13:12:16,697] ERROR {logger:26} - Please set env variable SPARK_VERSION
[2023-02-07 13:12:16,697] INFO {logger:27} - Using deequ: com.amazon.deequ:deequ:1.2.2-spark-3.0
/home/ec2-user/environment/datahub/lib64/python3.7/site-packages/datahub/ingestion/source/s3/source.py:317: ConfigurationWarning: env is deprecated and will be removed in a future release. Please use platform_instance instead.
config = DataLakeSourceConfig.parse_obj(config_dict)
[2023-02-07 13:12:17,144] INFO {datahub.ingestion.run.pipeline:196} - Source configured successfully.
[2023-02-07 13:12:17,146] INFO {datahub.cli.ingest_cli:120} - Starting metadata ingestion
-[2023-02-07 13:12:17,243] INFO {botocore.credentials:1253} - Found credentials in shared credentials file: ~/.aws/credentials
|[2023-02-07 13:13:26,208] INFO {datahub.cli.ingest_cli:133} - Finished metadata ingestion
Cli report:
{'cli_entry_location': '/home/ec2-user/environment/datahub/lib64/python3.7/site-packages/datahub/__init__.py',
'cli_version': '0.9.6.4',
'mem_info': '186.32 MB',
'os_details': 'Linux-4.14.301-224.520.amzn2.x86_64-x86_64-with-glibc2.2.5',
'py_exec_path': '/home/ec2-user/environment/datahub/bin/python3',
'py_version': '3.7.16 (default, Dec 15 2022, 23:24:54) \n[GCC 7.3.1 20180712 (Red Hat 7.3.1-15)]'}
Source (s3) report:
{'aspects': {},
'entities': {},
'events_produced': 0,
'events_produced_per_sec': 0,
'failures': {},
'filtered': [],
'running_time': '1 minute and 9.41 seconds',
'start_time': '2023-02-07 13:12:16.962307 (1 minute and 9.41 seconds ago)',
'warnings': {}}
Sink (datahub-rest) report:
{'current_time': '2023-02-07 13:13:26.378011 (now)',
'failures': [],
'gms_version': 'v0.8.45',
'pending_requests': 0,
'records_written_per_second': 0,
'start_time': '2023-02-07 13:12:16.302696 (1 minute and 10.08 seconds ago)',
'total_duration_in_seconds': 70.08,
'total_records_written': 0,
'warnings': []}
Pipeline finished successfully; produced 0 events in 1 minute and 9.41 seconds.
❗Client-Server Incompatible❗ Your client version 0.9.6.4 is older than your server version 0.8.45. Upgrading the cli to 0.8.45 is recommended.
Any idea, where I am doing wrong. Why I am not able to push it from S3 ?acceptable-account-83031
02/07/2023, 5:42 PMemit_s3_lineage: True
with ingestion source Glue.quiet-jelly-11365
02/07/2023, 5:51 PMalert-fall-82501
02/07/2023, 6:15 PMbland-barista-59197
02/07/2023, 8:58 PMnice-advantage-52080
02/07/2023, 9:50 PMbreezy-controller-54597
02/08/2023, 2:40 AMflat-painter-78331
02/08/2023, 7:46 AMincalculable-manchester-41314
02/08/2023, 9:39 AMripe-eye-60209
02/08/2023, 1:20 PMlemon-scooter-69730
02/08/2023, 1:41 PMelegant-salesmen-99143
02/08/2023, 6:06 PMschema_pattern:
allow:
to the transformer, but it didn't work, the ingest failed with an error in that part of recipe. Or maybe I messed up the nesting depth, where should it be?elegant-salesmen-99143
02/08/2023, 6:10 PMclean-tomato-22549
02/09/2023, 3:09 AMcold-airport-17919
02/09/2023, 4:10 AMclean-tomato-22549
02/09/2023, 6:04 AMthousands-bird-50049
02/09/2023, 6:48 AMlate-bear-87552
02/09/2023, 6:59 AMTraceback (most recent call last):
File "/home/airflow/.local/lib/python3.8/site-packages/datahub/ingestion/run/pipeline.py", line 114, in _add_init_error_context
yield
File "/home/airflow/.local/lib/python3.8/site-packages/datahub/ingestion/run/pipeline.py", line 192, in __init__
self.source = source_class.create(
File "/home/airflow/.local/lib/python3.8/site-packages/datahub/ingestion/source/bigquery_v2/bigquery.py", line 262, in create
return cls(ctx, config)
File "/home/airflow/.local/lib/python3.8/site-packages/datahub/ingestion/source/bigquery_v2/bigquery.py", line 199, in __init__
super(BigqueryV2Source, self).__init__(config, ctx)
File "/home/airflow/.local/lib/python3.8/site-packages/datahub/ingestion/source/state/stateful_ingestion_base.py", line 180, in __init__
self._initialize_checkpointing_state_provider()
File "/home/airflow/.local/lib/python3.8/site-packages/datahub/ingestion/source/state/stateful_ingestion_base.py", line 223, in _initialize_checkpointing_state_provider
checkpointing_state_provider_class.create(
File "/home/airflow/.local/lib/python3.8/site-packages/datahub/ingestion/source/state_provider/datahub_ingestion_checkpointing_provider.py", line 50, in create
graph = DataHubGraph(provider_config.datahub_api)
File "/home/airflow/.local/lib/python3.8/site-packages/datahub/ingestion/graph/client.py", line 72, in __init__
self.test_connection()
File "/home/airflow/.local/lib/python3.8/site-packages/datahub/emitter/rest_emitter.py", line 146, in test_connection
response = self._session.get(f"{self._gms_server}/config")
File "/home/airflow/.local/lib/python3.8/site-packages/requests/sessions.py", line 600, in get
return self.request("GET", url, **kwargs)
File "/home/airflow/.local/lib/python3.8/site-packages/requests/sessions.py", line 587, in request
resp = self.send(prep, **send_kwargs)
File "/home/airflow/.local/lib/python3.8/site-packages/requests/sessions.py", line 701, in send
r = adapter.send(request, **kwargs)
File "/home/airflow/.local/lib/python3.8/site-packages/requests/adapters.py", line 489, in send
resp = conn.urlopen(
File "/home/airflow/.local/lib/python3.8/site-packages/urllib3/connectionpool.py", line 703, in urlopen
httplib_response = self._make_request(
File "/home/airflow/.local/lib/python3.8/site-packages/urllib3/connectionpool.py", line 398, in _make_request
conn.request(method, url, **httplib_request_kw)
File "/home/airflow/.local/lib/python3.8/site-packages/urllib3/connection.py", line 239, in request
super(HTTPConnection, self).request(method, url, body=body, headers=headers)
File "/usr/local/lib/python3.8/http/client.py", line 1256, in request
self._send_request(method, url, body, headers, encode_chunked)
File "/usr/local/lib/python3.8/http/client.py", line 1297, in _send_request
self.putheader(hdr, value)
File "/home/airflow/.local/lib/python3.8/site-packages/urllib3/connection.py", line 224, in putheader
_HTTPConnection.putheader(self, header, *values)
File "/usr/local/lib/python3.8/http/client.py", line 1234, in putheader
raise ValueError('Invalid header value %r' % (values[i],))
ValueError: Invalid header value b'Bearer ***********\n'
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/home/airflow/.local/lib/python3.8/site-packages/airflow/operators/python.py", line 175, in execute
return_value = self.execute_callable()
File "/home/airflow/.local/lib/python3.8/site-packages/airflow/operators/python.py", line 192, in execute_callable
return self.python_callable(*self.op_args, **self.op_kwargs)
File "/opt/airflow/dags/repo/org/groww/dataplatform/datahub/bigquery/DAG_DATAHUB_BIGQUERY_META_WITH_DENY_DATASETS.py", line 85, in ingest_bigquery_metadata
pipeline = Pipeline.create(complete_json)
File "/home/airflow/.local/lib/python3.8/site-packages/datahub/ingestion/run/pipeline.py", line 303, in create
return cls(
File "/home/airflow/.local/lib/python3.8/site-packages/datahub/ingestion/run/pipeline.py", line 196, in __init__
<http://logger.info|logger.info>("Source configured successfully.")
File "/usr/local/lib/python3.8/contextlib.py", line 131, in __exit__
self.gen.throw(type, value, traceback)
File "/home/airflow/.local/lib/python3.8/site-packages/datahub/ingestion/run/pipeline.py", line 116, in _add_init_error_context
raise PipelineInitError(f"Failed to {step}: {e}") from e
limited-forest-73733
02/09/2023, 7:27 AMwhite-napkin-20729
02/09/2023, 9:23 AMelegant-salesmen-99143
02/09/2023, 9:39 AMpurple-printer-15193
02/09/2023, 9:54 AM