agreeable-address-71270
05/05/2023, 5:03 PMa. Open the filedocker/datahub-frontend/env/docker.env
b. Add the following configuration values to the file:My question is can the env variables in
docker.env
be set as docker-compose environment variables in the datahub-frontend-react
container?
AUTH_OIDC_ENABLED=true
AUTH_OIDC_CLIENT_ID=your-client-id
AUTH_OIDC_CLIENT_SECRET=your-client-secret
AUTH_OIDC_DISCOVERY_URI=<https://your-okta-domain.com/.well-known/openid-configuration>
AUTH_OIDC_BASE_URL=your-datahub-url
AUTH_OIDC_SCOPE="openid profile email groups"
I ask this because I am running the frontend container as an ECS service.early-hydrogen-27542
05/05/2023, 8:23 PMTier 1
term to a model via mapping, then swap it out for a Tier 3
term, both terms remain on the model. We have stateful ingestion enabled, and it works at the model level (e.g. soft deletes deleted models). Column meta mapping appears to be stateful.
2. How would I update my recipe to have column level terms prefaced with a term node (e.g. PII.
)? I would like to avoid the user having to type that prefix, but unsure how to add it when using add_terms
instead of just add_term
Version: 0.10.1
Recipe config:
meta_mapping:
datahub.owner:
match: ".*"
operation: "add_owner"
config:
owner_type: "group"
contains_pii:
match: True
operation: "add_term"
config:
term: "PersonalInformation.PII"
contains_pii:
match: False
operation: "add_term"
config:
term: "<http://PersonalInformation.No|PersonalInformation.No> PII"
tier:
match: "Tier 1|Tier 2|Tier 3"
operation: "add_term"
config:
term: "Tier.{{ $match }}"
column_meta_mapping:
glossary_terms:
match: ".*"
operation: "add_terms"
config:
separator: ","
nutritious-musician-70978
05/05/2023, 8:41 PMadamant-engine-29309
05/06/2023, 5:25 AM./gradlew quickstart
getting below error.
> Task :docker:elasticsearch-setup:docker
#12 10.24 go: <http://github.com/jwilder/dockerize@v0.6.1|github.com/jwilder/dockerize@v0.6.1>: <http://github.com/jwilder/dockerize@v0.6.1|github.com/jwilder/dockerize@v0.6.1>: Get "<https://proxy.golang.org/github.com/jwilder/dockerize/@v/v0.6.1.info>": dial tcp: lookup <http://proxy.golang.org|proxy.golang.org> on 192.168.65.5:53: read udp 172.17.0.2:59270->192.168.65.5:53: i/o timeout
#12 ERROR: executor failed running [/bin/sh -c go install <http://github.com/jwilder/dockerize@$DOCKERIZE_VERSION|github.com/jwilder/dockerize@$DOCKERIZE_VERSION>]: exit code: 1
------
> [binary 5/5] RUN go install <http://github.com/jwilder/dockerize@v0.6.1|github.com/jwilder/dockerize@v0.6.1>:
#12 10.24 go: <http://github.com/jwilder/dockerize@v0.6.1|github.com/jwilder/dockerize@v0.6.1>: <http://github.com/jwilder/dockerize@v0.6.1|github.com/jwilder/dockerize@v0.6.1>: Get "<https://proxy.golang.org/github.com/jwilder/dockerize/@v/v0.6.1.info>": dial tcp: lookup <http://proxy.golang.org|proxy.golang.org> on 192.168.65.5:53: read udp 172.17.0.2:59270->192.168.65.5:53: i/o timeout
------
ERROR: failed to solve: executor failed running [/bin/sh -c go install <http://github.com/jwilder/dockerize@$DOCKERIZE_VERSION|github.com/jwilder/dockerize@$DOCKERIZE_VERSION>]: exit code: 1
> Task :docker:elasticsearch-setup:docker FAILED
> Task :docker:mysql-setup:docker FAILED
#11 10.24 go: <http://github.com/jwilder/dockerize@v0.6.1|github.com/jwilder/dockerize@v0.6.1>: <http://github.com/jwilder/dockerize@v0.6.1|github.com/jwilder/dockerize@v0.6.1>: Get "<https://proxy.golang.org/github.com/jwilder/dockerize/@v/v0.6.1.info>": dial tcp: lookup <http://proxy.golang.org|proxy.golang.org> on 192.168.65.5:53: read udp 172.17.0.2:59270->192.168.65.5:53: i/o timeout
#11 ERROR: executor failed running [/bin/sh -c go install <http://github.com/jwilder/dockerize@$DOCKERIZE_VERSION|github.com/jwilder/dockerize@$DOCKERIZE_VERSION>]: exit code: 1
------
> [binary 5/5] RUN go install <http://github.com/jwilder/dockerize@v0.6.1|github.com/jwilder/dockerize@v0.6.1>:
#11 10.24 go: <http://github.com/jwilder/dockerize@v0.6.1|github.com/jwilder/dockerize@v0.6.1>: <http://github.com/jwilder/dockerize@v0.6.1|github.com/jwilder/dockerize@v0.6.1>: Get "<https://proxy.golang.org/github.com/jwilder/dockerize/@v/v0.6.1.info>": dial tcp: lookup <http://proxy.golang.org|proxy.golang.org> on 192.168.65.5:53: read udp 172.17.0.2:59270->192.168.65.5:53: i/o timeout
------
ERROR: failed to solve: executor failed running [/bin/sh -c go install <http://github.com/jwilder/dockerize@$DOCKERIZE_VERSION|github.com/jwilder/dockerize@$DOCKERIZE_VERSION>]: exit code: 1
> Task :datahub-web-react:yarnInstall
Done in 15.09s.
FAILURE: Build completed with 2 failures.
1: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':docker:elasticsearch-setup:docker'.
> Process 'command 'docker'' finished with non-zero exit value 1
* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================
2: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':docker:mysql-setup:docker'.
> Process 'command 'docker'' finished with non-zero exit value 1
* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================
* Get more help at <https://help.gradle.org>
Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See <https://docs.gradle.org/6.9.2/userguide/command_line_interface.html#sec:command_line_warnings>
BUILD FAILED in 1m 8s
209 actionable tasks: 82 executed, 127 up-to-date
nsimadas@bcd0746626a5 datahub %
powerful-answer-39247
05/06/2023, 10:06 AMrich-policeman-92383
05/07/2023, 2:06 AM[62]: index [datahubprod_datasetindex_v2], type [_doc], id [urn%3Ali%3Adataset%3A%28urn%3Ali%3AdataPlatform%3Aoracle%2Cbtvl.item_master%2CPROD%29], message [[datahubprod_datasetindex_v2/zU2-TAicR3aYyFw_r6niQg][[datahubprod_datasetindex_v2][0]] ElasticsearchException[Elasticsearch exception [type=document_missing_exception, reason=[_doc][urn%3Ali%3Adataset%3A%28urn%3Ali%3AdataPlatform%3Aoracle%2Cbtvl.item_master%2CPROD%29]: document missing]]]
bulky-vr-54429
05/08/2023, 8:04 AM"Packages below are known to contain vulnerabilities. Please update the affected packages and resubmit the solution
Image: elasticsearch:7.17.9
Note: CVE-2022-1471
Package: org.yaml:snakeyaml
Package Type: MAVEN
Affected Version: 1.33
Fixed Version: 2.0
Image: acryldata/datahub-actions:v0.0.12
Note: CVE-2023-24538
Package: go
Package Type: GO_STDLIB
Affected Version: 1.20.2
Fixed Version: 1.20.3
Note: CVE-2021-33036
Package: org.apache.hadoop:hadoop-yarn-server-common
Package Type: MAVEN
Affected Version: 3.2.0
Fixed Version: 3.2.3
Note: CVE-2022-37865
Package: org.apache.ivy:ivy
Package Type: MAVEN
Affected Version: 2.4.0
Fixed Version: 2.5.1
Note: CVE-2022-25168
Package: org.apache.hadoop:hadoop-common
Package Type: MAVEN
Affected Version: 3.2.0
Fixed Version: 3.2.4
Note: CVE-2023-22946
Package: org.apache.spark:spark-core_2.12
Package Type: MAVEN
Affected Version: 3.0.3
Fixed Version: 3.4.0
Note: CVE-2022-26612
Package: org.apache.hadoop:hadoop-common
Package Type: MAVEN
Affected Version: 3.2.0
Fixed Version: 3.2.3
Note: CVE-2019-0204
Package: org.apache.mesos:mesos
Package Type: MAVEN
Affected Version: 1.4.0
Fixed Version: 1.4.3
Image: acryldata/datahub-kafka-setup:v0.10.2.2
Note: CVE-2022-1471
Package: org.yaml:snakeyaml
Package Type: MAVEN
Affected Version: 1.32
Fixed Version: 2.0"
1 - I understand that it uses ES 7 and does not really run the vulnerability execution path with snakeYaml etc. It matters not. this is severe CVE and this software is going to be stopped out of any audited place (such as marketplace, etc.)
2 - Is there commitment to fix anything in that list ( I know about best efforts)
3 - Trying to run datahub without ES right now. We do not really use search. When Trying to build I get the following error:
2023-05-08T11:01:24.654+0300 [DEBUG] [org.gradle.internal.operations.DefaultBuildOperationRunner] Completing Build operation 'Configure build'
2023-05-08T11:01:24.654+0300 [DEBUG] [org.gradle.internal.operations.DefaultBuildOperationRunner] Build operation 'Configure build' completed
2023-05-08T11:01:24.660+0300 [ERROR] [org.gradle.internal.buildevents.BuildExceptionReporter]
2023-05-08T11:01:24.662+0300 [ERROR] [org.gradle.internal.buildevents.BuildExceptionReporter] FAILURE: Build failed with an exception.
2023-05-08T11:01:24.663+0300 [ERROR] [org.gradle.internal.buildevents.BuildExceptionReporter]
2023-05-08T11:01:24.663+0300 [ERROR] [org.gradle.internal.buildevents.BuildExceptionReporter] * Where:
2023-05-08T11:01:24.663+0300 [ERROR] [org.gradle.internal.buildevents.BuildExceptionReporter] Build file '/Users/podoleanu/work/datahub/buildSrc/build.gradle' line: 8
2023-05-08T11:01:24.663+0300 [ERROR] [org.gradle.internal.buildevents.BuildExceptionReporter]
2023-05-08T11:01:24.663+0300 [ERROR] [org.gradle.internal.buildevents.BuildExceptionReporter] * What went wrong:
2023-05-08T11:01:24.663+0300 [ERROR] [org.gradle.internal.buildevents.BuildExceptionReporter] A problem occurred evaluating project ':buildSrc'.
2023-05-08T11:01:24.663+0300 [ERROR] [org.gradle.internal.buildevents.BuildExceptionReporter] > Could not find method compile() for arguments [io.acryl:json-schema-avro:0.1.5, build_72c16s6ya15s0l3jdky658gr3$_run_closure1$_closure2@7c1b50a8] on object of type org.gradle.api.internal.artifacts.dsl.dependencies.DefaultDependencyHandler.
2023-05-08T11:01:24.664+0300 [ERROR] [org.gradle.internal.buildevents.BuildExceptionReporter]
2023-05-08T11:01:24.664+0300 [ERROR] [org.gradle.internal.buildevents.BuildExceptionReporter] * Exception is:
2023-05-08T11:01:24.664+0300 [ERROR] [org.gradle.internal.buildevents.BuildExceptionReporter] org.gradle.api.GradleScriptException: A problem occurred evaluating project ':buildSrc'.
2023-05-08T11:01:24.664+0300 [ERROR] [org.gradle.internal.buildevents.BuildExceptionReporter] at org.gradle.groovy.scripts.internal.DefaultScriptRunnerFactory$ScriptRunnerImpl.run(DefaultScriptRunnerFactory.java:93)
2023-05-08T11:01:24.664+0300 [ERROR] [org.gradle.internal.buildevents.BuildExceptionReporter] at org.gradle.configuration.DefaultScriptPluginFactory$ScriptPluginImpl.lambda$apply$0(DefaultScriptPluginFactory.java:135)
2023-05-08T11:01:24.669+0300 [ERROR] [org.gradle.internal.buildevents.BuildExceptionReporter] at org.gradle.configuration.ProjectScriptTarget.addConfiguration(ProjectScriptTarget.java:79)
2023-05-08T11:01:24.669+0300 [ERROR] [org.gradle.internal.buildevents.BuildExceptionReporter] at org.gradle.configuration.DefaultScriptPluginFactory$ScriptPluginImpl.apply(DefaultScriptPluginFactory.java:138)
2023-05-08T11:01:24.670+0300 [ERROR] [org.gradle.internal.buildevents.BuildExceptionReporter] at org.gradle.configuration.BuildOperationScriptPlugin$1.run(BuildOperationScriptPlugin.java:65)
2023-05-08T11:01:24.670+0300 [ERROR] [org.gradle.internal.buildevents.BuildExceptionReporter] at org.gradle.internal.operations.DefaultBuildOperationRunner$1.execute(DefaultBuildOperationRunner.java:29)
2023-05-08T11:01:24.670+0300 [ERROR] [org.gradle.internal.buildevents.BuildExceptionReporter] at org.gradle.internal.operations.DefaultBuildOperationRunner$1.execute(DefaultBuildOperationRunner.java:26)
2023-05-08T11:01:24.670+0300 [ERROR] [org.gradle.internal.buildevents.BuildExceptionReporter] at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:66)
2023-05-08T11:01:24.670+0300 [ERROR] [org.gradle.internal.buildevents.BuildExceptionReporter] at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:59)
2023-05-08T11:01:24.670+0300 [ERROR] [org.gradle.internal.buildevents.BuildExceptionReporter] at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:157)
2023-05-08T11:01:24.670+0300 [ERROR] [org.gradle.internal.buildevents.BuildExceptionReporter] at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:59)
2023-05-08T11:01:24.670+0300 [ERROR] [org.gradle.internal.buildevents.BuildExceptionReporter] at org.gradle.internal.operations.DefaultBuildOperationRunner.run(DefaultBuildOperationRunner.java:47)
2023-05-08T11:01:24.670+0300 [ERROR] [org.gradle.internal.buildevents.BuildExceptionReporter] at org.gradle.internal.operations.DefaultBuildOperationExecutor.run(DefaultBuildOperationExecutor.java:68)
2023-05-08T11:01:24.670+0300 [ERROR] [org.gradle.internal.buildevents.BuildExceptionReporter] at org.gradle.configuration.BuildOperationScriptPlugin.lambda$apply$0(BuildOperationScriptPlugin.java:62)
2023-05-08T11:01:24.670+0300 [ERROR] [org.gradle.internal.buildevents.BuildExceptionReporter] at org.gradle.configuration.internal.DefaultUserCodeApplicationContext.apply(DefaultUserCodeApplicationContext.java:44)
2023-05-08T11:01:24.670+0300 [ERROR] [org.gradle.internal.buildevents.BuildExceptionReporter] at org.gradle.configuration.BuildOperationScriptPlugin.apply(BuildOperationScriptPlugin.java:62)
Can I get help/answers please? Thanks!acoustic-kite-241
05/08/2023, 8:36 AMInternal Server Error(s) while executing query
Ingest log:
[2023-05-08 08:17:12,202] INFO {datahub.cli.ingest_cli:173} - DataHub CLI version: 0.10.2.2
[2023-05-08 08:17:12,256] INFO {datahub.ingestion.run.pipeline:204} - Sink configured successfully. DataHubRestEmitter: configured to talk to <http://datahub-gms.com> with token: eyJh**********SOTI
/usr/local/lib/python3.7/dist-packages/datahub/ingestion/source/tableau.py:2271: ConfigurationWarning: projects is deprecated and will be removed in a future release. Please removeit from your config.
config = TableauConfig.parse_obj(config_dict)
[2023-05-08 08:17:12,511] WARNING {datahub.ingestion.source.tableau:342} - project_pattern is not set but projects is set. projects is deprecated, please use project_pattern instead.
[2023-05-08 08:17:12,511] INFO {datahub.ingestion.source.tableau:345} - Initializing project_pattern from projects
[2023-05-08 08:17:12,842] INFO {tableau.endpoint.auth:50} - Signed into <https://my-tableau.org> as user with id d6948785-5cc9-4c58-8d7f-675a4e4f168b
[2023-05-08 08:17:12,842] INFO {datahub.ingestion.source.tableau:616} - Authenticated to Tableau server
[2023-05-08 08:17:12,842] INFO {datahub.ingestion.run.pipeline:221} - Source configured successfully.
[2023-05-08 08:17:12,843] INFO {datahub.cli.ingest_cli:129} - Starting metadata ingestion
-[2023-05-08 08:17:12,864] INFO {datahub.ingestion.source.tableau:596} - Initializing site project registry
[2023-05-08 08:17:12,865] INFO {tableau.endpoint.projects:31} - Querying all projects on site
2023-05-08 08:17:13,188] INFO {datahub.ingestion.source.tableau:517} - project(xxxx) is not allowed as per project_pattern
2023-05-08 08:17:13,188] INFO {datahub.ingestion.source.tableau:517} - project(xxxx) is not allowed as per project_pattern
2023-05-08 08:17:13,188] INFO {datahub.ingestion.source.tableau:517} - project(xxxx) is not allowed as per project_pattern
......
[2023-05-08 08:17:13,199] INFO {datahub.ingestion.source.tableau:517} - project(Paid Search) is not allowed as per project_pattern
[2023-05-08 08:17:13,200] INFO {tableau.endpoint.datasources:84} - Querying all datasources on site
[2023-05-08 08:17:13,306] INFO {tableau.endpoint.datasources:84} - Querying all datasources on site
|[2023-05-08 08:17:13,416] INFO {tableau.endpoint.workbooks:74} - Querying all workbooks on site
[2023-05-08 08:17:13,562] INFO {tableau.endpoint.workbooks:74} - Querying all workbooks on site
\[2023-05-08 08:17:13,694] INFO {tableau.endpoint.workbooks:74} - Querying all workbooks on site
[2023-05-08 08:17:13,807] INFO {tableau.endpoint.metadata:61} - Querying Metadata API
-[2023-05-08 08:17:13,877] ERROR {datahub.ingestion.run.pipeline:409} - Caught error
Traceback (most recent call last):
File "/usr/local/lib/python3.7/dist-packages/datahub/ingestion/run/pipeline.py", line 361, in run
self.preview_workunits if self.preview_mode else None,
File "/usr/local/lib/python3.7/dist-packages/datahub/utilities/source_helpers.py", line 91, in auto_stale_entity_removal
for wu in stream:
File "/usr/local/lib/python3.7/dist-packages/datahub/utilities/source_helpers.py", line 42, in auto_status_aspect
for wu in stream:
File "/usr/local/lib/python3.7/dist-packages/datahub/ingestion/source/tableau.py", line 2305, in get_workunits_internal
yield from self.emit_workbooks()
File "/usr/local/lib/python3.7/dist-packages/datahub/ingestion/source/tableau.py", line 738, in emit_workbooks
page_size_override=self.config.workbook_page_size,
File "/usr/local/lib/python3.7/dist-packages/datahub/ingestion/source/tableau.py", line 718, in get_connection_objects
offset,
File "/usr/local/lib/python3.7/dist-packages/datahub/ingestion/source/tableau.py", line 676, in get_connection_object_page
raise RuntimeError(f"Query {connection_type} error: {errors}")
RuntimeError: Query workbooksConnection error: [{'message': 'Internal Server Error(s) while executing query', 'extensions': None, 'path': None}]
[2023-05-08 08:17:13,895] INFO {datahub.cli.ingest_cli:135} - Source (tableau) report:
{'aspects': {'container': {'containerProperties': 1, 'dataPlatformInstance': 1, 'status': 1, 'subTypes': 1}},
'entities': {'container': ['urn:li:container:c6e27b6a2acce0003bc944ba693553f5']},
'events_produced': 4,
'events_produced_per_sec': 2,
'failures': {},
'running_time': '1.36 seconds',
'soft_deleted_stale_entities': [],
'start_time': '2023-05-08 08:17:12.531954 (1.36 seconds ago)',
'warnings': {}}
[2023-05-08 08:17:13,895] INFO {datahub.cli.ingest_cli:138} - Sink (datahub-rest) report:
{'current_time': '2023-05-08 08:17:13.895275 (now)',
'failures': [],
'gms_version': 'v0.9.5',
'pending_requests': 0,
'records_written_per_second': 2,
'start_time': '2023-05-08 08:17:12.249157 (1.65 seconds ago)',
'total_duration_in_seconds': 1.65,
'total_records_written': 4,
'warnings': []}
[2023-05-08 08:17:14,269] ERROR {datahub.entrypoints:195} - Command failed: Query workbooksConnection error: [{'message': 'Internal Server Error(s) while executing query', 'extensions': None, 'path': None}]
Traceback (most recent call last):
File "/usr/local/lib/python3.7/dist-packages/datahub/entrypoints.py", line 182, in main
sys.exit(datahub(standalone_mode=False, **kwargs))
File "/usr/local/lib/python3.7/dist-packages/click/core.py", line 1130, in __call__
return self.main(*args, **kwargs)
File "/usr/local/lib/python3.7/dist-packages/click/core.py", line 1055, in main
rv = self.invoke(ctx)
File "/usr/local/lib/python3.7/dist-packages/click/core.py", line 1657, in invoke
return _process_result(sub_ctx.command.invoke(sub_ctx))
File "/usr/local/lib/python3.7/dist-packages/click/core.py", line 1657, in invoke
return _process_result(sub_ctx.command.invoke(sub_ctx))
File "/usr/local/lib/python3.7/dist-packages/click/core.py", line 1404, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/usr/local/lib/python3.7/dist-packages/click/core.py", line 760, in invoke
return __callback(*args, **kwargs)
return __callback(*args, **kwargs)
File "/usr/local/lib/python3.7/dist-packages/click/decorators.py", line 26, in new_func
return f(get_current_context(), *args, **kwargs)
File "/usr/local/lib/python3.7/dist-packages/datahub/telemetry/telemetry.py", line 379, in wrapper
raise e
File "/usr/local/lib/python3.7/dist-packages/datahub/telemetry/telemetry.py", line 334, in wrapper
res = func(*args, **kwargs)
File "/usr/local/lib/python3.7/dist-packages/datahub/utilities/memory_leak_detector.py", line 95, in wrapper
return func(ctx, *args, **kwargs)
File "/usr/local/lib/python3.7/dist-packages/datahub/cli/ingest_cli.py", line 198, in run
loop.run_until_complete(run_func_check_upgrade(pipeline))
File "/usr/lib/python3.7/asyncio/base_events.py", line 579, in run_until_complete
return future.result()
File "/usr/local/lib/python3.7/dist-packages/datahub/cli/ingest_cli.py", line 158, in run_func_check_upgrade
ret = await the_one_future
File "/usr/local/lib/python3.7/dist-packages/datahub/cli/ingest_cli.py", line 150, in run_pipeline_async
None, functools.partial(run_pipeline_to_completion, pipeline)
File "/usr/lib/python3.7/concurrent/futures/thread.py", line 57, in run
result = self.fn(*self.args, **self.kwargs)
File "/usr/local/lib/python3.7/dist-packages/datahub/cli/ingest_cli.py", line 140, in run_pipeline_to_completion
raise e
File "/usr/local/lib/python3.7/dist-packages/datahub/cli/ingest_cli.py", line 132, in run_pipeline_to_completion
pipeline.run()
File "/usr/local/lib/python3.7/dist-packages/datahub/ingestion/run/pipeline.py", line 361, in run
self.preview_workunits if self.preview_mode else None,
File "/usr/local/lib/python3.7/dist-packages/datahub/utilities/source_helpers.py", line 91, in auto_stale_entity_removal
for wu in stream:
File "/usr/local/lib/python3.7/dist-packages/datahub/utilities/source_helpers.py", line 42, in auto_status_aspect
for wu in stream:
File "/usr/local/lib/python3.7/dist-packages/datahub/ingestion/source/tableau.py", line 2305, in get_workunits_internal
yield from self.emit_workbooks()
File "/usr/local/lib/python3.7/dist-packages/datahub/ingestion/source/tableau.py", line 738, in emit_workbooks
page_size_override=self.config.workbook_page_size,
File "/usr/local/lib/python3.7/dist-packages/datahub/ingestion/source/tableau.py", line 718, in get_connection_objects
offset,
File "/usr/local/lib/python3.7/dist-packages/datahub/ingestion/source/tableau.py", line 676, in get_connection_object_page
raise RuntimeError(f"Query {connection_type} error: {errors}")
RuntimeError: Query workbooksConnection error: [{'message': 'Internal Server Error(s) while executing query', 'extensions': None, 'path': None}]
tableau ingest yaml
# tableau
source:
type: tableau
config:
connect_uri: '${TABLEAU_ADDRESS}'
# site:
platform_instance: acryl_instance
# project_pattern:
project_pattern: ["^default$", "^Project 2$", "^/Project A/Nested Project B$"]
# projects: ["^default$", "^Project 2$", "^/Project A/Nested Project B$"]
username: '${TABLEAU_USER}'
password: '${TABLEAU_PASSWD}'
page_size: 10
ingest_tags: True
ingest_owner: True
stateful_ingestion:
enabled: True
remove_stale_metadata: true
and my datahub version: 0.10.0.7 tablue version: 2022.3.1
I want to figure out why this problem occurred and how to solve it. In fact, my Tableau service went through a version upgrade and was ingesting normally before the upgrade.
Thank u very much!powerful-cat-68806
05/08/2023, 12:29 PMerror was '
'(psycopg2.errors.FeatureNotSupported) Specified types or functions (one per INFO message) not supported on Redshift tables.
From some investigation we did, the recommendation is to install psycopg2
package
Not sure itβs relevant here, because Iβm only configuring my YAML to ingest the data
Iβm able to connect the cluster from my local prompt
Pls. adviseβ¦.
cc: @modern-garden-35830 @icy-controller-68116billowy-lock-72499
05/08/2023, 1:19 PMgentle-camera-33498
05/08/2023, 3:03 PMlate-smartphone-6255
05/08/2023, 4:27 PMhelm install datahub datahub/datahub
but get this error, could anyone help
UPGRADE FAILED: pre-upgrade hooks failed: 1 error occurred:
* timed out waiting for the condition
handsome-football-66174
05/08/2023, 6:07 PM- input-datasets:
- <s3://enterprise/Atemp>
output-datasets:
- <s3://enterprise/A>
- input-datasets:
- <s3://enterprise/A>
output-datasets:
- <s3://enterprise/B>
- input-datasets:
- <s3://enterprise/B>
output-datasets:
- <hdfs://enterprise/C>
- input-datasets:
- <hdfs://enterprise/C>
output-datasets:
- <s3://enterprise/D>
- input-datasets:
- <s3://enterprise/D>
output-datasets:
- <glue://db.table1>
GMS Logs:
2023-05-08 19:04:06,204 [qtp944427387-20] INFO c.l.m.r.entity.AspectResource:166 - INGEST PROPOSAL proposal: {aspectName=upstreamLineage, entityUrn=urn:li:dataset:(urn:li:dataPlatform:s3,enterprise/newA,PROD), entityType=dataset, aspect={contentType=application/json, value=ByteString(length=181,bytes=7b227570...227d5d7d)}, changeType=UPSERT}
[enterprise, newA]
2023-05-08 19:04:06,315 [ThreadPoolTaskExecutor-1] INFO c.l.m.k.h.s.SiblingAssociationHook:104 - Urn urn:li:dataset:(urn:li:dataPlatform:s3,enterprise/newA,PROD) received by Sibling Hook.
2023-05-08 19:04:06,346 [qtp944427387-97] INFO c.l.m.r.platform.PlatformResource:61 - Emitting platform event. name: entityChangeEvent, key: entityChangeEvent-urn:li:dataset:(urn:li:dataPlatform:s3,enterprise/newA,PROD)
2023-05-08 19:04:06,422 [pool-13-thread-1] INFO c.l.m.filter.RestliLoggingFilter:55 - POST /aspects?action=ingestProposal - ingestProposal - 200 - 218ms
2023-05-08 19:04:06,427 [pool-13-thread-2] INFO c.l.m.filter.RestliLoggingFilter:55 - POST /platform?action=producePlatformEvent - producePlatformEvent - 200 - 81ms
2023-05-08 19:04:06,429 [ThreadPoolTaskExecutor-1] INFO c.l.m.k.h.s.SiblingAssociationHook:104 - Urn urn:li:dataset:(urn:li:dataPlatform:s3,enterprise/newA,PROD) received by Sibling Hook.
2023-05-08 19:04:06,449 [qtp944427387-18] INFO c.l.m.r.entity.AspectResource:166 - INGEST PROPOSAL proposal: {aspectName=upstreamLineage, entityUrn=urn:li:dataset:(urn:li:dataPlatform:s3,enterprise/newB,PROD), entityType=dataset, aspect={contentType=application/json, value=ByteString(length=178,bytes=7b227570...227d5d7d)}, changeType=UPSERT}
2023-05-08 19:04:06,452 [ThreadPoolTaskExecutor-1] INFO c.d.event.PlatformEventProcessor:47 - Consuming a Platform Event
2023-05-08 19:04:06,495 [ThreadPoolTaskExecutor-1] INFO c.l.m.k.h.s.SiblingAssociationHook:104 - Urn urn:li:dataset:(urn:li:dataPlatform:s3,enterprise/newB,PROD) received by Sibling Hook.
[enterprise, newB]
2023-05-08 19:04:06,566 [qtp944427387-140] INFO c.l.m.r.platform.PlatformResource:61 - Emitting platform event. name: entityChangeEvent, key: entityChangeEvent-urn:li:dataset:(urn:li:dataPlatform:s3,enterprise/newB,PROD)
2023-05-08 19:04:06,568 [pool-13-thread-3] INFO c.l.m.filter.RestliLoggingFilter:55 - POST /platform?action=producePlatformEvent - producePlatformEvent - 200 - 2ms
2023-05-08 19:04:06,569 [ThreadPoolTaskExecutor-1] INFO c.l.m.k.h.s.SiblingAssociationHook:104 - Urn urn:li:dataset:(urn:li:dataPlatform:s3,enterprise/newB,PROD) received by Sibling Hook.
2023-05-08 19:04:06,578 [ThreadPoolTaskExecutor-1] INFO c.d.event.PlatformEventProcessor:47 - Consuming a Platform Event
2023-05-08 19:04:06,617 [pool-13-thread-4] INFO c.l.m.filter.RestliLoggingFilter:55 - POST /aspects?action=ingestProposal - ingestProposal - 200 - 168ms
2023-05-08 19:04:07,001 [I/O dispatcher 1] INFO c.l.m.s.e.update.BulkListener:47 - Successfully fed bulk request. Number of events: 20 Took time ms: -1
2023-05-08 19:04:32,386 [qtp944427387-18] INFO c.l.m.r.entity.AspectResource:166 - INGEST PROPOSAL proposal: {aspectName=upstreamLineage, entityUrn=urn:li:dataset:(urn:li:dataPlatform:s3,enterprise/newA,PROD), entityType=dataset, aspect={contentType=application/json, value=ByteString(length=181,bytes=7b227570...227d5d7d)}, changeType=UPSERT}
2023-05-08 19:04:32,420 [pool-13-thread-5] INFO c.l.m.filter.RestliLoggingFilter:55 - POST /aspects?action=ingestProposal - ingestProposal - 200 - 34ms
2023-05-08 19:04:32,421 [ThreadPoolTaskExecutor-1] INFO c.l.m.k.h.s.SiblingAssociationHook:104 - Urn urn:li:dataset:(urn:li:dataPlatform:s3,enterprise/newA,PROD) received by Sibling Hook.
2023-05-08 19:04:32,443 [qtp944427387-97] INFO c.l.m.r.entity.AspectResource:166 - INGEST PROPOSAL proposal: {aspectName=upstreamLineage, entityUrn=urn:li:dataset:(urn:li:dataPlatform:s3,enterprise/newB,PROD), entityType=dataset, aspect={contentType=application/json, value=ByteString(length=178,bytes=7b227570...227d5d7d)}, changeType=UPSERT}
2023-05-08 19:04:32,479 [ThreadPoolTaskExecutor-1] INFO c.l.m.k.h.s.SiblingAssociationHook:104 - Urn urn:li:dataset:(urn:li:dataPlatform:s3,enterprise/newB,PROD) received by Sibling Hook.
mysterious-scooter-52411
05/08/2023, 8:56 PMeager-river-28849
05/09/2023, 6:16 AMrich-pager-68736
05/09/2023, 7:17 AMproject_pattern:
allow:
- '^Common Analytics Domain/Production$'
I did not configure the field extract_project_hierarchy
, so that's supposed to be true
by default.wide-afternoon-79955
05/09/2023, 10:26 AMDetected undefined connector <Kafka Connector Name>, which is not in the customized connector list
. The Datahub document does not explains how do I make connector ingest the properties anyway.
b. Connector which I am trying to import are Snowflake-Sink and Generic connectors.
Thank you in advance.adorable-lawyer-88494
05/09/2023, 12:32 PM./gradlew metadata-service:war:docker
I am getting error like
FAILURE: Build completed with 2 failures.
1: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':metadata-models:processResources'.
> Cannot convert URL 'entity-registry.yml:Zone.Identifier' to a file.
* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights.
==============================================================================
2: Task failed with an exception.
-----------
* What went wrong:
Execution failed for task ':metadata-service:factories:processResources'.
Can anyone help me out from this.
Thankspurple-forest-88570
05/09/2023, 4:29 PM2023-05-09 15:13:26,155 [ForkJoinPool.commonPool-worker-201] DEBUG c.l.metadata.search.SearchService - Searching Search documents entities: [dataset], input: 2nd, postFilters: null, sortCriterion: null, from: 1200, size: 100
2023-05-09 15:13:26,156 [ForkJoinPool.commonPool-worker-87] DEBUG c.l.m.s.e.ElasticSearchService - Searching FullText Search documents entityName: dataset, input: 2nd, postFilters: null, sortCriterion: null, from: 0, size: 100
2023-05-09 15:13:27,191 [ForkJoinPool.commonPool-worker-87] DEBUG c.l.m.s.e.ElasticSearchService - Searching FullText Search documents entityName: dataset, input: 2nd, postFilters: null, sortCriterion: null, from: 0, size: 100
2023-05-09 15:13:28,225 [ForkJoinPool.commonPool-worker-87] DEBUG c.l.m.s.e.ElasticSearchService - Searching FullText Search documents entityName: dataset, input: 2nd, postFilters: null, sortCriterion: null, from: 100, size: 100
2023-05-09 15:13:29,010 [ForkJoinPool.commonPool-worker-87] DEBUG c.l.m.s.e.ElasticSearchService - Searching FullText Search documents entityName: dataset, input: 2nd, postFilters: null, sortCriterion: null, from: 0, size: 100
2023-05-09 15:13:29,764 [ForkJoinPool.commonPool-worker-87] DEBUG c.l.m.s.e.ElasticSearchService - Searching FullText Search documents entityName: dataset, input: 2nd, postFilters: null, sortCriterion: null, from: 100, size: 100
2023-05-09 15:13:30,549 [ForkJoinPool.commonPool-worker-87] DEBUG c.l.m.s.e.ElasticSearchService - Searching FullText Search documents entityName: dataset, input: 2nd, postFilters: null, sortCriterion: null, from: 200, size: 100
2023-05-09 15:13:31,417 [ForkJoinPool.commonPool-worker-87] DEBUG c.l.m.s.e.ElasticSearchService - Searching FullText Search documents entityName: dataset, input: 2nd, postFilters: null, sortCriterion: null, from: 0, size: 100
2023-05-09 15:13:32,378 [ForkJoinPool.commonPool-worker-87] DEBUG c.l.m.s.e.ElasticSearchService - Searching FullText Search documents entityName: dataset, input: 2nd, postFilters: null, sortCriterion: null, from: 100, size: 100
2023-05-09 15:13:33,462 [ForkJoinPool.commonPool-worker-87] DEBUG c.l.m.s.e.ElasticSearchService - Searching FullText Search documents entityName: dataset, input: 2nd, postFilters: null, sortCriterion: null, from: 200, size: 100
2023-05-09 15:13:34,105 [ForkJoinPool.commonPool-worker-87] DEBUG c.l.m.s.e.ElasticSearchService - Searching FullText Search documents entityName: dataset, input: 2nd, postFilters: null, sortCriterion: null, from: 300, size: 100
---------------<snip>---------------
2023-05-09 15:13:58,505 [ForkJoinPool.commonPool-worker-87] DEBUG c.l.m.s.e.ElasticSearchService - Searching FullText Search documents entityName: dataset, input: 2nd, postFilters: null, sortCriterion: null, from: 1100, size: 100
2023-05-09 15:13:58,586 [ForkJoinPool.commonPool-worker-87] DEBUG c.l.m.s.e.ElasticSearchService - Searching FullText Search documents entityName: dataset, input: 2nd, postFilters: null, sortCriterion: null, from: 0, size: 100
2023-05-09 15:13:59,617 [ForkJoinPool.commonPool-worker-87] DEBUG c.l.m.s.e.ElasticSearchService - Searching FullText Search documents entityName: dataset, input: 2nd, postFilters: null, sortCriterion: null, from: 100, size: 100
2023-05-09 15:14:00,402 [ForkJoinPool.commonPool-worker-87] DEBUG c.l.m.s.e.ElasticSearchService - Searching FullText Search documents entityName: dataset, input: 2nd, postFilters: null, sortCriterion: null, from: 200, size: 100
2023-05-09 15:14:01,266 [ForkJoinPool.commonPool-worker-87] DEBUG c.l.m.s.e.ElasticSearchService - Searching FullText Search documents entityName: dataset, input: 2nd, postFilters: null, sortCriterion: null, from: 300, size: 100
2023-05-09 15:14:01,430 [ForkJoinPool.commonPool-worker-87] DEBUG c.l.m.s.e.ElasticSearchService - Searching FullText Search documents entityName: dataset, input: 2nd, postFilters: null, sortCriterion: null, from: 400, size: 100
2023-05-09 15:14:01,515 [ForkJoinPool.commonPool-worker-87] DEBUG c.l.m.s.e.ElasticSearchService - Searching FullText Search documents entityName: dataset, input: 2nd, postFilters: null, sortCriterion: null, from: 500, size: 100
2023-05-09 15:14:01,585 [ForkJoinPool.commonPool-worker-87] DEBUG c.l.m.s.e.ElasticSearchService - Searching FullText Search documents entityName: dataset, input: 2nd, postFilters: null, sortCriterion: null, from: 600, size: 100
2023-05-09 15:14:01,664 [ForkJoinPool.commonPool-worker-87] DEBUG c.l.m.s.e.ElasticSearchService - Searching FullText Search documents entityName: dataset, input: 2nd, postFilters: null, sortCriterion: null, from: 700, size: 100
2023-05-09 15:14:01,736 [ForkJoinPool.commonPool-worker-87] DEBUG c.l.m.s.e.ElasticSearchService - Searching FullText Search documents entityName: dataset, input: 2nd, postFilters: null, sortCriterion: null, from: 800, size: 100
2023-05-09 15:14:01,814 [ForkJoinPool.commonPool-worker-87] DEBUG c.l.m.s.e.ElasticSearchService - Searching FullText Search documents entityName: dataset, input: 2nd, postFilters: null, sortCriterion: null, from: 900, size: 100
2023-05-09 15:14:01,883 [ForkJoinPool.commonPool-worker-87] DEBUG c.l.m.s.e.ElasticSearchService - Searching FullText Search documents entityName: dataset, input: 2nd, postFilters: null, sortCriterion: null, from: 1000, size: 100
2023-05-09 15:14:01,973 [ForkJoinPool.commonPool-worker-87] DEBUG c.l.m.s.e.ElasticSearchService - Searching FullText Search documents entityName: dataset, input: 2nd, postFilters: null, sortCriterion: null, from: 1100, size: 100
2023-05-09 15:14:02,056 [ForkJoinPool.commonPool-worker-87] DEBUG c.l.m.s.e.ElasticSearchService - Searching FullText Search documents entityName: dataset, input: 2nd, postFilters: null, sortCriterion: null, from: 1200, size: 100
GraphQL parameter
{
"input": {
"types": [
"DATASET"
],
"query": "2nd",
"start": 1200,
"count": 100,
"orFilters": []
}
}
GraphQL query
query getSearch($input:SearchAcrossEntitiesInput!){
searchAcrossEntities(input:$input){
total
count
searchResults{
entity{
urn
type
... on Dataset{
urn
name
}
}
matchedFields {
name
}
}
}
}
bland-orange-13353
05/09/2023, 6:55 PMrich-state-73859
05/09/2023, 7:37 PMPAC4J_SESSIONSTORE_PROVIDER=PlayCacheSessionStore
.
To reproduce, add PAC4J_SESSIONSTORE_PROVIDER=PlayCacheSessionStore
to docker-compose.yaml
and run quickstart.
Oops, cannot start the server.
com.google.inject.CreationException: Unable to create injector, see the following errors:
1) No implementation for play.cache.SyncCacheApi was bound.
at auth.AuthModule.configure(AuthModule.java:81) (via modules: com.google.inject.util.Modules$OverrideModule -> auth.AuthModule)
1 error
at com.google.inject.internal.Errors.throwCreationExceptionIfErrorsExist(Errors.java:554)
at com.google.inject.internal.InternalInjectorCreator.initializeStatically(InternalInjectorCreator.java:161)
at com.google.inject.internal.InternalInjectorCreator.build(InternalInjectorCreator.java:108)
at com.google.inject.Guice.createInjector(Guice.java:87)
at com.google.inject.Guice.createInjector(Guice.java:78)
at play.api.inject.guice.GuiceBuilder.injector(GuiceInjectorBuilder.scala:200)
at play.inject.guice.GuiceBuilder.injector(GuiceBuilder.java:211)
at play.inject.guice.GuiceApplicationBuilder.build(GuiceApplicationBuilder.java:121)
at play.inject.guice.GuiceApplicationLoader.load(GuiceApplicationLoader.java:32)
at play.api.ApplicationLoader$JavaApplicationLoaderAdapter$1.load(ApplicationLoader.scala:181)
at play.core.server.ProdServerStart$.start(ProdServerStart.scala:53)
at play.core.server.ProdServerStart$.main(ProdServerStart.scala:29)
at play.core.server.ProdServerStart.main(ProdServerStart.scala)
rapid-crowd-46218
05/10/2023, 6:48 AM"stacktrace": ["org.elasticsearch.bootstrap.StartupException: java.lang.IllegalStateException: cannot downgrade a node from version [7.17.3] to version [7.17.0]",
"at org.elasticsearch.bootstrap.Elasticsearch.init(Elasticsearch.java:170) ~[elasticsearch-7.17.0.jar:7.17.0]",
"at org.elasticsearch.bootstrap.Elasticsearch.execute(Elasticsearch.java:157) ~[elasticsearch-7.17.0.jar:7.17.0]",
"at org.elasticsearch.cli.EnvironmentAwareCommand.execute(EnvironmentAwareCommand.java:77) ~[elasticsearch-7.17.0.jar:7.17.0]",
"at org.elasticsearch.cli.Command.mainWithoutErrorHandling(Command.java:112) ~[elasticsearch-cli-7.17.0.jar:7.17.0]",
"at org.elasticsearch.cli.Command.main(Command.java:77) ~[elasticsearch-cli-7.17.0.jar:7.17.0]",
"at org.elasticsearch.bootstrap.Elasticsearch.main(Elasticsearch.java:122) ~[elasticsearch-7.17.0.jar:7.17.0]",
"at org.elasticsearch.bootstrap.Elasticsearch.main(Elasticsearch.java:80) ~[elasticsearch-7.17.0.jar:7.17.0]",
"Caused by: java.lang.IllegalStateException: cannot downgrade a node from version [7.17.3] to version [7.17.0]",
"at org.elasticsearch.env.NodeMetadata.upgradeToCurrentVersion(NodeMetadata.java:95) ~[elasticsearch-7.17.0.jar:7.17.0]",
"at org.elasticsearch.env.NodeEnvironment.loadNodeMetadata(NodeEnvironment.java:484) ~[elasticsearch-7.17.0.jar:7.17.0]",
"at org.elasticsearch.env.NodeEnvironment.<init>(NodeEnvironment.java:356) ~[elasticsearch-7.17.0.jar:7.17.0]",
"at org.elasticsearch.node.Node.<init>(Node.java:429) ~[elasticsearch-7.17.0.jar:7.17.0]",
"at org.elasticsearch.node.Node.<init>(Node.java:309) ~[elasticsearch-7.17.0.jar:7.17.0]",
"at org.elasticsearch.bootstrap.Bootstrap$5.<init>(Bootstrap.java:234) ~[elasticsearch-7.17.0.jar:7.17.0]",
"at org.elasticsearch.bootstrap.Bootstrap.setup(Bootstrap.java:234) ~[elasticsearch-7.17.0.jar:7.17.0]",
"at org.elasticsearch.bootstrap.Bootstrap.init(Bootstrap.java:434) ~[elasticsearch-7.17.0.jar:7.17.0]",
"at org.elasticsearch.bootstrap.Elasticsearch.init(Elasticsearch.java:166) ~[elasticsearch-7.17.0.jar:7.17.0]",
uncaught exception in thread [main]
"... 6 more"] }
java.lang.IllegalStateException: cannot downgrade a node from version [7.17.3] to version [7.17.0]
at org.elasticsearch.env.NodeMetadata.upgradeToCurrentVersion(NodeMetadata.java:95)
at org.elasticsearch.env.NodeEnvironment.loadNodeMetadata(NodeEnvironment.java:484)
at org.elasticsearch.env.NodeEnvironment.<init>(NodeEnvironment.java:356)
at org.elasticsearch.node.Node.<init>(Node.java:429)
at org.elasticsearch.node.Node.<init>(Node.java:309)
at org.elasticsearch.bootstrap.Bootstrap$5.<init>(Bootstrap.java:234)
at org.elasticsearch.bootstrap.Bootstrap.setup(Bootstrap.java:234)
at org.elasticsearch.bootstrap.Bootstrap.init(Bootstrap.java:434)
at org.elasticsearch.bootstrap.Elasticsearch.init(Elasticsearch.java:166)
at org.elasticsearch.bootstrap.Elasticsearch.execute(Elasticsearch.java:157)
at org.elasticsearch.cli.EnvironmentAwareCommand.execute(EnvironmentAwareCommand.java:77)
at org.elasticsearch.cli.Command.mainWithoutErrorHandling(Command.java:112)
at org.elasticsearch.cli.Command.main(Command.java:77)
at org.elasticsearch.bootstrap.Elasticsearch.main(Elasticsearch.java:122)
at org.elasticsearch.bootstrap.Elasticsearch.main(Elasticsearch.java:80)
For complete error details, refer to the log at /usr/share/elasticsearch/logs/elasticsearch.log
strong-parrot-78481
05/10/2023, 12:51 PMpowerful-shampoo-81990
05/10/2023, 10:35 PMastonishing-dusk-99990
05/11/2023, 4:21 AMCrashLoopBackOff
on pods prerequisites-kafka
?
Recently I have installed datahub using helm deployment on kubernetes and after 3 days one of the pod named prerequisites-kafka
got status CrashLoopBackOff
and always restarting but the pods always falling.
Did someone know how to fix it? Currently I always redeploy again but I think itβs not the best solution so far.
Here I attach some screenshoots from UI kubernetes, Openlens and kubectl
*Iβm using version v0.10.0
*Also Iβm using default installation for kafka not customizedsilly-ability-65278
05/11/2023, 8:17 AMANTLR Tool version 4.5 used for code generation does not match the current runtime version 4.7.2ANTLR Runtime version 4.5 used for parser compilation does not match the current runtime version 4.7.2ANTLR Tool version 4.5 used for code generation does not match the current runtime version 4.7.2ANTLR Runtime version 4.5 used for parser compilation does not match the current runtime version 4.7.2
helpful-doctor-67337
05/11/2023, 8:56 AM[2023-05-10 14:02:14,659] ERROR {datahub.entrypoints:213} - Command failed: Failed to configure the source (unity-catalog): type object 'Retry' has no attribute 'DEFAULT_METHOD_WHITELIST'
Traceback (most recent call last):
File "/tmp/datahub/ingest/venv-unity-catalog-0.9.6/lib/python3.10/site-packages/datahub/ingestion/run/pipeline.py", line 114, in _add_init_error_context
yield
File "/tmp/datahub/ingest/venv-unity-catalog-0.9.6/lib/python3.10/site-packages/datahub/ingestion/run/pipeline.py", line 192, in __init__
self.source = source_class.create(
File "/tmp/datahub/ingest/venv-unity-catalog-0.9.6/lib/python3.10/site-packages/datahub/ingestion/source/unity/source.py", line 158, in create
return cls(ctx=ctx, config=config)
File "/tmp/datahub/ingest/venv-unity-catalog-0.9.6/lib/python3.10/site-packages/datahub/ingestion/source/unity/source.py", line 110, in __init__
self.unity_catalog_api_proxy = proxy.UnityCatalogApiProxy(
File "/tmp/datahub/ingest/venv-unity-catalog-0.9.6/lib/python3.10/site-packages/datahub/ingestion/source/unity/proxy.py", line 125, in __init__
ApiClient(
File "/tmp/datahub/ingest/venv-unity-catalog-0.9.6/lib/python3.10/site-packages/databricks_cli/sdk/api_client.py", line 106, in __init__
method_whitelist=set({'POST'}) | set(Retry.DEFAULT_METHOD_WHITELIST),
AttributeError: type object 'Retry' has no attribute 'DEFAULT_METHOD_WHITELIST'
A Google search revealed that urllib3 had an update which changed some things: https://stackoverflow.com/questions/76183443/azure-devops-release-pipeline-attributeerror-type-object-retry-has-no-attribu
So I guess that pinning the urllib3 requirement to a 1.x version could solve this.mysterious-table-75773
05/11/2023, 9:03 AMancient-queen-15575
05/11/2023, 2:34 PMdatahub delete --urn "urn:li:dataPlatform:awsdms_apply_exceptions"
datahub delete --urn "urn:li:dataPlatform:mongodb"
I get an error about json decoding:
json.decoder.JSONDecodeError: Expecting value: line 1 column 1 (char 0)
If I try with a filter like
datahub delete --entity_type dataset --platform mongodb
I get an error about the client not being authorised
requests.exceptions.HTTPError: 401 Client Error: Unauthorized for url: <http://3.72.60.6:8080/entities?action=search>
I am using an API key and my user is an Admin in Datahub.average-nail-72662
05/11/2023, 4:24 PM