handsome-football-66174
11/03/2022, 3:23 PMlively-dusk-19162
11/03/2022, 6:05 PMsilly-lock-22001
11/03/2022, 6:30 PMfierce-garage-74290
11/04/2022, 11:06 AMchilly-elephant-51826
11/04/2022, 11:37 AM2022-11-04 11:27:15.085 ERROR 1 --- [pool-7-thread-1] c.d.authorization.DataHubAuthorizer : Failed to retrieve policy urns! Skipping updating policy cache until next refresh. start: 0, count: 30
com.datahub.util.exception.ESQueryException: Search query failed:
at com.linkedin.metadata.search.elasticsearch.query.ESSearchDAO.executeAndExtract(ESSearchDAO.java:73) ~[metadata-io.jar!/:na]
at com.linkedin.metadata.search.elasticsearch.query.ESSearchDAO.search(ESSearchDAO.java:100) ~[metadata-io.jar!/:na]
at com.linkedin.metadata.search.elasticsearch.ElasticSearchService.search(ElasticSearchService.java:67) ~[metadata-io.jar!/:na]
at com.linkedin.entity.client.JavaEntityClient.search(JavaEntityClient.java:280) ~[restli-client.jar!/:na]
at com.datahub.authorization.PolicyFetcher.fetchPolicies(PolicyFetcher.java:50) ~[auth-impl.jar!/:na]
at com.datahub.authorization.PolicyFetcher.fetchPolicies(PolicyFetcher.java:42) ~[auth-impl.jar!/:na]
at com.datahub.authorization.DataHubAuthorizer$PolicyRefreshRunnable.run(DataHubAuthorizer.java:229) ~[auth-impl.jar!/:na]
witty-television-74309
11/04/2022, 2:21 PMmysterious-hamburger-65313
11/04/2022, 4:29 PMdatahub docker quickstart
step. I run into the following issue: see image below and file as stated in the cmd | I'm on a Windows computer 4 CPUs, 16GB RAM and I believe I satisfy the 2GB Swap area and 10GB disk space. thanks for any help!bland-orange-13353
11/04/2022, 4:29 PMgreen-intern-1667
11/04/2022, 4:37 PMTest Connection
button I can only see Testing you connection for several minutes
.
Any clue on that?billowy-pilot-93812
11/04/2022, 7:09 PMlittle-spring-72943
11/05/2022, 10:34 PMfew-sunset-43876
11/06/2022, 3:56 PMStarting upgrade with id NoCodeDataMigration...
Cleanup has not been requested.
Skipping Step 1/6: RemoveAspectV2TableStep...
Executing Step 2/6: GMSQualificationStep...
Completed Step 2/6: GMSQualificationStep successfully.
Executing Step 3/6: UpgradeQualificationStep...
-- V1 table does not exist
Failed to qualify upgrade candidate. Aborting the upgrade...
Step with id UpgradeQualificationStep requested an abort of the in-progress update. Aborting the upgrade...
Upgrade NoCodeDataMigration completed with result ABORTED. Exiting...
It seems that a table was missed? What should I do in this case?
Thanks!billowy-pilot-93812
11/07/2022, 3:43 AM~~~~ Execution Summary ~~~~
RUN_INGEST - {'errors': [],
'exec_id': 'e84cc000-058f-4fa0-a2e1-6ad532762012',
'infos': ['2022-11-07 03:41:59.324788 [exec_id=e84cc000-058f-4fa0-a2e1-6ad532762012] INFO: Starting execution for task with name=RUN_INGEST',
'2022-11-07 03:42:05.614780 [exec_id=e84cc000-058f-4fa0-a2e1-6ad532762012] INFO: stdout=venv setup time = 0\n'
'This version of datahub supports report-to functionality\n'
'datahub ingest run -c /tmp/datahub/ingest/e84cc000-058f-4fa0-a2e1-6ad532762012/recipe.yml --report-to '
'/tmp/datahub/ingest/e84cc000-058f-4fa0-a2e1-6ad532762012/ingestion_report.json\n'
'[2022-11-07 03:42:03,375] INFO {datahub.cli.ingest_cli:182} - DataHub CLI version: 0.9.1\n'
'[2022-11-07 03:42:03,432] INFO {datahub.ingestion.run.pipeline:175} - Sink configured successfully. DataHubRestEmitter: configured '
'to talk to <http://datahub-gms:8080>\n'
'[2022-11-07 03:42:04,848] ERROR {datahub.entrypoints:192} - \n'
'Traceback (most recent call last):\n'
' File "/tmp/datahub/ingest/venv-superset-0.9.1/lib/python3.10/site-packages/datahub/ingestion/run/pipeline.py", line 196, in __init__\n'
' self.source: Source = source_class.create(\n'
' File "/tmp/datahub/ingest/venv-superset-0.9.1/lib/python3.10/site-packages/datahub/ingestion/source/superset.py", line 168, in create\n'
' return cls(ctx, config)\n'
' File "/tmp/datahub/ingest/venv-superset-0.9.1/lib/python3.10/site-packages/datahub/ingestion/source/superset.py", line 148, in '
'__init__\n'
' self.access_token = login_response.json()["access_token"]\n'
"KeyError: 'access_token'\n"
'\n'
'The above exception was the direct cause of the following exception:\n'
'\n'
'Traceback (most recent call last):\n'
' File "/tmp/datahub/ingest/venv-superset-0.9.1/lib/python3.10/site-packages/datahub/cli/ingest_cli.py", line 197, in run\n'
' pipeline = Pipeline.create(\n'
' File "/tmp/datahub/ingest/venv-superset-0.9.1/lib/python3.10/site-packages/datahub/ingestion/run/pipeline.py", line 317, in create\n'
' return cls(\n'
' File "/tmp/datahub/ingest/venv-superset-0.9.1/lib/python3.10/site-packages/datahub/ingestion/run/pipeline.py", line 202, in __init__\n'
' self._record_initialization_failure(\n'
' File "/tmp/datahub/ingest/venv-superset-0.9.1/lib/python3.10/site-packages/datahub/ingestion/run/pipeline.py", line 129, in '
'_record_initialization_failure\n'
' raise PipelineInitError(msg) from e\n'
'datahub.ingestion.run.pipeline.PipelineInitError: Failed to configure source (superset)\n'
'[2022-11-07 03:42:04,848] ERROR {datahub.entrypoints:195} - Command failed: \n'
'\tFailed to configure source (superset) due to \n'
"\t\t''access_token''.\n"
'\tRun with --debug to get full stacktrace.\n'
"\te.g. 'datahub --debug ingest run -c /tmp/datahub/ingest/e84cc000-058f-4fa0-a2e1-6ad532762012/recipe.yml --report-to "
"/tmp/datahub/ingest/e84cc000-058f-4fa0-a2e1-6ad532762012/ingestion_report.json'\n",
"2022-11-07 03:42:05.615068 [exec_id=e84cc000-058f-4fa0-a2e1-6ad532762012] INFO: Failed to execute 'datahub ingest'",
'2022-11-07 03:42:05.616189 [exec_id=e84cc000-058f-4fa0-a2e1-6ad532762012] INFO: Caught exception EXECUTING '
'task_id=e84cc000-058f-4fa0-a2e1-6ad532762012, name=RUN_INGEST, stacktrace=Traceback (most recent call last):\n'
' File "/usr/local/lib/python3.10/site-packages/acryl/executor/execution/default_executor.py", line 123, in execute_task\n'
' task_event_loop.run_until_complete(task_future)\n'
' File "/usr/local/lib/python3.10/asyncio/base_events.py", line 646, in run_until_complete\n'
' return future.result()\n'
' File "/usr/local/lib/python3.10/site-packages/acryl/executor/execution/sub_process_ingestion_task.py", line 227, in execute\n'
' raise TaskError("Failed to execute \'datahub ingest\'")\n'
"acryl.executor.execution.task.TaskError: Failed to execute 'datahub ingest'\n"]}
Execution finished with errors.
Hi all, I'm ingesting data from superset and got this issued. Any clue on that? Thank youcolossal-laptop-87082
11/07/2022, 6:10 AMbrave-zebra-97479
11/07/2022, 7:10 AMgray-telephone-67568
11/07/2022, 7:39 AMmicroscopic-mechanic-13766
11/07/2022, 11:45 AM{plugins_manager.py:235} ERROR - Failed to import plugin acryl-datahub-airflow-plugin
Traceback (most recent call last):
File "/home/airflow/.local/lib/python3.7/site-packages/airflow/plugins_manager.py", line 227, in load_entrypoint_plugins
plugin_class = entry_point.load()
File "/home/airflow/.local/lib/python3.7/site-packages/importlib_metadata/__init__.py", line 203, in load
module = import_module(match.group('module'))
File "/usr/local/lib/python3.7/importlib/__init__.py", line 127, in import_module
return _bootstrap._gcd_import(name[level:], package, level)
File "<frozen importlib._bootstrap>", line 1006, in _gcd_import
File "<frozen importlib._bootstrap>", line 983, in _find_and_load
File "<frozen importlib._bootstrap>", line 967, in _find_and_load_unlocked
File "<frozen importlib._bootstrap>", line 677, in _load_unlocked
File "<frozen importlib._bootstrap_external>", line 728, in exec_module
File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed
File "/home/airflow/.local/lib/python3.7/site-packages/datahub_airflow_plugin/datahub_plugin.py", line 14, in <module>
from datahub_provider.hooks.datahub import AIRFLOW_1, DatahubGenericHook
ImportError: cannot import name 'AIRFLOW_1' from 'datahub_provider.hooks.datahub' (/home/airflow/.local/lib/python3.7/site-packages/datahub_provider/hooks/datahub.py)
I don't know the source of this error as it has never shown before.
I am using the v0.8.44 version of Datahub's plugin on Airflow as with later versions I have a problem version with sqlalchemy that causes the following error:
File "/home/airflow/.local/lib/python3.7/site-packages/airflow/configuration.py", line 267, in _validate_config_dependencies
raise AirflowConfigException(f"error: cannot use sqlite with the {self.get('core', 'executor')}")
airflow.exceptions.AirflowConfigException: error: cannot use sqlite with the CeleryExecutor
Thanks in advance!!salmon-angle-92685
11/07/2022, 2:15 PMmicroscopic-mechanic-13766
11/07/2022, 3:45 PMSign in with SSO
I don't have to put my credentials again)clever-garden-23538
11/07/2022, 4:19 PMwitty-television-74309
11/07/2022, 4:20 PMbusy-eye-72759
11/07/2022, 4:21 PMswift-farmer-36942
11/07/2022, 6:22 PMswift-farmer-36942
11/07/2022, 6:24 PMswift-farmer-36942
11/07/2022, 6:24 PMwitty-television-74309
11/07/2022, 7:32 PMhandsome-football-66174
11/07/2022, 8:26 PM{search(input: {type:DATASET,start: 0, count: 1000, query: "*", filters: [{field:"tags",value:"urn:li:tag:testtag"}]} ) {searchResults {entity {urn ... on Dataset{urn status{ removed } editableProperties{ description } schemaMetadata{ fields{description type fieldPath } } domain{domain{ properties{name } } } } } } } }
salmon-jackal-36326
11/07/2022, 9:16 PM[application-akka.actor.default-dispatcher-25] WARN application - The submitted callback is unsupported!
ERROR controllers.AuthenticationController - Caught exception while attempting to redirect to SSO identity provider! It's likely that SSO integration is mis-configured.
org.pac4j.core.exception.TechnicalException: com.nimbusds.oauth2.sdk.ParseException: The scope must include an "openid" value
able-evening-90828
11/08/2022, 1:53 AMEntity
in SearchQueryResponse
of a GraphQL search request?
I got the following exception when trying to turn the code below it. I used the auto generated graphql client code and I confirmed that the raw response string the client received is correct.
com.fasterxml.jackson.databind.exc.InvalidDefinitionException: Cannot construct instance of com.linkedin.datahub.graphql.generated.Entity (no Creators, like default constructor, exist): abstract types either need to be mapped to concrete types, have custom deserializer, or contain additional type information
final SearchInput searchInput =
SearchInput.builder()
.setType(EntityType.DATASET)
.setQuery("")
.setStart(0)
.setCount(10)
.build();
final SearchQueryRequest searchQueryRequest =
SearchQueryRequest.builder().setInput(searchInput).build();
GraphQLResponseProjection graphQLResponseProjection =
new SearchResultsResponseProjection()
.start()
.count()
.total()
.searchResults(
new SearchResultResponseProjection()
.entity(new EntityResponseProjection().urn().type()));
GraphQLRequest graphQLRequest =
new GraphQLRequest(searchQueryRequest, graphQLResponseProjection);
final SearchQueryResponse searchQueryResponse =
getRestTemplate()
.exchange(
URI.create(GRAPHQL_ENDPOINT),
<http://HttpMethod.POST|HttpMethod.POST>,
createHttpEntity(graphQLRequest),
SearchQueryResponse.class)
.getBody();
System.out.println(searchQueryResponse.search());
@green-football-43791 @bulky-soccer-26729billowy-pilot-93812
11/08/2022, 4:25 AM"'container-urn:li:container:f4f87664cea5ae66d80ae56c7893eef5-to-urn:li:dataset:(urn:li:dataPlatform:postgres,tech.tech_sch.ga2_metrics_hour,PROD)',\n"
" 'tech.tech_sch.landers_category',\n"
" 'tech.tech_sch.search_query_category_rnk1-subtypes',\n"
" '... sampled of 2271 total elements'],\n"
' \'warnings\': {\'tech.location_tree._locations_vn_geo\': ["unable to map type Geometry(from_text=\'ST_GeomFromEWKT\', '
'name=\'geometry\') to metadata schema"],\n'
' \'tech.location_tree.locations_vn\': ["unable to map type Geometry(from_text=\'ST_GeomFromEWKT\', name=\'geometry\') to '
'metadata schema"],\n'
' \'tech.location_tree.locations_ph\': ["unable to map type Geometry(from_text=\'ST_GeomFromEWKT\', name=\'geometry\') to '
'metadata schema"]},\n'
" 'failures': {'Stateful Ingestion': ['Fail safe mode triggered, entity difference percent:97.61904761904762 > "
"fail_safe_threshold:{self.stateful_ingestion_config.fail_safe_threshold}']},\n"
" 'soft_deleted_stale_entities': [],\n"
" 'tables_scanned': '656',\n"