jolly-pager-18761
07/05/2023, 2:51 PMbusy-analyst-35820
07/06/2023, 3:58 AMkind-whale-9577
07/06/2023, 4:19 AMkind-whale-9577
07/06/2023, 4:19 AMkind-whale-9577
07/06/2023, 4:19 AM23/07/06 04:16:52 ERROR DatahubSparkListener: java.lang.NullPointerException
at datahub.spark.DatahubSparkListener.processExecution(DatahubSparkListener.java:296)
at datahub.spark.DatahubSparkListener.onOtherEvent(DatahubSparkListener.java:241)
at org.apache.spark.scheduler.SparkListenerBus.doPostEvent(SparkListenerBus.scala:100)
kind-whale-9577
07/06/2023, 4:19 AMkind-whale-9577
07/06/2023, 4:40 AM23/07/06 04:40:11 ERROR DatahubSparkListener: java.lang.NullPointerException
at datahub.spark.DatahubSparkListener$3.apply(DatahubSparkListener.java:262)
at datahub.spark.DatahubSparkListener$3.apply(DatahubSparkListener.java:258)
worried-rocket-84695
07/06/2023, 5:31 AMadventurous-lighter-8945
07/06/2023, 6:20 AMpowerful-tent-14193
07/06/2023, 8:02 AMpowerful-tent-14193
07/06/2023, 8:09 AMacceptable-computer-51491
07/06/2023, 9:55 AM[2023-07-06 09:16:10,979] DEBUG {datahub.emitter.rest_emitter:247} - Attempting to emit to DataHub GMS; using curl equivalent to:\n',
'2023-07-06 09:16:11.149010 [exec_id=280a9dbb-5208-4212-95ee-d28a9e4d4afc] INFO: Caught exception EXECUTING '
'task_id=280a9dbb-5208-4212-95ee-d28a9e4d4afc, name=RUN_INGEST, stacktrace=Traceback (most recent call last):\n'
' File "/usr/local/lib/python3.10/asyncio/streams.py", line 525, in readline\n'
' line = await self.readuntil(sep)\n'
' File "/usr/local/lib/python3.10/asyncio/streams.py", line 603, in readuntil\n'
' raise exceptions.LimitOverrunError(\n'
'asyncio.exceptions.LimitOverrunError: Separator is not found, and chunk exceed the limit\n'
'\n'
'During handling of the above exception, another exception occurred:\n'
'\n'
'Traceback (most recent call last):\n'
' File "/usr/local/lib/python3.10/site-packages/acryl/executor/execution/default_executor.py", line 123, in execute_task\n'
' task_event_loop.run_until_complete(task_future)\n'
' File "/usr/local/lib/python3.10/asyncio/base_events.py", line 646, in run_until_complete\n'
' return future.result()\n'
' File "/usr/local/lib/python3.10/site-packages/acryl/executor/execution/sub_process_ingestion_task.py", line 147, in execute\n'
' await tasks.gather(_read_output_lines(), _report_progress(), _process_waiter())\n'
' File "/usr/local/lib/python3.10/site-packages/acryl/executor/execution/sub_process_ingestion_task.py", line 99, in _read_output_lines\n'
' line_bytes = await ingest_process.stdout.readline()\n'
' File "/usr/local/lib/python3.10/asyncio/streams.py", line 534, in readline\n'
' raise ValueError(e.args[0])\n'
'ValueError: Separator is not found, and chunk exceed the limit\n']}
limited-dentist-50437
07/06/2023, 12:08 PMbroad-pager-68280
07/06/2023, 12:54 PMkind-whale-9577
07/06/2023, 4:13 PMkind-whale-9577
07/06/2023, 4:13 PM23/07/06 04:16:52 ERROR DatahubSparkListener: java.lang.NullPointerException
at datahub.spark.DatahubSparkListener.processExecution(DatahubSparkListener.java:296)
at datahub.spark.DatahubSparkListener.onOtherEvent(DatahubSparkListener.java:241)
at org.apache.spark.scheduler.SparkListenerBus.doPostEvent(SparkListenerBus.scala:100)
kind-whale-9577
07/06/2023, 4:15 PMdelightful-beard-43126
07/06/2023, 6:34 PM2023-07-06 15:20:20,856 [R2 Nio Event Loop-1-1] WARN c.l.r.t.h.c.c.ChannelPoolLifecycle:139 - Failed to create channel, remote=localhost/127.0.0.1:8080
io.netty.channel.AbstractChannel$AnnotatedConnectException: Connection refused: localhost/127.0.0.1:8080
Caused by: java.net.ConnectException: Connection refused
at java.base/sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
at java.base/sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:777)
at io.netty.channel.socket.nio.NioSocketChannel.doFinishConnect(NioSocketChannel.java:337)
at io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.finishConnect(AbstractNioChannel.java:334)
at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:776)
at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:724)
at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:650)
at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:562)
at io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997)
at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
at java.base/java.lang.Thread.run(Thread.java:829)
2023-07-06 15:20:29,258 [R2 Nio Event Loop-1-2] WARN c.l.r.t.h.c.c.ChannelPoolLifecycle:139 - Failed to create channel, remote=localhost/127.0.0.1:8080
io.netty.channel.AbstractChannel$AnnotatedConnectException: Connection refused: localhost/127.0.0.1:8080
Caused by: java.net.ConnectException: Connection refused
at java.base/sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
at java.base/sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:777)
at io.netty.channel.socket.nio.NioSocketChannel.doFinishConnect(NioSocketChannel.java:337)
at io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.finishConnect(AbstractNioChannel.java:334)
at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:776)
at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:724)
at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:650)
at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:562)
at io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997)
at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
at java.base/java.lang.Thread.run(Thread.java:829)
2023-07-06 15:20:29,259 [ThreadPoolTaskExecutor-1] ERROR c.l.m.kafka.hydrator.EntityHydrator:49 - Error while calling GMS to hydrate entity for urn urn:li:corpuser:${company_email}
2023-07-06 15:20:29,259 [ThreadPoolTaskExecutor-1] INFO c.l.m.k.t.DataHubUsageEventTransformer:128 - No matches for urn urn:li:corpuser:${company_email}
2023-07-06 15:20:29,548 [I/O dispatcher 1] ERROR c.l.m.s.e.update.BulkListener:44 - Failed to feed bulk request. Number of events: 1 Took time ms: -1 Message: failure in bulk execution:
[0]: index [datahub_usage_event-000059], type [_doc], id [PageViewEvent_urn%3Ali%3Acorpuser%3A${company_email}_1688570490840_11171], message [[datahub_usage_event-000059/v6wPC48TQBOr5UQ9q9kygw][[datahub_usage_event-000059][0]] ElasticsearchException[Elasticsearch exception [type=version_conflict_engine_exception, reason=[PageViewEvent_urn%3Ali%3Acorpuser%3A${company_email}_1688570490840_11171]: version conflict, document already exists (current version [1])]]]
2023-07-06 15:20:30,065 [R2 Nio Event Loop-1-1] WARN c.l.r.t.h.c.c.ChannelPoolLifecycle:139 - Failed to create channel, remote=localhost/127.0.0.1:8080
io.netty.channel.AbstractChannel$AnnotatedConnectException: Connection refused: localhost/127.0.0.1:8080
What could be happening here?victorious-monkey-86128
07/06/2023, 7:11 PMgradlew/quickstart
in the datahub repo but I keep getting the error:
Execution failed for task ':docker:kafka-setup:docker'.
> Process 'command 'docker'' finished with non-zero exit value 1
nice-rocket-26538
07/06/2023, 7:38 PMimportant-pager-98358
07/06/2023, 8:39 PMgreen-monitor-16572
07/07/2023, 7:17 AM^
symbol: method schema()
location: variable event of type PlatformEvent
/Users/r0b0d1h/open source code/datahub/metadata-events/mxe-utils-avro-1.7/src/main/java/com/linkedin/metadata/EventUtils.java:314: error: cannot find symbol
DataTranslator.dataMapToGenericRecord(event.data(), event.schema(), ORIGINAL_DUHE_AVRO_SCHEMA);
^
symbol: method data()
location: variable event of type DataHubUpgradeHistoryEvent
/Users/r0b0d1h/open source code/datahub/metadata-events/mxe-utils-avro-1.7/src/main/java/com/linkedin/metadata/EventUtils.java:314: error: cannot find symbol
DataTranslator.dataMapToGenericRecord(event.data(), event.schema(), ORIGINAL_DUHE_AVRO_SCHEMA);
^
symbol: method schema()
location: variable event of type DataHubUpgradeHistoryEvent
Note: /Users/r0b0d1h/open source code/datahub/metadata-events/mxe-utils-avro-1.7/src/main/java/com/linkedin/metadata/EventUtils.java uses or overrides a deprecated API.
Note: Recompile with -Xlint:deprecation for details.
Note: /Users/r0b0d1h/open source code/datahub/metadata-events/mxe-utils-avro-1.7/src/main/java/com/linkedin/metadata/EventUtils.java uses unchecked or unsafe operations.
Note: Recompile with -Xlint:unchecked for details.
36 errors
Kindly help me why I am facing with the above compiler errorgreen-autumn-94171
07/07/2023, 7:28 AMadorable-lawyer-88494
07/07/2023, 7:31 AM[13:00] Sejal Jain
RUN_INGEST - {'errors': [],
'exec_id': '2663b6f4-59b2-4b55-a5e8-9b7d89f1be66',
'infos': ['2023-07-07 06:32:09.122532 [exec_id=2663b6f4-59b2-4b55-a5e8-9b7d89f1be66] INFO: Starting execution for task with name=RUN_INGEST',
'2023-07-07 06:32:31.297837 [exec_id=2663b6f4-59b2-4b55-a5e8-9b7d89f1be66] INFO: stdout=venv setup time = 0\n'
'This version of datahub supports report-to functionality\n'
'datahub ingest run -c /tmp/datahub/ingest/2663b6f4-59b2-4b55-a5e8-9b7d89f1be66/recipe.yml --report-to '
'/tmp/datahub/ingest/2663b6f4-59b2-4b55-a5e8-9b7d89f1be66/ingestion_report.json\n'
'[2023-07-07 06:32:10,812] INFO {datahub.cli.ingest_cli:173} - DataHub CLI version: 0.10.3\n'
'[2023-07-07 06:32:10,848] INFO {datahub.ingestion.run.pipeline:210} - Sink configured successfully. DataHubRestEmitter: configured '
'to talk to <http://datahub-datahub-gms:8080>\n'
'[2023-07-07 06:32:10,862] INFO {datahub.ingestion.run.pipeline:227} - Source configured successfully.\n'
'[2023-07-07 06:32:10,863] INFO {datahub.cli.ingest_cli:129} - Starting metadata ingestion\n'
'[2023-07-07 06:32:29,861] INFO {datahub.ingestion.reporting.file_reporter:52} - Wrote SUCCESS report successfully to '
"<_io.TextIOWrapper name='/tmp/datahub/ingest/2663b6f4-59b2-4b55-a5e8-9b7d89f1be66/ingestion_report.json' mode='w' encoding='UTF-8'>\n"
'[2023-07-07 06:32:29,861] INFO {datahub.cli.ingest_cli:142} - Finished metadata ingestion\n'
'\n'
'Cli report:\n'
"{'cli_version': '0.10.3',\n"
" 'cli_entry_location': '/tmp/datahub/ingest/venv-openapi-0.10.3/lib/python3.10/site-packages/datahub/__init__.py',\n"
" 'py_version': '3.10.7 (main, Sep 13 2022, 14:31:33) [GCC 10.2.1 20210110]',\n"
" 'py_exec_path': '/tmp/datahub/ingest/venv-openapi-0.10.3/bin/python3',\n"
" 'os_details': 'Linux-5.10.179-168.710.amzn2.x86_64-x86_64-with-glibc2.31',\n"
" 'peak_memory_usage': '66.3 MB',\n"
" 'mem_info': '66.3 MB',\n"
" 'peak_disk_usage': '17.64 GB',\n"
" 'disk_info': {'total': '21.46 GB', 'used': '17.64 GB', 'free': '3.82 GB'}}\n"
'Source (openapi) report:\n'
"{'events_produced': 0,\n"
" 'events_produced_per_sec': 0,\n"
" 'entities': {},\n"
" 'aspects': {},\n"
" 'warnings': {'/api/health': ['Unable to find an example for endpoint. Please add it to the list of forced examples.'],\n"
" '/resources/agents/CommPayable/{commPayableID}': ['Unable to find an example for endpoint. Please add it to the list of "
"forced examples.'],\n"
" '/resources/agents/Commissions': ['Unable to find an example for endpoint. Please add it to the list of forced "
"examples.'],\n"
" '/resources/agents/{agentID}/carrierAppointments': ['Unable to find an example for endpoint. Please add it to the list of "
"forced examples.'],\n"
" '/resources/agents/{agentID}/leads': ['Unable to find an example for endpoint. Please add it to the list of forced "
"examples.'],\n"
" '/resources/customers/{ssn}/GetCustomersBySSN': ['Unable to find an example for endpoint. Please add it to the list of "
"forced examples.'],\n"
" '/resources/eapp/{applicationNumber}/covarageMatchedParties': ['Unable to find an example for endpoint. Please add it to "
"the list of forced examples.'],\n"
" '/resources/policies/{policyNumber}/notes': ['Unable to find an example for endpoint. Please add it to the list of forced "
"examples.'],\n"
" '/resources/portalForms/{formCode}': ['Unable to find an example for endpoint. Please add it to the list of forced "
"examples.'],\n"
" '/resources/portalForms/{formCode}/localizeConfig': ['Unable to find an example for endpoint. Please add it to the list of "
"forced examples.'],\n"
" 'sampled': '10 sampled of at most 120 entries.'},\n"
" 'failures': {},\n"
" 'start_time': '2023-07-07 06:32:10.862463 (19.17 seconds ago)',\n"
" 'running_time': '19.17 seconds'}\n"
'Sink (datahub-rest) report:\n'
"{'total_records_written': 0,\n"
" 'records_written_per_second': 0,\n"
" 'warnings': [],\n"
" 'failures': [],\n"
" 'start_time': '2023-07-07 06:32:10.844377 (19.19 seconds ago)',\n"
" 'current_time': '2023-07-07 06:32:30.029331 (now)',\n"
" 'total_duration_in_seconds': 19.18,\n"
" 'gms_version': 'v0.9.6',\n"
" 'pending_requests': 0}\n"
'\n'
' Pipeline finished with at least 120 warnings; produced 0 events in 19.17 seconds.\n'
'❗Client-Server Incompatible❗ Your client version 0.10.3 is newer than your server version 0.9.6. Downgrading the cli to 0.9.6 is '
'recommended.\n'
' ➡️ Downgrade via `"pip install \'acryl-datahub==0.9.6\'"\n',
"2023-07-07 06:32:31.298067 [exec_id=2663b6f4-59b2-4b55-a5e8-9b7d89f1be66] INFO: Successfully executed 'datahub ingest'"],
'structured_report': '{"cli": {"cli_version": "0.10.3", "cli_entry_location": '
'"/tmp/datahub/ingest/venv-openapi-0.10.3/lib/python3.10/site-packages/datahub/__init__.py", "py_version": "3.10.7 (main, Sep '
'13 2022, 14:31:33) [GCC 10.2.1 20210110]", "py_exec_path": "/tmp/datahub/ingest/venv-openapi-0.10.3/bin/python3", '
'"os_details": "Linux-5.10.179-168.710.amzn2.x86_64-x86_64-with-glibc2.31", "peak_memory_usage": "66.03 MB", "mem_info": '
'"66.03 MB", "peak_disk_usage": "17.64 GB", "disk_info": {"total": "21.46 GB", "used": "17.64 GB", "free": "3.82 GB"}}, '
'"source": {"type": "openapi", "report": {"events_produced": 0, "events_produced_per_sec": 0, "entities": {}, "aspects": {}, '
'"warnings": {"/api/health": ["Unable to find an example for endpoint. Please add it to the list of forced examples."], '
'"/resources/agents/CommPayable/{commPayableID}": ["Unable to find an example for endpoint. Please add it to the list of '
'forced examples."], "/resources/agents/Commissions": ["Unable to find an example for endpoint. Please add it to the list of '
'forced examples."], "/resources/agents/{agentID}/carrierAppointments": ["Unable to find an example for endpoint. Please add '
'it to the list of forced examples."], "/resources/agents/{agentID}/leads": ["Unable to find an example for endpoint. Please '
'add it to the list of forced examples."], "/resources/customers/{ssn}/GetCustomersBySSN": ["Unable to find an example for '
'endpoint. Please add it to the list of forced examples."], "/resources/eapp/{applicationNumber}/covarageMatchedParties": '
'["Unable to find an example for endpoint. Please add it to the list of forced examples."], '
'"/resources/policies/{policyNumber}/notes": ["Unable to find an example for endpoint. Please add it to the list of forced '
'examples."], "/resources/portalForms/{formCode}": ["Unable to find an example for endpoint. Please add it to the list of '
'forced examples."], "/resources/portalForms/{formCode}/localizeConfig": ["Unable to find an example for endpoint. Please add '
'it to the list of forced examples."], "sampled": "10 sampled of at most 120 entries."}, "failures": {}, "start_time": '
'"2023-07-07 06:32:10.862463 (19 seconds ago)", "running_time": "19 seconds"}}, "sink": {"type": "datahub-rest", "report": '
'{"total_records_written": 0, "records_written_per_second": 0, "warnings": [], "failures": [], "start_time": "2023-07-07 '
'06:32:10.844377 (19.02 seconds ago)", "current_time": "2023-07-07 06:32:29.861069 (now)", "total_duration_in_seconds": 19.02, '
'"gms_version": "v0.9.6", "pending_requests": 0}}}'}
Execution finished successfully!
broad-pager-68280
07/07/2023, 11:30 AMnice-waiter-58576
07/07/2023, 12:25 PMsome-crowd-4662
07/07/2023, 1:47 PMchilly-elephant-51826
07/08/2023, 11:17 AM2023-07-08 08:56:12,569 [application-akka.actor.default-dispatcher-15] WARN p.api.mvc.LegacySessionCookieBaker - Cookie failed message authentication check
2023-07-08 08:56:12,684 [application-akka.actor.default-dispatcher-15] WARN p.api.mvc.LegacySessionCookieBaker - Cookie failed message authentication check
2023-07-08 08:56:12,684 [application-akka.actor.default-dispatcher-9] WARN p.api.mvc.LegacySessionCookieBaker - Cookie failed message authentication check
2023-07-08 08:56:12,685 [application-akka.actor.default-dispatcher-13] WARN p.api.mvc.LegacySessionCookieBaker - Cookie failed message authentication check
2023-07-08 08:56:12,692 [application-akka.actor.default-dispatcher-15] WARN p.api.mvc.LegacySessionCookieBaker - Cookie failed message authentication check
2023-07-08 08:56:12,692 [application-akka.actor.default-dispatcher-9] WARN p.api.mvc.LegacySessionCookieBaker - Cookie failed message authentication check
2023-07-08 08:56:13,725 [application-akka.actor.default-dispatcher-10] WARN p.api.mvc.LegacySessionCookieBaker - Cookie failed message authentication check
2023-07-08 08:56:13,726 [application-akka.actor.default-dispatcher-15] ERROR controllers.SsoCallbackController - Caught exception while attempting to handle SSO callback! It's likely that SSO integration is mis-configured.
java.util.concurrent.CompletionException: org.pac4j.core.exception.TechnicalException: State cannot be determined
at java.base/java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:314)
at java.base/java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:319)
at java.base/java.util.concurrent.CompletableFuture$AsyncSupply.run(CompletableFuture.java:1702)
at play.core.j.HttpExecutionContext.$anonfun$execute$1(HttpExecutionContext.scala:64)
at akka.dispatch.TaskInvocation.run(AbstractDispatcher.scala:49)
at akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(ForkJoinExecutorConfigurator.scala:48)
at java.base/java.util.concurrent.ForkJoinTask.doExec(ForkJoinTask.java:290)
at java.base/java.util.concurrent.ForkJoinPool$WorkQueue.topLevelExec(ForkJoinPool.java:1020)
at java.base/java.util.concurrent.ForkJoinPool.scan(ForkJoinPool.java:1656)
at java.base/java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1594)
at java.base/java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:183)
Caused by: org.pac4j.core.exception.TechnicalException: State cannot be determined
at org.pac4j.oidc.credentials.extractor.OidcExtractor.lambda$extract$0(OidcExtractor.java:100)
at java.base/java.util.Optional.orElseThrow(Optional.java:408)
at org.pac4j.oidc.credentials.extractor.OidcExtractor.extract(OidcExtractor.java:100)
at org.pac4j.core.client.BaseClient.retrieveCredentials(BaseClient.java:66)
at org.pac4j.core.client.IndirectClient.getCredentials(IndirectClient.java:143)
at org.pac4j.core.engine.DefaultCallbackLogic.perform(DefaultCallbackLogic.java:85)
at auth.sso.oidc.OidcCallbackLogic.perform(OidcCallbackLogic.java:100)
at controllers.SsoCallbackController$SsoCallbackLogic.perform(SsoCallbackController.java:91)
at controllers.SsoCallbackController$SsoCallbackLogic.perform(SsoCallbackController.java:77)
at org.pac4j.play.CallbackController.lambda$callback$0(CallbackController.java:54)
at java.base/java.util.concurrent.CompletableFuture$AsyncSupply.run(CompletableFuture.java:1700)
... 8 common frames omitted
there is an indication that oidc config are mis configured, but can that be an issue ? as on redeploying the application application everything looks finefancy-crayon-39356
07/09/2023, 5:01 PMdatahub-frontend
, it just doesn't find the tableau assets in Datahub: "Sorry, we are unable to find this entity in DataHub".
Digging into the problem, I see that the extension uses the useGetSearchResultsForMultipleQuery
function (https://github.com/datahub-project/datahub/pull/8033) that basically calls the searchAcrossEntities
query using externalUrl, chartUrl, dashboardUrl
fields as filters. So I figured out that the problem is with this query - on my side, the filter simply doesn't work. I've managed to replicate the query that the plugin does on GraphiQL, and the server is not able to filter for externalUrl
, chartUrl
or dashboardUrl
. The strange thing is that externalUrl
is now a searchable field: https://github.com/datahub-project/datahub/pull/7953
Example of query I'm running:
searchAcrossEntities(input: {query: "*", start: 0, count: 2, orFilters: [{and: [{field: "externalUrl", values: ["<https://my-tableau-server-url>..."]}]}]}) {
searchResults {
entity {
urn,
type
}
}
}
searchResults
is empty here - and I'm providing the correct externalUrl, checked many times.
Does anyone have an idea why I can't search using the externalUrl
field on the filters? My thinking is that this is the root cause behind the Chrome Extension for Tableau not working.
Datahub version: v0.10.3
Deployment method: Helm charts
Help would be much appreciated 🙏 @big-carpet-38439numerous-account-62719
07/10/2023, 11:07 AM