astonishing-dream-54015
06/12/2023, 3:00 PMStarting up DataHub...
[+] Running 7/8
⠿ Container mysql Healthy 0.5s
⠿ Container zookeeper Healthy 5.9s
⠿ Container mysql-setup Started 0.8s
⠿ Container elasticsearch Healthy 1.9s
⠿ Container elasticsearch-setup Started 2.1s
⠿ Container broker Waiting 821.1s
⠿ Container schema-registry Created 0.0s
⠿ Container kafka-setup Created 0.0s
In broker log there is an error:
[2023-06-12 14:45:40,243] INFO Initiating client connection, connectString=zookeeper:2181 sessionTimeout=18000 watcher=kafka.zookeeper.ZooKeeperClient$ZooKeeperClientWatcher$@1ecee32c (org.apache.zookeeper.ZooKeeper)
[2023-06-12 14:45:40,247] INFO jute.maxbuffer value is 4194304 Bytes (org.apache.zookeeper.ClientCnxnSocket)
[2023-06-12 14:45:40,251] INFO zookeeper.request.timeout value is 0. feature enabled=false (org.apache.zookeeper.ClientCnxn)
[2023-06-12 14:45:40,255] INFO Opening socket connection to server zookeeper/172.19.0.4:2181. (org.apache.zookeeper.ClientCnxn)
[2023-06-12 14:45:40,257] INFO [ZooKeeperClient Kafka server] Waiting until connected. (kafka.zookeeper.ZooKeeperClient)
[2023-06-12 14:45:40,260] INFO Socket connection established, initiating session, client: /172.19.0.6:50088, server: zookeeper/172.19.0.4:2181 (org.apache.zookeeper.ClientCnxn)
[2023-06-12 14:45:40,268] INFO Session establishment complete on server zookeeper/172.19.0.4:2181, session id = 0x10000010f140001, negotiated timeout = 18000 (org.apache.zookeeper.ClientCnxn)
[2023-06-12 14:45:40,271] INFO [ZooKeeperClient Kafka server] Connected. (kafka.zookeeper.ZooKeeperClient)
[2023-06-12 14:45:40,343] INFO [feature-zk-node-event-process-thread]: Starting (kafka.server.FinalizedFeatureChangeListener$ChangeNotificationProcessorThread)
[2023-06-12 14:45:40,355] INFO Feature ZK node at path: /feature does not exist (kafka.server.FinalizedFeatureChangeListener)
[2023-06-12 14:45:40,355] INFO Cleared cache (kafka.server.FinalizedFeatureCache)
[2023-06-12 14:45:40,470] INFO Cluster ID = VBPvRVOsQmK-gMNc1hRWRg (kafka.server.KafkaServer)
[2023-06-12 14:45:40,476] ERROR Fatal error during KafkaServer startup. Prepare to shutdown (kafka.server.KafkaServer)
kafka.common.InconsistentClusterIdException: The Cluster ID VBPvRVOsQmK-gMNc1hRWRg doesn't match stored clusterId Some(5J_OZE3xSnicskM-BRhbyA) in meta.properties. The broker is trying to join the wrong cluster. Configured zookeeper.connect may be wrong.
at kafka.server.KafkaServer.startup(KafkaServer.scala:230)
at kafka.Kafka$.main(Kafka.scala:109)
at kafka.Kafka.main(Kafka.scala)
[2023-06-12 14:45:40,478] INFO shutting down (kafka.server.KafkaServer)
[2023-06-12 14:45:40,479] INFO [feature-zk-node-event-process-thread]: Shutting down (kafka.server.FinalizedFeatureChangeListener$ChangeNotificationProcessorThread)
[2023-06-12 14:45:40,479] INFO [feature-zk-node-event-process-thread]: Stopped (kafka.server.FinalizedFeatureChangeListener$ChangeNotificationProcessorThread)
[2023-06-12 14:45:40,479] INFO [feature-zk-node-event-process-thread]: Shutdown completed (kafka.server.FinalizedFeatureChangeListener$ChangeNotificationProcessorThread)
[2023-06-12 14:45:40,480] INFO [ZooKeeperClient Kafka server] Closing. (kafka.zookeeper.ZooKeeperClient)
[2023-06-12 14:45:40,585] INFO Session: 0x10000010f140001 closed (org.apache.zookeeper.ZooKeeper)
[2023-06-12 14:45:40,585] INFO EventThread shut down for session: 0x10000010f140001 (org.apache.zookeeper.ClientCnxn)
[2023-06-12 14:45:40,586] INFO [ZooKeeperClient Kafka server] Closed. (kafka.zookeeper.ZooKeeperClient)
[2023-06-12 14:45:40,591] INFO App info kafka.server for 1 unregistered (org.apache.kafka.common.utils.AppInfoParser)
[2023-06-12 14:45:40,591] INFO shut down completed (kafka.server.KafkaServer)
[2023-06-12 14:45:40,591] ERROR Exiting Kafka. (kafka.Kafka$)
[2023-06-12 14:45:40,591] INFO shutting down (kafka.server.KafkaServer)
Any help would be appreciated!! Thanks in advance~~~~bland-barista-59197
06/12/2023, 6:15 PMpurple-printer-15193
06/12/2023, 6:58 PMelegant-guitar-28442
06/13/2023, 2:35 AMadorable-lawyer-88494
06/13/2023, 7:21 AMFailure 1 of 1The :li-utils:compileMainGeneratedDataTemplateJava task failed.View task in console log
Compilation failed; see the compiler error output for details.
Exception
org.gradle.api.tasks.TaskExecutionException: Execution failed for task ':li-utils:compileMainGeneratedDataTemplateJava'.
at org.gradle.api.internal.tasks.execution.ExecuteActionsTaskExecuter.lambda$executeIfValid$1(ExecuteActionsTaskExecuter.java:145)
•••
Caused by: org.gradle.api.internal.tasks.compile.CompilationFailedException: Compilation failed; see the compiler error output for details.
at org.gradle.api.internal.tasks.compile.JdkJavaCompiler.execute(JdkJavaCompiler.java:56)
•••
swift-dream-78272
06/13/2023, 12:04 PMbillions-baker-82097
06/13/2023, 1:07 PMmysterious-advantage-78411
06/13/2023, 2:40 PMcreamy-battery-20182
06/13/2023, 5:09 PM023-06-12 22:40:55,684 [qtp944427387-17466] INFO c.l.m.r.entity.AspectResource:166 - INGEST PROPOSAL proposal: {aspectName=assertionInfo, systemMetadata={lastObserved=1686609651915, runId=dbt-2023_06_12-22_40_42}, entityUrn=urn:li:assertion:d8691f1c759e159221940a3696e48cf8, entityType=assertion, aspect={contentType=application/json, value=ByteString(length=1375,bytes=7b226375...6e227d7d)}, changeType=UPSERT}
2023-06-12 22:40:55,687 [qtp944427387-17421] ERROR c.l.m.filter.RestliLoggingFilter:38 - <http://Rest.li|Rest.li> error:
com.linkedin.restli.server.RestLiServiceException: com.datahub.util.exception.RetryLimitReached: Failed to add after 3 retries
But these are the underlying exceptions (logs are from the GMS pod):
Caused by: io.ebean.DuplicateKeyException: Error when batch flush on sql: insert into metadata_aspect_v2 (urn, aspect, version, metadata, createdOn, createdBy, createdFor, systemmetadata) values (?,?,?,?,?,?,?,?)
Caused by: java.sql.BatchUpdateException: Duplicate entry 'urn:li:assertion:04063f0fbcbe627b390598a883fb0272-assertionInfo-' for key 'PRIMARY'
Caused by: java.sql.SQLIntegrityConstraintViolationException: Duplicate entry 'urn:li:assertion:04063f0fbcbe627b390598a883fb0272-assertionInfo-' for key 'PRIMARY'
Has anyone seen these before? What could be the underlying issue here, is there an issue with the data itself?flat-engineer-75197
06/13/2023, 5:24 PMcool_kids
group on the UI and added my user to it. The group’s URN tells me it’s of type corpGroup but when I queried the user’s aspects, it instead shows up as a native group.victorious-monkey-86128
06/13/2023, 8:29 PMincalculable-portugal-45517
06/14/2023, 3:31 AMNo such file or directory
for /docker/datahub-upgrade/datahub-upgrade.sh
and ./datahub-upgrade.sh
(I also can't find any datahub-upgrade.sh
in the container when I run it locally with docker run)bumpy-shoe-90203
06/14/2023, 6:38 AMhelm install datahub datahub/datahub
Error: INSTALLATION FAILED: YAML parse error on datahub/templates/datahub-upgrade/datahub-cleanup-job-template.yml: error converting YAML to JSON: yaml: line 91: did not find expected '-' indicator
bland-gigabyte-28270
06/14/2023, 8:55 AM0.10.2
to 0.10.3
, and seems like the secret set before cannot be accessed anymore. Is this expected? Details in 🧵enough-football-92033
06/14/2023, 12:09 PMuser.props
file and able to login with them but this user can't access Manage Permissions
tab. For now I able to create policies only with default datahub user, I guess it a bug and how I can fix it?proud-dusk-671
06/14/2023, 12:10 PMnumerous-autumn-22862
06/14/2023, 2:42 PMError while getting broker list.
java.util.concurrent.ExecutionException: org.apache.kafka.common.errors.TimeoutException: Timed out waiting for a node assignment. Call: listNodes
this is an error happening in the kafka-setup-job
We are running kafka in MSK from AWSlate-dawn-4912
06/14/2023, 11:41 PMfaint-oyster-25890
06/15/2023, 3:56 AM/etc/datahub/plugins/retention/retention.yaml
like this
- entity: "dataProcessInstance"
aspect: "*"
config:
retention:
version:
maxVersions: 1
time:
maxAgeInSeconds: 2592000 # 30 days
But, it did not work. Basically, my goal is to keep retaining every DataProcessInstance history for 1 monthbest-wire-59738
06/15/2023, 4:34 AM./gradlew build
.
When I tried to build frontend image using the below dockerfile the Image build is getting stuck in the middle at gradle build command and it’s not moving forward. I had check the logs using --debug
mode but didn’t figure out the actual issue. I had also attached the logs for your reference. Could you please help me out with the issue.
docker buildx build . -t datahub --platform=linux/arm64
brief-nail-41206
06/15/2023, 5:01 AMget_usage_aspects_from_urn
method. I was able to get this at the table level but not at a project or even dataset level. It gives an empty response when I use a container entity (like BQ dataset or project).
Would you know how I would get these stats (like most queried tables in my project) at a project level using graphql?stocky-guitar-68560
06/15/2023, 6:41 AMpowerful-tent-14193
06/15/2023, 7:49 AMicy-flag-80360
06/15/2023, 8:53 AMpowerful-cat-68806
06/15/2023, 9:10 AMproud-dusk-671
06/15/2023, 9:49 AMReading rows 1000 through 2000 from the aspects table completed.
metrics so far RestoreIndicesResult(ignored=0, rowsMigrated=1000, timeSqlQueryMs=86, timeGetRowMs=0, timeUrnMs=4, timeEntityRegistryCheckMs=1, aspectCheckMs=76, createRecordMs=3814, sendMessageMs=26099)
Successfully sent MAEs for 1000/12805 rows (7.81% of total). 0 rows ignored (0.00% of total)
0.64 mins taken. 7.52 est. mins to completion. Total mins est. = 8.15.
Args are RestoreIndicesArgs(start=2000, batchSize=1000, numThreads=1, batchDelayMs=100, aspectName=null, urn=null, urnLike=null)
Reading rows 2000 through 3000 from the aspects table started.
Reading rows 2000 through 3000 from the aspects table completed.
ANTLR Tool version 4.5 used for code generation does not match the current runtime version 4.7.2ANTLR Runtime version 4.5 used for parser compilation does not match the current runtime version 4.7.2ANTLR Tool version 4.5 used for code generation does not match the current runtime version 4.7.2ANTLR Runtime version 4.5 used for parser compilation does not match the current runtime version 4.7.2java.util.concurrent.ExecutionException: java.lang.OutOfMemoryError: Java heap space
at java.base/java.util.concurrent.FutureTask.report(FutureTask.java:122)
at java.base/java.util.concurrent.FutureTask.get(FutureTask.java:191)
at com.linkedin.datahub.upgrade.restoreindices.SendMAEStep.iterateFutures(SendMAEStep.java:71)
at com.linkedin.datahub.upgrade.restoreindices.SendMAEStep.lambda$executable$0(SendMAEStep.java:138)
at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.executeStepInternal(DefaultUpgradeManager.java:106)
at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.executeInternal(DefaultUpgradeManager.java:65)
at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.executeInternal(DefaultUpgradeManager.java:39)
at com.linkedin.datahub.upgrade.impl.DefaultUpgradeManager.execute(DefaultUpgradeManager.java:30)
at com.linkedin.datahub.upgrade.UpgradeCli.run(UpgradeCli.java:80)
at org.springframework.boot.SpringApplication.callRunner(SpringApplication.java:768)
at org.springframework.boot.SpringApplication.callRunners(SpringApplication.java:752)
at org.springframework.boot.SpringApplication.run(SpringApplication.java:314)
at org.springframework.boot.builder.SpringApplicationBuilder.run(SpringApplicationBuilder.java:164)
at com.linkedin.datahub.upgrade.UpgradeCliApplication.main(UpgradeCliApplication.java:23)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.base/java.lang.reflect.Method.invoke(Method.java:566)
at org.springframework.boot.loader.MainMethodRunner.run(MainMethodRunner.java:49)
at org.springframework.boot.loader.Launcher.launch(Launcher.java:108)
at org.springframework.boot.loader.Launcher.launch(Launcher.java:58)
Config in helm -
datahubUpgrade:
enabled: true
image:
repository: acryldata/datahub-upgrade
# tag: "v0.10.0" # defaults to .global.datahub.version
batchSize: 1000
batchDelayMs: 100
noCodeDataMigration:
sqlDbType: "MYSQL"
# sqlDbType: "POSTGRES"
podSecurityContext: {}
# fsGroup: 1000
securityContext: {}
# runAsUser: 1000
podAnnotations:
"<http://sidecar.istio.io/inject|sidecar.istio.io/inject>": 'false'
# Add extra sidecar containers to job pod
extraSidecars: []
# - name: my-image-name
# image: my-image
# imagePullPolicy: Always
cleanupJob:
resources:
limits:
cpu: 500m
memory: 512Mi
requests:
cpu: 300m
memory: 256Mi
# Add extra sidecar containers to job pod
extraSidecars: []
# - name: my-image-name
# image: my-image
# imagePullPolicy: Always
restoreIndices:
resources:
limits:
cpu: 500m
memory: 512Mi
requests:
cpu: 300m
memory: 256Mi
flat-engineer-75197
06/15/2023, 2:22 PMcorpUser
not a valid entity type in a transformer? My transform_aspect
method refuses to run when I have this:
def entity_types(self) -> List[str]:
return ["corpUser"]
But it’s fine if I do:
def entity_types(self) -> List[str]:
return ["*"]
silly-fish-85029
06/15/2023, 2:37 PMabundant-grass-62044
06/15/2023, 2:53 PMswift-processor-45491
06/15/2023, 3:59 PM