Howdy everyone - would you please lend me a hand ?...
# troubleshoot
t
Howdy everyone - would you please lend me a hand ? I am now deploying datahub with EKS / MSK / ES. However i met a wired issue, which was when i turned on multi partitions for MSK(aws managed kafka)
Copy code
kafka:
    bootstrap:
      server: "xxx"
    zookeeper:
      server: "xxx"
    ## For AWS MSK set this to a number larger than 1
    partitions: 3
    replicationFactor: 3
however, that led to datahub-gms ran abnormally, kept reporting issues as below, which resulted in unable to create AccessToken/Glossary etc. The only way to resolve this was to comment the
partitions & replicationFactor
. May i seek help here - any chance to enable multi partitions along with gms work properly ? Thanks.
i
Hello Vincent, Do you have logs for the kafka setup job?
That job is responsible for creating the topics with the adequate partitions and replication factor, it may take a while and that is why GMS fails. One thing you can do workaround this is to disable the kafka setup job, assuming the topic exists and then go in to MSK console and manually edit the partition counts
t
Thanks Pedro for response. Comes with the log
Copy code
[main] INFO org.apache.kafka.clients.admin.AdminClientConfig - AdminClientConfig values:
        bootstrap.servers = [<http://b-2.xxxxxxx.amazonaws.com:9092|b-2.xxxxxxx.amazonaws.com:9092>, <http://b-1.xxxxx.amazonaws.com:9092|b-1.xxxxx.amazonaws.com:9092>]
        client.dns.lookup = use_all_dns_ips
        client.id =
        <http://connections.max.idle.ms|connections.max.idle.ms> = 300000
        <http://default.api.timeout.ms|default.api.timeout.ms> = 60000
        <http://metadata.max.age.ms|metadata.max.age.ms> = 300000
        metric.reporters = []
        metrics.num.samples = 2
        metrics.recording.level = INFO
        <http://metrics.sample.window.ms|metrics.sample.window.ms> = 30000
        receive.buffer.bytes = 65536
        <http://reconnect.backoff.max.ms|reconnect.backoff.max.ms> = 1000
        <http://reconnect.backoff.ms|reconnect.backoff.ms> = 50
        <http://request.timeout.ms|request.timeout.ms> = 30000
        retries = 2147483647
        <http://retry.backoff.ms|retry.backoff.ms> = 100
        sasl.client.callback.handler.class = null
        sasl.jaas.config = null
        sasl.kerberos.kinit.cmd = /usr/bin/kinit
        sasl.kerberos.min.time.before.relogin = 60000
        sasl.kerberos.service.name = null
        sasl.kerberos.ticket.renew.jitter = 0.05
        sasl.kerberos.ticket.renew.window.factor = 0.8
        sasl.login.callback.handler.class = null
        sasl.login.class = null
        sasl.login.refresh.buffer.seconds = 300
        sasl.login.refresh.min.period.seconds = 60
        sasl.login.refresh.window.factor = 0.8
        sasl.login.refresh.window.jitter = 0.05
        sasl.mechanism = GSSAPI
        security.protocol = PLAINTEXT
        security.providers = null
        send.buffer.bytes = 131072
        <http://socket.connection.setup.timeout.max.ms|socket.connection.setup.timeout.max.ms> = 127000
        <http://socket.connection.setup.timeout.ms|socket.connection.setup.timeout.ms> = 10000
        ssl.cipher.suites = null
        ssl.enabled.protocols = [TLSv1.2, TLSv1.3]
        ssl.endpoint.identification.algorithm = https
        ssl.engine.factory.class = null
        ssl.key.password = null
        ssl.keymanager.algorithm = SunX509
        ssl.keystore.certificate.chain = null
        ssl.keystore.key = null
        ssl.keystore.location = null
        ssl.keystore.password = null
        ssl.keystore.type = JKS
        ssl.protocol = TLSv1.3
        ssl.provider = null
        ssl.secure.random.implementation = null
        ssl.trustmanager.algorithm = PKIX
        ssl.truststore.certificates = null
        ssl.truststore.location = null
        ssl.truststore.password = null
        ssl.truststore.type = JKS

[main] INFO org.apache.kafka.common.utils.AppInfoParser - Kafka version: 6.1.4-ccs
[main] INFO org.apache.kafka.common.utils.AppInfoParser - Kafka commitId: c9124241a6ff43bc
[main] INFO org.apache.kafka.common.utils.AppInfoParser - Kafka startTimeMs: 1668072757195
WARNING: Due to limitations in metric names, topics with a period ('.') or underscore ('_') could collide. To avoid issues it is best to use either, but not both.
WARNING: Due to limitations in metric names, topics with a period ('.') or underscore ('_') could collide. To avoid issues it is best to use either, but not both.
WARNING: Due to limitations in metric names, topics with a period ('.') or underscore ('_') could collide. To avoid issues it is best to use either, but not both.
WARNING: Due to limitations in metric names, topics with a period ('.') or underscore ('_') could collide. To avoid issues it is best to use either, but not both.
WARNING: Due to limitations in metric names, topics with a period ('.') or underscore ('_') could collide. To avoid issues it is best to use either, but not both.
WARNING: Due to limitations in metric names, topics with a period ('.') or underscore ('_') could collide. To avoid issues it is best to use either, but not both.
WARNING: Due to limitations in metric names, topics with a period ('.') or underscore ('_') could collide. To avoid issues it is best to use either, but not both.
WARNING: Due to limitations in metric names, topics with a period ('.') or underscore ('_') could collide. To avoid issues it is best to use either, but not both.
WARNING: Due to limitations in metric names, topics with a period ('.') or underscore ('_') could collide. To avoid issues it is best to use either, but not both.
Completed updating config for topic _schemas.
everything looks good..
or, it may take a while and that is why GMS fails.
i cannot reproduce that anymore (i deleted & reinstall datahub/datahub but not redeploy everything from scratch). The screenshot i sent before was what i saw from gms continuously. So you mean, i may need to wait long enough time ?
i
Correct
The Kafka setup job may take longer than expected while performing some operations.
t
thanks Pedro. Btw - did you meet an issue which from WebUI when you creating the
integration
?
[(com.datahub.authentication.authenticator.DataHubSystemAuthenticator,Failed to authenticate inbound request: Authorization header is missing Authorization header.), (com.datahub.authentication.authenticator.DataHubTokenAuthenticator,Failed to authenticate inbound request: Request is missing ‘Authorization’ header.)]
i can only create the integration from cli, not from webui..