bitter-waitress-17567
01/03/2023, 4:42 PMbright-egg-51769
01/04/2023, 7:27 PMgreat-monkey-52307
12/28/2022, 9:43 PMstrong-belgium-32572
01/05/2023, 4:39 PMError: template: datahub/charts/datahub-ingestion-cron/templates/cron.yaml:38:109: executing "datahub/charts/datahub-ingestion-cron/templates/cron.yaml" at <.Values.image.tag>: nil pointer evaluating interface {}.image
microscopic-mechanic-13766
01/09/2023, 12:39 PMstrong-belgium-32572
01/09/2023, 1:10 PMrefined-energy-76018
01/09/2023, 11:51 PMreplicaCount
of datahub-gms from 1 to 2-3, do the mae/mce consumers have to be run in standalone mode as to not run into any issues?microscopic-mechanic-13766
01/10/2023, 9:44 AMlimited-library-89060
01/11/2023, 10:57 AMbitter-waitress-17567
01/11/2023, 3:49 PMred-waitress-53338
01/11/2023, 6:34 PMwitty-motorcycle-52108
01/11/2023, 9:43 PMSSLError(OSError(24, 'Too many open files')))
. we're using the pre-built container from docker hub so im not sure what the ulimit is, but it seems like something in actions may not be releasing open file descriptors. has anyone seen this before?red-waitress-53338
01/11/2023, 11:40 PMlate-book-30206
01/12/2023, 9:28 AMwide-butcher-58942
01/12/2023, 3:36 PMwide-butcher-58942
01/12/2023, 3:48 PM./gradlew :datahub-frontend:dist -x yarnTest -x yarnLint
is Failing on this step Task :datahub-frontend:compileScala FAILED
. (Error in Thread)
* What went wrong:
Execution failed for task ':datahub-frontend:compileScala'.
> java.io.IOException: Cannot run program "/usr/lib/jvm/java-11-openjdk-amd64/bin/javac" (in directory "/home/rupesh/.gradle/workers"): error=2, No such file or directory
Any thoughts on how to resolve. (I am on openJDK 11)
Thanksred-waitress-53338
01/13/2023, 2:36 AMgentle-portugal-21014
01/13/2023, 5:04 PMred-waitress-53338
01/15/2023, 10:41 PMKAFKA_BOOTSTRAP_SERVER=<http://kafka.xxxxx.com:443|kafka.xxxxx.com:443>
KAFKA_SCHEMAREGISTRY_URL=<https://schemaregistry.xxxxx.com:443>
SPRING_KAFKA_PROPERTIES_SECURITY_PROTOCOL=SASL_SSL
SPRING_KAFKA_PROPERTIES_SASL_JAAS_CONFIG=org.apache.kafka.common.security.plain.PlainLoginModule required username='xxxxx' password='xxxxx';
SPRING_KAFKA_PROPERTIES_SASL_MECHANISM=PLAIN
SPRING_KAFKA_PROPERTIES_CLIENT_DNS_LOOKUP=use_all_dns_ips
SPRING_KAFKA_PROPERTIES_SSL_TRUSTSTORE_LOCATION=/src/main/resources/truststore.jks
SPRING_KAFKA_PROPERTIES_SSL_TRUSTSTORE_PASSWORD=xxxxx
I am getting the following error when running GMS docker image locally. I think the issue is with the Kafka SSL.
22:30:32.237 [ThreadPoolTaskExecutor-1] INFO org.apache.kafka.clients.Metadata:277 - [Consumer clientId=consumer-mce-consumer-job-client-1, groupId=mce-consumer-job-client] Cluster ID: XpLVkk39TyK_obCIQyz4rA
22:30:32.237 [ThreadPoolTaskExecutor-1] INFO org.apache.kafka.clients.Metadata:277 - [Consumer clientId=consumer-mce-consumer-job-client-1, groupId=mce-consumer-job-client] Cluster ID: XpLVkk39TyK_obCIQyz4rA
22:30:32.239 [ThreadPoolTaskExecutor-1] ERROR o.s.k.l.KafkaMessageListenerContainer$ListenerConsumer:149 - Authorization Exception and no authorizationExceptionRetryInterval set
org.apache.kafka.common.errors.GroupAuthorizationException: Not authorized to access group: mce-consumer-job-client
22:30:32.239 [ThreadPoolTaskExecutor-1] ERROR o.s.k.l.KafkaMessageListenerContainer$ListenerConsumer:140 - Fatal consumer exception; stopping container
22:30:32.250 [ThreadPoolTaskExecutor-1] INFO o.s.k.l.KafkaMessageListenerContainer$ListenerConsumer:292 - mce-consumer-job-client: Consumer stopped
22:30:32.280 [ThreadPoolTaskExecutor-1] INFO org.apache.kafka.clients.Metadata:277 - [Consumer clientId=consumer-generic-mce-consumer-job-client-2, groupId=generic-mce-consumer-job-client] Cluster ID: XpLVkk39TyK_obCIQyz4rA
22:30:32.280 [ThreadPoolTaskExecutor-1] INFO org.apache.kafka.clients.Metadata:277 - [Consumer clientId=consumer-generic-mce-consumer-job-client-2, groupId=generic-mce-consumer-job-client] Cluster ID: XpLVkk39TyK_obCIQyz4rA
22:30:32.339 [ThreadPoolTaskExecutor-1] INFO org.apache.kafka.clients.Metadata:277 - [Consumer clientId=consumer-datahub-usage-event-consumer-job-client-3, groupId=datahub-usage-event-consumer-job-client] Cluster ID: XpLVkk39TyK_obCIQyz4rA
22:30:32.339 [ThreadPoolTaskExecutor-1] INFO org.apache.kafka.clients.Metadata:277 - [Consumer clientId=consumer-datahub-usage-event-consumer-job-client-3, groupId=datahub-usage-event-consumer-job-client] Cluster ID: XpLVkk39TyK_obCIQyz4rA
22:30:32.341 [ThreadPoolTaskExecutor-1] ERROR o.s.k.l.KafkaMessageListenerContainer$ListenerConsumer:149 - Authorization Exception and no authorizationExceptionRetryInterval set
org.apache.kafka.common.errors.GroupAuthorizationException: Not authorized to access group: datahub-usage-event-consumer-job-client
22:30:32.341 [ThreadPoolTaskExecutor-1] ERROR o.s.k.l.KafkaMessageListenerContainer$ListenerConsumer:140 - Fatal consumer exception; stopping container
22:30:32.344 [ThreadPoolTaskExecutor-1] INFO o.s.k.l.KafkaMessageListenerContainer$ListenerConsumer:292 - datahub-usage-event-consumer-job-client: Consumer stopped
22:30:32.387 [ThreadPoolTaskExecutor-1] INFO org.apache.kafka.clients.Metadata:277 - [Consumer clientId=consumer-generic-mae-consumer-job-client-4, groupId=generic-mae-consumer-job-client] Cluster ID: XpLVkk39TyK_obCIQyz4rA
22:30:32.387 [ThreadPoolTaskExecutor-1] INFO org.apache.kafka.clients.Metadata:277 - [Consumer clientId=consumer-generic-mae-consumer-job-client-4, groupId=generic-mae-consumer-job-client] Cluster ID: XpLVkk39TyK_obCIQyz4rA
22:30:32.388 [ThreadPoolTaskExecutor-1] ERROR o.s.k.l.KafkaMessageListenerContainer$ListenerConsumer:149 - Authorization Exception and no authorizationExceptionRetryInterval set
org.apache.kafka.common.errors.GroupAuthorizationException: Not authorized to access group: generic-mae-consumer-job-client
22:30:32.388 [ThreadPoolTaskExecutor-1] ERROR o.s.k.l.KafkaMessageListenerContainer$ListenerConsumer:140 - Fatal consumer exception; stopping container
22:30:32.390 [ThreadPoolTaskExecutor-1] INFO o.s.k.l.KafkaMessageListenerContainer$ListenerConsumer:292 - generic-mae-consumer-job-client: Consumer stopped
22:30:32.468 [ThreadPoolTaskExecutor-1] INFO org.apache.kafka.clients.Metadata:277 - [Consumer clientId=consumer-generic-platform-event-job-client-5, groupId=generic-platform-event-job-client] Cluster ID: XpLVkk39TyK_obCIQyz4rA
22:30:32.468 [ThreadPoolTaskExecutor-1] INFO org.apache.kafka.clients.Metadata:277 - [Consumer clientId=consumer-generic-platform-event-job-client-5, groupId=generic-platform-event-job-client] Cluster ID: XpLVkk39TyK_obCIQyz4rA
22:30:32.497 [main] INFO c.l.metadata.boot.BootstrapManager:33 - Executing bootstrap step 2/10 with name IngestPoliciesStep...
22:30:32.498 [main] INFO c.l.m.boot.steps.IngestPoliciesStep:60 - Ingesting default access policies...
22:30:32.500 [main] INFO c.l.m.boot.steps.IngestPoliciesStep:85 - Ingesting default policy with urn urn:li:dataHubPolicy:0
22:30:32.581 [ThreadPoolTaskExecutor-1] ERROR o.s.k.l.KafkaMessageListenerContainer$ListenerConsumer:149 - Authorization Exception and no authorizationExceptionRetryInterval set
org.apache.kafka.common.errors.GroupAuthorizationException: Not authorized to access group: generic-mce-consumer-job-client
22:30:32.582 [ThreadPoolTaskExecutor-1] ERROR o.s.k.l.KafkaMessageListenerContainer$ListenerConsumer:140 - Fatal consumer exception; stopping container
22:30:32.584 [ThreadPoolTaskExecutor-1] INFO o.s.k.l.KafkaMessageListenerContainer$ListenerConsumer:292 - generic-mce-consumer-job-client: Consumer stopped
22:30:32.769 [ThreadPoolTaskExecutor-1] ERROR o.s.k.l.KafkaMessageListenerContainer$ListenerConsumer:149 - Authorization Exception and no authorizationExceptionRetryInterval set
org.apache.kafka.common.errors.GroupAuthorizationException: Not authorized to access group: generic-platform-event-job-client
22:30:32.769 [ThreadPoolTaskExecutor-1] ERROR o.s.k.l.KafkaMessageListenerContainer$ListenerConsumer:140 - Fatal consumer exception; stopping container
red-waitress-53338
01/15/2023, 10:41 PM22:30:32.771 [ThreadPoolTaskExecutor-1] INFO o.s.k.l.KafkaMessageListenerContainer$ListenerConsumer:292 - generic-platform-event-job-client: Consumer stopped
22:30:33.085 [pool-7-thread-1] WARN org.elasticsearch.client.RestClient:65 - request [POST <https://99b1f46f43124c9dbfe6a6de8cee78e8.psc.us-central1.gcp.cloud.es.io:9243/datahubpolicyindex_v2/_search?typed_keys=true&max_concurrent_shard_requests=5&ignore_unavailable=false&expand_wildcards=open&allow_no_indices=true&ignore_throttled=true&search_type=query_then_fetch&batched_reduce_size=512&ccs_minimize_roundtrips=true>] returned 1 warnings: [299 Elasticsearch-7.17.8-120eabe1c8a0cb2ae87cffc109a5b65d213e9df1 "[ignore_throttled] parameter is deprecated because frozen indices have been deprecated. Consider cold or frozen tiers in place of frozen indices."]
22:30:33.170 [main] INFO c.l.m.boot.steps.IngestPoliciesStep:85 - Ingesting default policy with urn urn:li:dataHubPolicy:1
22:30:34.329 [main] INFO c.l.m.boot.steps.IngestPoliciesStep:93 - Skipping ingestion of editable policy with urn urn:li:dataHubPolicy:7
22:30:34.713 [main] INFO c.l.m.boot.steps.IngestPoliciesStep:93 - Skipping ingestion of editable policy with urn urn:li:dataHubPolicy:view-entity-page-all
22:30:34.869 [main] INFO c.l.m.boot.steps.IngestPoliciesStep:93 - Skipping ingestion of editable policy with urn urn:li:dataHubPolicy:view-dataset-sensitive
22:30:34.870 [main] INFO c.l.m.boot.steps.IngestPoliciesStep:85 - Ingesting default policy with urn urn:li:dataHubPolicy:admin-platform-policy
22:30:35.494 [main] INFO c.l.m.boot.steps.IngestPoliciesStep:85 - Ingesting default policy with urn urn:li:dataHubPolicy:admin-metadata-policy
22:30:36.128 [main] INFO c.l.m.boot.steps.IngestPoliciesStep:85 - Ingesting default policy with urn urn:li:dataHubPolicy:editor-platform-policy
22:30:36.757 [main] INFO c.l.m.boot.steps.IngestPoliciesStep:85 - Ingesting default policy with urn urn:li:dataHubPolicy:editor-metadata-policy
22:30:37.384 [main] INFO c.l.m.boot.steps.IngestPoliciesStep:85 - Ingesting default policy with urn urn:li:dataHubPolicy:reader-platform-policy
22:30:38.004 [main] INFO c.l.m.boot.steps.IngestPoliciesStep:85 - Ingesting default policy with urn urn:li:dataHubPolicy:reader-metadata-policy
22:30:38.792 [main] INFO c.l.m.boot.steps.IngestPoliciesStep:93 - Skipping ingestion of editable policy with urn urn:li:dataHubPolicy:asset-owners-metadata-policy
22:30:38.835 [main] WARN org.elasticsearch.client.RestClient:65 - request [POST <https://99b1f46f43124c9dbfe6a6de8cee78e8.psc.us-central1.gcp.cloud.es.io:9243/datahubpolicyindex_v2/_count?ignore_throttled=true&ignore_unavailable=false&expand_wildcards=open&allow_no_indices=true>] returned 1 warnings: [299 Elasticsearch-7.17.8-120eabe1c8a0cb2ae87cffc109a5b65d213e9df1 "[ignore_throttled] parameter is deprecated because frozen indices have been deprecated. Consider cold or frozen tiers in place of frozen indices."]
22:30:38.838 [main] INFO c.l.m.boot.steps.IngestPoliciesStep:102 - Successfully ingested default access policies.
22:30:38.838 [main] INFO c.l.metadata.boot.BootstrapManager:41 - Starting asynchronous bootstrap step 3/10 with name IngestRolesStep...
22:30:38.839 [main] INFO c.l.metadata.boot.BootstrapManager:33 - Executing bootstrap step 4/10 with name IngestDataPlatformsStep...
22:30:47.135 [main] INFO c.l.metadata.boot.BootstrapManager:41 - Starting asynchronous bootstrap step 5/10 with name IngestDataPlatformInstancesStep...
22:30:47.136 [main] INFO c.l.metadata.boot.BootstrapManager:41 - Starting asynchronous bootstrap step 6/10 with name IngestRetentionPoliciesStep...
22:30:47.136 [pool-10-thread-2] INFO c.l.m.b.s.IngestDataPlatformInstancesStep:51 - Checking for DataPlatformInstance
22:30:47.137 [pool-10-thread-3] INFO c.l.m.b.s.IngestRetentionPoliciesStep:48 - Ingesting default retention...
22:30:47.136 [main] INFO c.l.metadata.boot.BootstrapManager:41 - Starting asynchronous bootstrap step 7/10 with name RestoreGlossaryIndices...
22:30:47.138 [main] INFO c.l.metadata.boot.BootstrapManager:41 - Starting asynchronous bootstrap step 8/10 with name RemoveClientIdAspectStep...
22:30:47.138 [pool-10-thread-4] INFO c.linkedin.metadata.boot.UpgradeStep:42 - Attempting to run RestoreGlossaryIndices Upgrade Step..
22:30:47.139 [pool-10-thread-4] INFO c.linkedin.metadata.boot.UpgradeStep:43 - Waiting 120 seconds..
22:30:47.139 [main] INFO c.l.metadata.boot.BootstrapManager:41 - Starting asynchronous bootstrap step 9/10 with name RestoreDbtSiblingsIndices...
22:30:47.139 [main] INFO c.l.metadata.boot.BootstrapManager:41 - Starting asynchronous bootstrap step 10/10 with name IndexDataPlatformsStep...
22:30:47.140 [pool-10-thread-3] INFO c.l.m.b.s.IngestRetentionPoliciesStep:64 - Setting 2 policies
22:30:47.143 [main] INFO o.s.web.context.ContextLoader:307 - Root WebApplicationContext initialized in 32668 ms
22:30:47.149 [main] INFO c.d.a.filter.AuthenticationFilter:175 - Auth is disabled. Building no-op authenticator chain...
2023-01-15 22:30:47.196:INFO:oejshC.ROOT:main: Initializing Spring DispatcherServlet 'apiServlet'
22:30:47.196 [main] INFO o.s.web.servlet.DispatcherServlet:525 - Initializing Servlet 'apiServlet'
22:30:47.637 [main] INFO o.s.web.servlet.DispatcherServlet:547 - Completed initialization in 441 ms
2023-01-15 22:30:47.638:INFO:oejshC.ROOT:main: Initializing Spring DispatcherServlet 'authServlet'
22:30:47.638 [main] INFO o.s.web.servlet.DispatcherServlet:525 - Initializing Servlet 'authServlet'
22:30:47.695 [main] INFO o.s.web.servlet.DispatcherServlet:547 - Completed initialization in 57 ms
2023-01-15 22:30:47.695:INFO:oejshC.ROOT:main: Initializing Spring DispatcherServlet 'openapiServlet'
22:30:47.695 [main] INFO o.s.web.servlet.DispatcherServlet:525 - Initializing Servlet 'openapiServlet'
22:30:48.062 [pool-10-thread-5] INFO c.l.m.b.s.RemoveClientIdAspectStep:43 - Unknown aspects have been removed. Skipping...
22:30:48.064 [pool-10-thread-5] INFO c.l.m.b.s.RestoreDbtSiblingsIndices:61 - Attempting to run RestoreDbtSiblingsIndices upgrade..
22:30:48.065 [pool-10-thread-5] INFO c.l.m.b.s.RestoreDbtSiblingsIndices:62 - Waiting 120 seconds..
22:30:48.225 [pool-10-thread-2] INFO c.l.m.b.s.IngestDataPlatformInstancesStep:61 - Reading urns 0 to 1000 from the aspects table to generate dataplatform instance aspects
22:30:48.319 [pool-10-thread-2] INFO c.l.m.b.s.IngestDataPlatformInstancesStep:76 - Finished ingesting DataPlatformInstance for urn 0 to 1000
22:30:48.319 [pool-10-thread-2] INFO c.l.m.b.s.IngestDataPlatformInstancesStep:79 - Finished ingesting DataPlatformInstance for all entities
22:30:48.320 [pool-10-thread-2] INFO c.linkedin.metadata.boot.UpgradeStep:42 - Attempting to run IndexDataPlatformsStep Upgrade Step..
22:30:48.320 [pool-10-thread-2] INFO c.linkedin.metadata.boot.UpgradeStep:43 - Waiting 120 seconds..
22:30:49.089 [main] INFO o.s.web.servlet.DispatcherServlet:547 - Completed initialization in 1393 ms
2023-01-15 22:30:49.089:INFO:oejsh.ContextHandler:main: Started o.e.j.w.WebAppContext@6eda5c9{Open source GMS,/,[file:///tmp/jetty-0_0_0_0-8080-war_war-_-any-8600198876037324782/webapp/, jar:file:///tmp/jetty-0_0_0_0-8080-war_war-_-any-8600198876037324782/webapp/WEB-INF/lib/swagger-ui-4.10.3.jar!/META-INF/resources],AVAILABLE}{file:///datahub/datahub-gms/bin/war.war}
2023-01-15 22:30:49.106:INFO:oejs.AbstractConnector:main: Started ServerConnector@4387b79e{HTTP/1.1, (http/1.1)}{0.0.0.0:8080}
2023-01-15 22:30:49.107:INFO:oejs.Server:main: Started @61430ms
red-waitress-53338
01/15/2023, 10:41 PM22:30:49.907 [pool-10-thread-3] ERROR i.c.k.s.client.rest.RestService:267 - Failed to send HTTP request to endpoint: <https://schemaregistry.xxxxx.com:443/subjects/a54808-preprod-MetadataChangeLog_Versioned_v1-value/versions>
javax.net.ssl.SSLHandshakeException: PKIX path building failed: sun.security.provider.certpath.SunCertPathBuilderException: unable to find valid certification path to requested target
at java.base/sun.security.ssl.Alert.createSSLException(Alert.java:131)
at java.base/sun.security.ssl.TransportContext.fatal(TransportContext.java:353)
at java.base/sun.security.ssl.TransportContext.fatal(TransportContext.java:296)
at java.base/sun.security.ssl.TransportContext.fatal(TransportContext.java:291)
at java.base/sun.security.ssl.CertificateMessage$T12CertificateConsumer.checkServerCerts(CertificateMessage.java:654)
at java.base/sun.security.ssl.CertificateMessage$T12CertificateConsumer.onCertificate(CertificateMessage.java:473)
at java.base/sun.security.ssl.CertificateMessage$T12CertificateConsumer.consume(CertificateMessage.java:369)
at java.base/sun.security.ssl.SSLHandshake.consume(SSLHandshake.java:392)
at java.base/sun.security.ssl.HandshakeContext.dispatch(HandshakeContext.java:443)
at java.base/sun.security.ssl.HandshakeContext.dispatch(HandshakeContext.java:421)
at java.base/sun.security.ssl.TransportContext.dispatch(TransportContext.java:183)
at java.base/sun.security.ssl.SSLTransport.decode(SSLTransport.java:172)
at java.base/sun.security.ssl.SSLSocketImpl.decode(SSLSocketImpl.java:1506)
at java.base/sun.security.ssl.SSLSocketImpl.readHandshakeRecord(SSLSocketImpl.java:1416)
at java.base/sun.security.ssl.SSLSocketImpl.startHandshake(SSLSocketImpl.java:456)
at java.base/sun.security.ssl.SSLSocketImpl.startHandshake(SSLSocketImpl.java:427)
at java.base/sun.net.www.protocol.https.HttpsClient.afterConnect(HttpsClient.java:572)
at java.base/sun.net.www.protocol.https.AbstractDelegateHttpsURLConnection.connect(AbstractDelegateHttpsURLConnection.java:201)
at java.base/sun.net.www.protocol.http.HttpURLConnection.getOutputStream0(HttpURLConnection.java:1367)
at java.base/sun.net.www.protocol.http.HttpURLConnection.getOutputStream(HttpURLConnection.java:1342)
at java.base/sun.net.www.protocol.https.HttpsURLConnectionImpl.getOutputStream(HttpsURLConnectionImpl.java:246)
at io.confluent.kafka.schemaregistry.client.rest.RestService.sendHttpRequest(RestService.java:263)
at io.confluent.kafka.schemaregistry.client.rest.RestService.httpRequest(RestService.java:351)
at io.confluent.kafka.schemaregistry.client.rest.RestService.registerSchema(RestService.java:494)
at io.confluent.kafka.schemaregistry.client.rest.RestService.registerSchema(RestService.java:485)
at io.confluent.kafka.schemaregistry.client.rest.RestService.registerSchema(RestService.java:458)
at io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.registerAndGetId(CachedSchemaRegistryClient.java:206)
at io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.register(CachedSchemaRegistryClient.java:268)
at io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.register(CachedSchemaRegistryClient.java:244)
at io.confluent.kafka.serializers.AbstractKafkaAvroSerializer.serializeImpl(AbstractKafkaAvroSerializer.java:74)
at io.confluent.kafka.serializers.KafkaAvroSerializer.serialize(KafkaAvroSerializer.java:59)
at org.apache.kafka.common.serialization.Serializer.serialize(Serializer.java:62)
at org.apache.kafka.clients.producer.KafkaProducer.doSend(KafkaProducer.java:902)
at org.apache.kafka.clients.producer.KafkaProducer.send(KafkaProducer.java:862)
at com.linkedin.metadata.dao.producer.KafkaEventProducer.produceMetadataChangeLog(KafkaEventProducer.java:145)
at com.linkedin.metadata.entity.EntityService.produceMetadataChangeLog(EntityService.java:1284)
at com.linkedin.metadata.entity.EntityService.emitChangeLog(EntityService.java:1049)
at com.linkedin.metadata.entity.EntityService.ingestProposal(EntityService.java:893)
at com.linkedin.metadata.entity.RetentionService.setRetention(RetentionService.java:113)
at com.linkedin.metadata.boot.steps.IngestRetentionPoliciesStep.execute(IngestRetentionPoliciesStep.java:67)
at com.linkedin.metadata.boot.BootstrapManager.lambda$start$0(BootstrapManager.java:44)
at java.base/java.util.concurrent.CompletableFuture$AsyncRun.run(CompletableFuture.java:1736)
at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
at java.base/java.lang.Thread.run(Thread.java:829)
Caused by: sun.security.validator.ValidatorException: PKIX path building failed: sun.security.provider.certpath.SunCertPathBuilderException: unable to find valid certification path to requested target
at java.base/sun.security.validator.PKIXValidator.doBuild(PKIXValidator.java:439)
at java.base/sun.security.validator.PKIXValidator.engineValidate(PKIXValidator.java:306)
at java.base/sun.security.validator.Validator.validate(Validator.java:264)
at java.base/sun.security.ssl.X509TrustManagerImpl.validate(X509TrustManagerImpl.java:313)
at java.base/sun.security.ssl.X509TrustManagerImpl.checkTrusted(X509TrustManagerImpl.java:222)
at java.base/sun.security.ssl.X509TrustManagerImpl.checkServerTrusted(X509TrustManagerImpl.java:129)
at java.base/sun.security.ssl.CertificateMessage$T12CertificateConsumer.checkServerCerts(CertificateMessage.java:638)
... 40 common frames omitted
Caused by: sun.security.provider.certpath.SunCertPathBuilderException: unable to find valid certification path to requested target
at java.base/sun.security.provider.certpath.SunCertPathBuilder.build(SunCertPathBuilder.java:141)
at java.base/sun.security.provider.certpath.SunCertPathBuilder.engineBuild(SunCertPathBuilder.java:126)
at java.base/java.security.cert.CertPathBuilder.build(CertPathBuilder.java:297)
at java.base/sun.security.validator.PKIXValidator.doBuild(PKIXValidator.java:434)
... 46 common frames omitted
22:30:49.908 [pool-10-thread-3] ERROR c.l.metadata.boot.BootstrapManager:46 - Caught exception while executing bootstrap step IngestRetentionPoliciesStep. Continuing...
org.apache.kafka.common.errors.SerializationException: Error serializing Avro message
Caused by: javax.net.ssl.SSLHandshakeException: PKIX path building failed: sun.security.provider.certpath.SunCertPathBuilderException: unable to find valid certification path to requested target
at java.base/sun.security.ssl.Alert.createSSLException(Alert.java:131)
at java.base/sun.security.ssl.TransportContext.fatal(TransportContext.java:353)
at java.base/sun.security.ssl.TransportContext.fatal(TransportContext.java:296)
at java.base/sun.security.ssl.TransportContext.fatal(TransportContext.java:291)
at java.base/sun.security.ssl.CertificateMessage$T12CertificateConsumer.checkServerCerts(CertificateMessage.java:654)
at java.base/sun.security.ssl.CertificateMessage$T12CertificateConsumer.onCertificate(CertificateMessage.java:473)
at java.base/sun.security.ssl.CertificateMessage$T12CertificateConsumer.consume(CertificateMessage.java:369)
at java.base/sun.security.ssl.SSLHandshake.consume(SSLHandshake.java:392)
at java.base/sun.security.ssl.HandshakeContext.dispatch(HandshakeContext.java:443)
at java.base/sun.security.ssl.HandshakeContext.dispatch(HandshakeContext.java:421)
at java.base/sun.security.ssl.TransportContext.dispatch(TransportContext.java:183)
at java.base/sun.security.ssl.SSLTransport.decode(SSLTransport.java:172)
at java.base/sun.security.ssl.SSLSocketImpl.decode(SSLSocketImpl.java:1506)
at java.base/sun.security.ssl.SSLSocketImpl.readHandshakeRecord(SSLSocketImpl.java:1416)
at java.base/sun.security.ssl.SSLSocketImpl.startHandshake(SSLSocketImpl.java:456)
at java.base/sun.security.ssl.SSLSocketImpl.startHandshake(SSLSocketImpl.java:427)
at java.base/sun.net.www.protocol.https.HttpsClient.afterConnect(HttpsClient.java:572)
at java.base/sun.net.www.protocol.https.AbstractDelegateHttpsURLConnection.connect(AbstractDelegateHttpsURLConnection.java:201)
at java.base/sun.net.www.protocol.http.HttpURLConnection.getOutputStream0(HttpURLConnection.java:1367)
at java.base/sun.net.www.protocol.http.HttpURLConnection.getOutputStream(HttpURLConnection.java:1342)
at java.base/sun.net.www.protocol.https.HttpsURLConnectionImpl.getOutputStream(HttpsURLConnectionImpl.java:246)
at io.confluent.kafka.schemaregistry.client.rest.RestService.sendHttpRequest(RestService.java:263)
at io.confluent.kafka.schemaregistry.client.rest.RestService.httpRequest(RestService.java:351)
at io.confluent.kafka.schemaregistry.client.rest.RestService.registerSchema(RestService.java:494)
at io.confluent.kafka.schemaregistry.client.rest.RestService.registerSchema(RestService.java:485)
at io.confluent.kafka.schemaregistry.client.rest.RestService.registerSchema(RestService.java:458)
at io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.registerAndGetId(CachedSchemaRegistryClient.java:206)
at io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.register(CachedSchemaRegistryClient.java:268)
at io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.register(CachedSchemaRegistryClient.java:244)
at io.confluent.kafka.serializers.AbstractKafkaAvroSerializer.serializeImpl(AbstractKafkaAvroSerializer.java:74)
at io.confluent.kafka.serializers.KafkaAvroSerializer.serialize(KafkaAvroSerializer.java:59)
at org.apache.kafka.common.serialization.Serializer.serialize(Serializer.java:62)
at org.apache.kafka.clients.producer.KafkaProducer.doSend(KafkaProducer.java:902)
at org.apache.kafka.clients.producer.KafkaProducer.send(KafkaProducer.java:862)
at com.linkedin.metadata.dao.producer.KafkaEventProducer.produceMetadataChangeLog(KafkaEventProducer.java:145)
at com.linkedin.metadata.entity.EntityService.produceMetadataChangeLog(EntityService.java:1284)
at com.linkedin.metadata.entity.EntityService.emitChangeLog(EntityService.java:1049)
at com.linkedin.metadata.entity.EntityService.ingestProposal(EntityService.java:893)
at com.linkedin.metadata.entity.RetentionService.setRetention(RetentionService.java:113)
at com.linkedin.metadata.boot.steps.IngestRetentionPoliciesStep.execute(IngestRetentionPoliciesStep.java:67)
at com.linkedin.metadata.boot.BootstrapManager.lambda$start$0(BootstrapManager.java:44)
at java.base/java.util.concurrent.CompletableFuture$AsyncRun.run(CompletableFuture.java:1736)
at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
at java.base/java.lang.Thread.run(Thread.java:829)
quick-student-61408
01/16/2023, 1:53 PMmicroscopic-mechanic-13766
01/16/2023, 4:16 PMpip install kerberos
pip install thrift==0.13.0
pip install great-expectations==0.15.43
I don't know if it would worker with a lower GE version, but I am sure that it has to be a version higher than 0.15.2 (that, if I am not mistaken is the version installed in such module by default).
Then you would have to select the other
option when creating the datasource (see image).
Once the datasource was created, the connection_string looked like this in my case --> hive+<http://hiveserver1:10001/default?auth=KERBEROS&kerberos_service_name=hive-server>
Before executing the datasource, you would have to make a kinit of a user in order to be able to obtain the TGT of your hive service.
The rest is really similar to creating any other type of validation tests, so I will not go into detail.polite-actor-701
01/17/2023, 5:36 AMbetter-orange-49102
01/17/2023, 10:21 AMrefined-tent-35319
01/17/2023, 11:22 AMboundless-piano-94348
01/17/2023, 3:15 PMInput and Output
1. Ensure that input data is strongly typed, validated, range or length checked, or at worst, sanitized or filtered.
2. Ensure that validation happen not only on client side but also on the corresponding server side.
3. Ensure that output encoding happen in the interpreter. This can be handled by the framework or manually.
4. Ensure that database query is protected against SQL injection
Cryptography
1. Ensure that any cryptograpic keys used by the system are documented: key purpose, where it's stored, and key specifications.
2. Ensure that any cryptograpic keys stored in a secured secret manager.
3. Ensure that database encryption is enabled.
4. When using any cryptographic algorithm, ensure that using secured and appropriate algorithm/keys.
Errors, Logging, and Auditing
1. Ensure that application use JSON format & send log to cloud logging.
2. Ensure that PII data, authentication data, sensisive financial data is not being logged.
3. Ensure that authentication events are logged properly (upon success, failed, and changes of authentication method such as password reset).
4. Ensure that there is log event for sensitive transactions or admin actions.
5. Ensure that event log has information on user who perform the action, time of event, type of event, source of event (IP client).
6. Ensure that the application return a useful and generic error message to user instead of the verbose application error.
7. Ensure that application fail securely. There is a "last resort" error handler to catch all unhandled exception.
red-waitress-53338
01/18/2023, 12:01 AMgreat-monkey-52307
01/18/2023, 5:32 AMfast-ice-59096
01/18/2023, 11:04 AM