This message was deleted.
# ask-for-help
s
This message was deleted.
🦄 1
s
Hi Shaohong, some of the team is on the Lunar New Year break. So please expect some delays in responses. In your Yatai setup, is it a new setup or an upgrade from 1.0?
s
Ahhh, no worries. Take time, it is a new step up given recent launch
Happy new year!
x
Hi, thanks for your feedback. I need you to provide some information to help troubleshoot, first you should provide the version of yatai you have installed:
Copy code
helm list -A
You will also be asked to provide the following role information:
Copy code
kubectl -n yatai-system get role yatai-deployment-in-yatai-system -o yaml | grep -i secrets
s
versions
this looks no seretes right?
x
It looks like the role of secrets exists, can you provide the logs of yatai-deployment again?
Copy code
kubectl -n yatai-deployment logs deploy/yatai-deployment --tail 200
s
1.6745654232867305e+09 INFO start cleaning up abandoned runner services {“func”: “doCleanUpAbandonedRunnerServices”} 1.6745654232907972e+09 INFO finished cleaning up abandoned runner services {“func”: “doCleanUpAbandonedRunnerServices”} W0124 130410.157654 1 warnings.go:70] autoscaling/v2beta2 HorizontalPodAutoscaler is deprecated in v1.23+, unavailable in v1.26+; use autoscaling/v2 HorizontalPodAutoscaler 1.674565453286925e+09 INFO start cleaning up abandoned runner services {“func”: “doCleanUpAbandonedRunnerServices”} 1.6745654532937553e+09 INFO finished cleaning up abandoned runner services {“func”: “doCleanUpAbandonedRunnerServices”} 1.674565483286653e+09 INFO start cleaning up abandoned runner services {“func”: “doCleanUpAbandonedRunnerServices”} 1.6745654832948124e+09 INFO finished cleaning up abandoned runner services {“func”: “doCleanUpAbandonedRunnerServices”} 1.6745655132864134e+09 INFO start cleaning up abandoned runner services {“func”: “doCleanUpAbandonedRunnerServices”} 1.6745655132938392e+09 INFO finished cleaning up abandoned runner services {“func”: “doCleanUpAbandonedRunnerServices”} 1.6745655432862902e+09 INFO start cleaning up abandoned runner services {“func”: “doCleanUpAbandonedRunnerServices”} 1.6745655432985623e+09 INFO finished cleaning up abandoned runner services {“func”: “doCleanUpAbandonedRunnerServices”} 1.6745655586679792e+09 ERROR Failed to reconcile BentoDeployment. {“controller”: “bentodeployment”, “controllerGroup”: “serving.yatai.ai”, “controllerKind”: “BentoDeployment”, “BentoDeployment”: {“name”“adthena iris test”,“namespace”“yatai”}, “namespace”: “yatai”, “name”: “adthena-iris-test”, “reconcileID”: “24a61e0d-8eed-4f51-a3f9-e876ed86881e”, “bentoDeployment”: “adthena-iris-test”, “namespace”: “yatai”, “error”: “get domain suffix: failed to wait for ingress default-domain-rht49 to be ready: timed out waiting for the condition”, “errorVerbose”: “timed out waiting for the condition\nfailed to wait for ingress default-domain-rht49 to be ready\ngithub.com/bentoml/yatai-common/system.GetIngressIP\n\t/go/pkg/mod/github.com/bentoml/yatai-common@v0.0.0-20230110044251-1eb442339c38/system/ingress.go:156\ngithub.com/bentoml/yatai-common/system.GetDomainSuffix\n\t/go/pkg/mod/github.com/bentoml/yatai-common@v0.0.0-20230110044251-1eb442339c38/system/ingress.go:198\ngithub.com/bentoml/yatai-deployment/controllers.(*BentoDeploymentReconciler).generateDefaultHostname\n\t/workspace/controllers/bentodeployment_controller.go:2644\ngithub.com/bentoml/yatai-deployment/controllers.(*BentoDeploymentReconciler).generateIngressHost\n\t/workspace/controllers/bentodeployment_controller.go:2634\ngithub.com/bentoml/yatai-deployment/controllers.(*BentoDeploymentReconciler).generateIngresses\n\t/workspace/controllers/bentodeployment_controller.go:2719\ngithub.com/bentoml/yatai-deployment/controllers.(*BentoDeploymentReconciler).createOrUpdateIngresses\n\t/workspace/controllers/bentodeployment_controller.go:1152\ngithub.com/bentoml/yatai-deployment/controllers.(*BentoDeploymentReconciler).Reconcile\n\t/workspace/controllers/bentodeployment_controller.go:403\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.13.0/pkg/internal/controller/controller.go:121\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.13.0/pkg/internal/controller/controller.go:320\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.13.0/pkg/internal/controller/controller.go:273\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/go/pkg/mod/sigs.k8s.io/controller-runtime@v0.13.0/pkg/internal/controller/controller.go234\nruntime.goexit\n\t/usr/local/go/src/runtime/asm amd64.s1594\nget domain suffix\ngithub.com/bentoml/yatai-deployment/controllers.(*BentoDeploymentReconciler).generateDefaultHostname\n\t/workspace/controllers/bentodeployment_controller.go:2646\ngithub.com/bentoml/yatai-deployment/controllers.(*BentoDeploymentReconciler).generateIngressHost\n\t/workspace/controllers/bentodeployment_controller.go:2634\ngithub.com/bentoml/yatai-deployment/controllers.(*BentoDeploymentReconciler).generateIngresses\n\t/workspace/controllers/bentodeployment_controller.go:2719\ngithub.com/bentoml/yatai-deployment/controllers.(*BentoDeploymentReconciler).createOrUpdateIngresses\n\t/workspace/controllers/bentodeployment_controller.go:1152\ngithub.com/bentoml/yatai-deployment/controllers.(*BentoDeploymentReconciler).Reconcile\n\t/workspace/controllers/bentodeployment_controller.go:403\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/go/pkg/mod/sigs.k8s.io/controller-
basically keep doing this and in between have that erro message
x
It doesn't seem to have the error you posted
s
i think because the cleaning up thing still running, i can try in stall deployment again and give u log
that is not a problem
Found 2 pods, using pod/yatai-deployment-default-domain-fg7px time=“2023-01-24T150654Z” level=info msg=“Creating ingress default-domain- to get a ingress IP automatically” time=“2023-01-24T150654Z” level=info msg=“Waiting for ingress default-domain-64ntf to be ready” looks much clear now
it is odd no such error today, but still time out
do u think it is ingress config issue?
there are more coming afterwards
s
enjoy your holiday maybe, we discuss when you come back, I was following that installation, no luck so far, bash <(curl -s “https://raw.githubusercontent.com/bentoml/yatai-deployment/main/scripts/quick-install-yatai-deployment.sh”), I will have a look
x
This document also applies to this type of installation
s
yeah i assume so, just something odd for the domain bit, i will have to install part to part to debug, will update u guys