This message was deleted.
# ask-for-help
s
This message was deleted.
c
that looks right! it should create two endpoints
/predict
and
/predict_2
🙏 1
s
@Chaoyu Thanks! How do I verify if I setup adaptive batching correctly in the code?
c
@Shihgian Lee do you have
batchable: True
for the runner? If so, it should do adaptive batching when running with
--production
flag
we recently added batching related metrics too, you can view batch size histogram from a grafana UI too https://github.com/bentoml/BentoML/pull/2902
Note that batching only happens at the runner level now, not the API level, if you need to batch custom code/operations, you can do that via custom runners
s
Hi @Chaoyu Yes, I have batchable set to true in the runner. The adaptive batching for BentoML 1.0 is very confusing for me. I am still working it out with Sean. Can you take a look at this thread when you get a chance? Sean might be tied up with work and Kubeflow summit. This is the last step that is blocking me from deploying to Kubernetes cluster for load testing.
(removed due to posting to the wrong thread)