Join Slack
Powered by
This message was deleted.
# ask-for-help
s
Slackbot
01/11/2023, 6:58 AM
This message was deleted.
c
Chaoyu
01/11/2023, 7:05 AM
Yes BentoML will initialize the model in memory for online inference requests and minimize latency. Is it because you want to host many models and unload the models not actively being used?
3
Views
Open in Slack
Previous
Next