Slackbot
06/05/2023, 1:57 PMYilun Zhang
06/05/2023, 2:03 PMtag.path
and load the tokenizer first:
tag = bentoml.transformers.get(model_name)
tokenizer = transformers.AutoTokenizer.from_pretrained(tag.path)
pipeline = bentoml.transformers.load_model(tag, trust_remote_code=True, torch_dtype=torch.bfloat16, tokenizer=tokenizer)
Is this the desired way of doing this?
I’m suspecting it might have something to do with how the pipeline/config is setup for this falcon model, and thus causing some inconsistency with bentoml backend.Aaron Pham
06/05/2023, 6:49 PMYilun Zhang
06/06/2023, 9:07 AMYilun Zhang
06/06/2023, 12:25 PM