Including tokenizer to onnx model / basic usage of the onnxruntime-extensions #798
Open
Description
Hi and thanks for this great library!
Iam very new to onnx and Iam trying to include the Roberta tokenizer into a Roberta onnx model.
As far as I have understood, one can get the onnx graph for the tokenizer using:
import onnxruntime as _ort
from transformers import RobertaTokenizer
from onnxruntime_extensions import OrtPyFunction, gen_processing_models
# Roberta tokenizer
tokenizer = AutoTokenizer.from_pretrained("roberta-base", model_max_length=512)
tokenizer_onnx = OrtPyFunction(gen_processing_models(spm_hf_tokenizer, pre_kwargs={})[0])
Now Iam wondering what the next step is? How can I combine the onnx tokenizer (or graph) with a model?
Thanks in advance for any help,
cheers,
M
Metadata
Assignees
Labels
No labels