volexbpe is a low latency high throughput Byte-Pair encoding derived tokenizer providng exceptional performance & streamline interface.
- o200k_base - used in o3, o1, gpt-4o.
- cl100k_base - used in gpt-4, gpt-3.5 turbo, gpt-3.5, most openai text embedding endpoints.
- r50k_base - majority decreptad.
- gpt-2 - gpt-2, open source
pip install veloxbpe
git clone https://github.com/friedhar/veloxbpe.git
maturin develop
All benchmarks can be run locally. After you've built from source, run
uv run bench/benchmark_bandwidth_0.py
- Add support for custom BPE training.