tiktokenizer 1.0.0
Install from the command line:
Learn more about npm packages
$ npm install @microsoft/tiktokenizer@1.0.0
Install via package.json:
"@microsoft/tiktokenizer": "1.0.0"
About this version
This is a Typescript implementation of byte pair encoding(BPE) tokenizer for OpenAI LLMs, it's based on open sourced rust implementation in the OpenAI tiktoken. It's valuable to run prompt tokenization in Nodejs or web browser before feeding prompt into a LLM.
Install the npm package in your project:
npm install @microsoft/tiktokenizer
Example Typescript code to use @microsoft/tiktokenizer in your code. In production setting, you should pre-download the BPE rank file and call createTokenizer
API to avoid downloading the BPE rank file on the fly.
import {
createByModelName
} from '@microsoft/tiktokenizer';
const IM_START = "<|im_start|>";
const IM_END = "<|im_end|>";
const specialTokens: ReadonlyMap<string, number> = new Map([
[IM_START, 100264],
[IM_END, 100265],
]);
const str = "<|im_start|>Hello World<|im_end|>";
let tokenizer = null;
const createTokenizer = async () => {
tokenizer = await createByModelName("gpt-3.5-turbo", specialTokens);
var out = tokenizer.encodeTrimSuffix(str, 3, Array.from(specialTokens.keys()));
console.log(out);
}
createTokenizer();
We welcome contributions. Please follow this guideline.
This project may contain trademarks or logos for projects, products, or services. Authorized use of Microsoft trademarks or logos is subject to and must follow Microsoft's Trademark & Brand Guidelines. Use of Microsoft trademarks or logos in modified versions of this project must not cause confusion or imply Microsoft sponsorship. Any use of third-party trademarks or logos are subject to those third-party's policies.