Skip to content

Latest commit

 

History

History
31 lines (16 loc) · 1.1 KB

readme.md

File metadata and controls

31 lines (16 loc) · 1.1 KB

Context Length Visualized - How much is 128k tokens

To visualize how big the context windows are for the current LLMs I currated a list of different token lengths.

Context Length Terminology

Since we started intially with a context length of 4096 I assumed that the follow up context lengths are multiples of 4096.

But to keep it simple the big companies shortend it to 32k, 64k and 128k respectivelly.

However I kept it to the multiples of 4096 and added 128k tokens.

Tokenizer

The tokenizer used is tiktoken from OpenAI

Every company is using their own tokenizer but in principal they are all working the same

The tokenizer base for those files is o200k_base which means that the vocabulary used to create those tokens is 200k big!

Tokenizer base file

I tried to extract all tokens from the tiktoken tokenizer

It holds the tokens in text format o200k_base.txt

I had troubles to decode it. Some decodes didn't work properly.

Does someone knows how to decode it propperly?

You can run the decode with python main.py