Tokens were considered individual words or 3 to 4 characters, but it’s false.

Tokens can be individual or partial words, as seen in the above image.

Large Language Models use tokens to measure 3 things →

https://media.beehiiv.com/cdn-cgi/image/fit=scale-down,format=auto,onerror=redirect,quality=80/uploads/asset/file/c5376152-d7cb-41d4-a536-2515a16393d4/to.png?t=1714804988

OpenAI tokenizer - Himanshu Ramchandani

The tokens will be converted into numeric embeddings, as all types of models process numbers only.

The GPT was trained on more than 500 billion tokens.

The GPT was trained on 175 billion parameters.

Both the statements are true.