What is a Token?

In the context of AI, particularly in multimodal large language models (LLMs), a token refers to the basic unit of information used for processing and representation. It's analogous to how words are the building blocks of sentences in human language. When engaging with a model it's responses or usage is measured by the number of tokens required.

So the more tokens the gen-AI model can process the better? Not necessarily. Two models with similar token processing capacity might demonstrate vastly different capabilities if their architectures, training data, and optimization differ significantly. Here is a breakdown:

General AI:

Multimodal LLMs:

Here are some additional points to consider: