Context length / window. Think of them and the "number of words" that the model can effectively process. 1 token is roughly equal to 4 characters or 0.75 words for English text. The number of tokens is the total number that can fit into a context window, which again is the space of "input" i.e. prompts and output (response/ completions) that the model can handle