Tokenization
Tokenization is a method of tagging the words in a corpus or dataset with numbers. The total tokenized collection is called a word index, which is a dictionary mapping words to numbers.
In ML/DL models, tokenization is an essential processing techniques used for preprocessing datatsets as learning models generally deal with models that take numerical input.