Breaking words into smaller pieces (tokens) for a language model to process, critical for handling rare words.