Tokenization breaks raw text into smaller units like words or subwords to enable LLM processing. This foundational step converts human language into numerical representations that machines can compute. While basic, the choice between word-level and byte-pair encoding affects model efficiency. Practitioners must optimize this process to reduce latency and improve token density.