Latent reasoning models perform chain-of-thought processing within a model's latent space by bypassing the language model head. Current versions remain small, mostly GPT-2 scale, and limited to narrow tasks. This architecture potentially simplifies the analysis of internal model logic. Researchers are weighing these benefits against the risks of hidden, non-human-readable reasoning.