Learning-Deep-Learning

Coconut: Training Large Language Models to Reason in a Continuous Latent Space

January 2026

tl;dr: Reason in unconstraint latent space vs language space to bypass the “Vocabulary Bottleneck”.

Overall impression

Coconut allows more efficient reasoning, reducing token num and also induces new thining patterns.

The latent thinking requires a normal training with explicit CoT first. It creates a latent thinking stage marked by and , and iteratively repalces each step of explicit thinking (1 thinking step ≈ a short sentence or clause, x language tokens) with c latent thought tokens.

One possible drawback is that this cannot decode into human language and is hard to visualize. Maybe DLCM is a better method.

Key ideas

Technical details

Notes