RLM: Recursive Language Model
The RLM architecture is designed for “thinking” at scale. It uses a recursive loop structure to handle inputs that exceed standard context windows or require multiple rounds of deliberation.Using RLM Zero-Shot
You can use the RLM architecture immediately by targeting thezero_shot_rlm_single graph ID. This allows you to provide a massive context and a complex system prompt without any prior optimization.
RLM for Standalone Reasoning (Policy)
Use RLM when you have a high-complexity task that requires deep reasoning but you don’t need a verifier score.RLM as a Verifier
RLM can also be used as a high-fidelity judge for verification tasks. It is the default architecture used when traces exceed 500k tokens or 100+ events. However, RLMs aren’t only for verification—they’re a general-purpose architecture for handling massive context and complex recursive reasoning.Why use RLM?
- Unlimited Context: Automatically handles massive inputs that would crash a standard single-prompt LLM.
- Recursive Auditing: RLM nodes can check each other’s work, significantly reducing hallucinations in long-context tasks.
- Optimizable: While
zero_shot_rlmworks out of the box, you can use Graph Gen to optimize the RLM recursive prompts for your specific domain.