Reasoning language models (RLMs) are large language models that have been further trained to solve multi-step reasoning tasks. These models perform better on logical, mathematical or programmatic tasks than traditional autoregressive LLMs, have the ability to backtrack, and employ test-time compute as an additional scaling axis beyond training examples, parameter count, and train-time compute.