Reasoning language model

Reasoning language models (RLMs) are large language models that have been further trained to solve multi-step reasoning tasks.[1] These models perform better on logical, mathematical or programmatic tasks than traditional autoregressive LLMs, have the ability to backtrack, and employ test-time compute as an additional scaling axis beyond training examples, parameter count, and train-time compute.

  1. ^ Besta, Maciej (2025-01-23). "Reasoning Language Models: A Blueprint". arXiv:2501.11223 [cs.CL].

© MMXXIII Rich X Search. We shall prevail. All rights reserved. Rich X Search