There is speculation that Anthropic’s Claude Mythos is a Looped Language Model
Paper: https://arxiv.org/abs/2510.25741
Claude Mythos hardly needs an introduction at this point, since so many people already know about it. What is less familiar is the idea of a Looped Language Model, a concept proposed by the ByteDance team in a paper published in late 2025. That paper argues that graph search is one of the areas where looping offers a very large theoretical advantage over standard RLVR.
Interestingly, Mythos’s benchmark result in this area (Graphwalks BFS) is 80%, far ahead of Claude Opus (38%) and GPT-5.4 (21.4%). This also seems to be the first time many people in ML have even heard of Graphwalks BFS.
Main points:
Ouro is a Looped Language Model (LoopLM), a new architecture for LLMs
Instead of stacking many different layers, Ouro reuses the same group of layers multiple times in a loop
It has an exit gate to decide when to stop (adaptive computation)
It is trained with an entropy-regularized objective
With only 1.4B and 2.6B parameters, it matches the performance of 4B–12B models
The reason is not that it memorizes more, but that it manipulates knowledge more effectively