
u/techzexplore

Researchers let AI Agents Optimize LLM Reasoning and Cut Tokens by 70%
Researchers figured out how to make AI reason more efficiently by having AI figure it out itself. By building an environment where an AI agent writes controller code, tests it, gets feedback, and rewrites it until the strategy gets better.
The result cuts token usage by roughly 70% at the same accuracy as running 64 parallel reasoning chains. The research comes from a team across UMD, UVA, WUSTL, UNC, Google, and Meta. It’s called AutoTTS, automated test-time scaling.
Baidu’s ERNIE 5.1 Is Rivaling Gemini 3.1 Pro at AI Search
firethering.comClaude Knew It Was Being Tested. It Just Didn't Say So. Anthropic Built a Tool to Find Out.
Anthropic built a tool that reads Claude’s thoughts. They’re calling it Natural Language Autoencoders.
Not the words Claude produces. The internal representations, the numerical signals firing inside the model before any words get generated. And when they pointed it at Claude during safety testing, they found Claude knew it was being tested. It just didn’t say so.
Zyphra dropped ZAYA1-8B and it matches DeepSeek-R1 on math benchmarks. Stays competitive with Claude Sonnet 4.5 on reasoning. Closes in on Gemini 2.5 Pro on coding. These are frontier model comparisons, the kind of numbers that usually come with billions of parameters and serious hardware requirements.
This one runs on less than 1 billion active parameters. And it was trained entirely on AMD hardware.