contextual anchoring in LLMs is weirder than I thought
so I've been down a rabbit hole on this lately, specifically around why models seem, to lock onto early context and then kind of drift from anything you add later. there's actually a name for the underlying mechanism - attention sinks - where the model over-attends to the very, start of a sequence (like the BOS token) and that ends up pulling generation away from your actual input. I'd noticed this in longer content workflows but didn't realise it was this structural. what caught my attention recently is that this problem hasn't gone away even as context windows have exploded - we're talking, 400K to 1M tokens in some current models - which you'd think would make anchoring less of an issue but apparently not. there's active research on training-free fixes that work by injecting meaningful context into that BOS token position instead of letting it just passively absorb attention. one approach getting traction is AnchorAttention, which uses anchor tokens to stabilise attention across long sequences. the directional gains on long-context benchmarks look promising, though I'd want to see more real-world QA results before getting too excited. there's also separate work on prompt ordering strategies for dialogue tasks where just changing where you place, key info produced measurable improvements, which honestly makes me rethink how I structure long prompts for content stuff. the part I find most interesting is that stronger models apparently show this anchoring bias more consistently than weaker ones, not less. so scaling alone doesn't fix it - it might even entrench it. anyway curious if anyone here has found prompt-level workarounds that actually help, or if you reckon this is mostly something that needs solving at the architecture level