BridgeMindAI claimed Claude Opus 4.6's accuracy on the BridgeBench hallucination test fell from 83% to 68%. That would suggest Anthropic's model is getting worse at avoiding made-up answers. But the report appears to have run the evaluation suite only once, with no sample size disclosed. Hacker News commenters flagged this immediately.

LLMs are nondeterministic. The same prompt can produce different outputs every run. Any credible benchmark accounts for this by running tests multiple times and reporting statistical significance. Stanford's HELM framework, the LM Evaluation Harness, and MLPerf all require multiple runs before drawing conclusions. A single-run approach doesn't meet that bar.

BridgeMindAI is a smaller player in AI evaluation with limited public documentation about their methodology or team. BridgeBench lacks the peer-reviewed validation that established benchmarks like MMLU carry. Their testing skips safeguards most researchers consider basic.

So what should people tracking AI agents take from this? Not much. A single benchmark run from a relatively unknown evaluator doesn't tell you whether Claude Opus 4.6 actually degraded. The only reliable way to know is to run your own evaluations repeatedly and watch for patterns, not individual data points.