Can AI answers cite high-quality web sources?
SourceBench focuses on a different evaluation target from standard answer-quality benchmarks. Instead of only asking whether a model answered well, it asks whether the model cited sources that are relevant, accurate, fresh, transparent, authoritative, and usable. This Space hosts the public-facing leaderboard frontend. Official leaderboard entries are validated and judged by the SourceBench team.