Google launches new benchmark to test AI models’ factual accuracy
Google has introduced FACTS Grounding, a new benchmark system to evaluate how accurately large language models (LLMs) use source material in their responses. The benchmark comprises 1,719 examples across various domains including finance, technology, and medicine. The FACTS team at Google DeepMind and Google Research developed the system, which uses three frontier LLM judges – …