Anthropic has updated its AI safety policy to prevent misuse, reports VentureBeat author Michael Nuñez. The new “Capability Thresholds” define benchmarks for risky capabilities of AI models, such as in the area of bioweapons or autonomous AI research. If a model reaches such a threshold, additional safeguards are triggered. The revised policy also sets out more detailed responsibilities for a “Responsible Scaling Officer” who oversees compliance with safety standards. Anthropic hopes the policy will serve as a blueprint for the entire AI industry and lead to a race for the best safety standards.