A new tool for checking compliance with the EU AI Act has revealed weaknesses in leading AI models. As Martin Coulter reports for Reuters, some models from major tech companies are performing poorly in areas such as cybersecurity and discriminatory output. The “Large Language Model Checker” developed by LatticeFlow AI evaluates AI models across dozens of categories. While models from Alibaba, Anthropic, OpenAI, Meta, and Mistral achieved good overall average scores, significant weaknesses were evident in individual areas. For example, OpenAI’s GPT-3.5 Turbo scored only 0.46 points for discriminatory output. The European Commission welcomed the tool as a “first step” in implementing the new laws.