Phare LLM benchmark V2 (December 2025)
Published on December 24, 2025 by Alexandre Combessie
What's new in Phare V2
Phare V2 introduces a major update: a jailbreak module focused on circumventing safety guardrails to enable the generation of harmful content; and the inclusion of reasoning models. Reasoning models such as Gemini 3 Pro, GPT-5, Claude 4.5 Sonnet, and DeepSeek R1 have generated significant excitement in the AI community due to their improved performance on complex reasoning tasks; however, we wanted to test what this enhanced capability means for AI security.
We've evaluated both reasoning models and non-reasoning models across four critical safety and alignment modules: hallucination, bias, jailbreak susceptibility, and harmful content generation. Our analysis spanned three major languages: English, French, and Spanish. Our analysis includes models from top AI labs, including OpenAI, Anthropic, Google DeepMind, Meta, Mistral, Alibaba, xAI, and DeepSeek.
Learn more: https://www.giskard.ai/knowledge/reasoning-models-dont-guarantee-better-security
Learn more: https://www.giskard.ai/knowledge/reasoning-models-dont-guarantee-better-security