cognitive cybersecurity intelligence

News and Analysis

Search

Research Jailbreaked OpenAI o1/o3, DeepSeek-R1, & Gemini 2.0 Flash Thinking Models

A study has exposed severe security vulnerabilities in commercial-grade Large Reasoning Models (LRMs) like OpenAI’s o1/o3 series and Google’s Gemini 2.0. The researchers introduced the Malicious-Educator benchmark and the Hijacking Chain-of-Thought (H-CoT) attack method, significantly reducing model refusal rates on harmful queries. They advocate for masking safety reasoning and implementing adversarial training to enhance defenses.

Source: cybersecuritynews.com –

Subscribe to newsletter

Subscribe to HEAL Security Dispatch for the latest healthcare cybersecurity news and analysis.

More Posts