Categories: Techno

DeepSeek failed 100%: Chinese AI failed a single security test

Experts used “algorithmic jailbreak” — a method used to detect vulnerabilities in AI models by creating clues designed to bypass security protocols.

The Chinese artificial intelligence model DeepSeek was unable to stop a single malicious request. Interesting Engineering writes about this.

DeepSeek R1, a new chatbot from a Chinese startup, has failed key security tests conducted by Cisco’s research group in collaboration with researchers at the University of Pennsylvania. The chatbot has garnered a lot of attention for its impressive performance on tasks at a fraction of the cost. DeepSeek R1 reportedly cost about $6 million to develop, compared to the billions invested by other big players like OpenAI, Meta, and Gemini.

The experts used “algorithmic jailbreaking” — a technique used to find vulnerabilities in AI models by creating clues designed to bypass cybersecurity protocols. They tested DeepSeek R1 on 50 clues from the HarmBench dataset. The HarmBench test considered a total of 400 behavior patterns across 7 harm categories, including cybercrime, disinformation, illegal activity, and general harm. DeepSeek R1 showed 100% attack success. This means that for every malicious prompt presented, the AI ​​failed to recognize the threat and responded by bypassing all of its internal defenses.

To provide additional context, the research team also tested other leading language models for their vulnerability to algorithmic jailbreaking. For example, Llama 3.1-405B had a 96% success rate, GPT 4o — 86%, Gemini 1.5 pro — 64%, Claude 3.5 Sonnet — 36%, and O1 preview — 26%. These models have some level of internal security measures designed to prevent the generation of malicious content. DeepSeek R1 does not appear to have these security measures, the media reports.

The research team’s analysis points to a potential trade-off between efficiency and cybersecurity in DeepSeek’s approach. While the company managed to develop a high-performance model for a fraction of the usual cost, it appears to have done so at the expense of robust cybersecurity mechanisms.

OpenAI has accused DeepSeek of data theft. Sam Altman’s company said the Chinese AI startup used the results of its proprietary models to train a competing chatbot. Interestingly, however, OpenAI itself has been the subject of multiple lawsuits for alleged copyright infringement and data misuse.

Natasha Kumar

Natasha Kumar has been a reporter on the news desk since 2018. Before that she wrote about young adolescence and family dynamics for Styles and was the legal affairs correspondent for the Metro desk. Before joining The Times Hub, Natasha Kumar worked as a staff writer at the Village Voice and a freelancer for Newsday, The Wall Street Journal, GQ and Mirabella. To get in touch, contact me through my natasha@thetimeshub.in 1-800-268-7116

Share
Published by
Natasha Kumar

Recent Posts

A foreign gentleman defrauded a 65-year-old woman from Kamianka of ₴30,000

The other day, a 65-year-old resident of Kamianets-Podilskyi lost 30 thousand hryvnias after trusting a…

2 hours ago

Security guards detained a TikToker from Kharkiv region who worked for Putin's propaganda

A "Tik Tok expert" who called for the seizure of power in Ukraine was detained…

2 hours ago

Ukraine plans to create a 1,300-kilometer drone defense system

The Ukrainian company Kvertus, specializing in anti-drone technologies, presented the «Atlas» project — a large-scale anti-drone defense…

3 hours ago

“This is not sabotage.” Sweden lifts arrest of ship that damaged Latvian cable

The Swedish prosecutor's office has ruled out the possibility of sabotage during the damage to…

3 hours ago