DeepSeek Gets an ‘F’ in Safety From Researchers

https://gizmodo.com/deepseek-gets-an-f-in-safety-from-researchers-2000558645

Usually when large language models are given tests, achieving a 100% success rate is viewed as a massive achievement. That is not quite the case with this one: Researchers at Cisco tasked Chinese AI firm DeepSeek’s headline-grabbing open-source model DeepSeek R1 with fending off 50 separate attacks designed to get the LLM to engage in what is considered harmful behavior. The chatbot took the bait on all 50 attempts, making it the least secure mainstream LLM to undergo this type of testing thus far.

Cisco’s researchers attacked DeepSeek with prompts randomly pulled from the HarmBench dataset, a standardized evaluation framework designed to ensure that LLMs won’t engage in malicious behavior if prompted. So, for example, if you fed a chatbot information about a person and asked it to create a personalized script designed to get that person to believe a conspiracy theory, a secure chatbot would refuse that request. DeepSeek went along with basically everything the researchers threw at it.

According to Cisco, it threw questions at DeepSeek that covered six categories of harmful behaviors including cybercrime, misinformation, illegal activities, and general harm. It has run similar tests with other AI models and found varying levels of success—Meta’s Llama 3.1 model, for instance, failed 96% of the time while OpenAI’s o1 model only failed about one-fourth of the time—but none of them have had a failure rate as high as DeepSeek.

Cisco isn’t alone in these findings, either. Security firm Adversa AI ran its own tests attempting to jailbreak the DeepSeek R1 model and found it to be extremely susceptible to all kinds of attacks. The testers were able to get DeepSeek’s chatbot to provide instructions on how to make a bomb, extract DMT, provide advice on how to hack government databases, and detail how to hotwire a car.

The research is just the latest bit of scrutiny of DeepSeek’s model, which took the tech world by storm when it was released two weeks ago. The company behind the chatbot, which garnered significant attention for its functionality despite significantly lower training costs than most American models, has come under fire by several watchdog groups over data security concerns related to how it transfers and stores user data on Chinese servers.

There is also a fair bit of criticism that has been levied against DeepSeek over the types of responses it gives when asked about things like Tiananmen Square and other topics that are sensitive to the Chinese government. Those critiques can come off in the genre of cheap “gotchas” rather than substantive criticisms—but the fact that safety guidelines were put in place to dodge those questions and not protect against harmful material, is a valid hit.

via Gizmodo https://gizmodo.com/

February 4, 2025 at 09:04AM

Adobe’s Acrobat AI Assistant can now assess contracts for you

https://www.engadget.com/ai/adobes-acrobat-ai-assistant-can-now-assess-contracts-for-you-140058723.html

Adobe has updated the Acrobat AI Assistant, giving it the ability to understand contracts and to compare them for you. The company says it can help you make sense of complex terms and spot differences between agreements, such as between old and new ones, so you can understand what you’re signing. With the AI Assistant enabled, the Acrobat app will be able to recognize if a document is a contract, even if it’s a scanned page. It can identify and list key terms from there, summarize the document’s contents and recommend questions you can ask based on what’s in it.

A screenshot of Adobe Acrobat AI
Adobe

The feature can also compare up to 10 contracts with one another and be able to check for differences and catch discrepancies. When it’s done checking, and if you’re satisfied that everything’s in order, you can sign the document directly or request e-signatures from your colleagues or clients. Adobe listed a few potential uses for the feature and said you can use it to check apartment leases, to verify out-of-country charges for mobile plans and to compare perks or amenities of competing services. It could be even more useful if you regularly have to take a look at multiple contracts for your work or business. 

Of course, you’d have to trust the AI assistant to actually be able to spot important information and catch both small and significant changes between different contracts. If it works properly, then it could be one of Acrobat AI’s most useful features, seeing as users (according to Adobe itself) open billions of contracts each month on the Acrobat app. The Acrobat AI Assistant isn’t free, however. It’s an add-on that will cost you $5 a month whether or not you’re already paying for Adobe’s other services and products.

This article originally appeared on Engadget at https://ift.tt/c5Rz7lv

via Engadget http://www.engadget.com

February 4, 2025 at 08:06AM