Researchers cause GitLab AI developer assistant to turn safe code malicious

Date:

Share:



Marketers promote AI-assisted developer tools as workhorses that are essential for today’s software engineer. Developer platform GitLab, for instance, claims its Duo chatbot can “instantly generate a to-do list” that eliminates the burden of “wading through weeks of commits.” What these companies don’t say is that these tools are, by temperament if not default, easily tricked by malicious actors into performing hostile actions against their users.

Researchers from security firm Legit on Thursday demonstrated an attack that induced Duo into inserting malicious code into a script it had been instructed to write. The attack could also leak private code and confidential issue data, such as zero-day vulnerability details. All that’s required is for the user to instruct the chatbot to interact with a merge request or similar content from an outside source.

AI assistants’ double-edged blade

The mechanism for triggering the attacks is, of course, prompt injections. Among the most common forms of chatbot exploits, prompt injections are embedded into content a chatbot is asked to work with, such as an email to be answered, a calendar to consult, or a webpage to summarize. Large language model-based assistants are so eager to follow instructions that they’ll take orders from just about anywhere, including sources that can be controlled by malicious actors.

The attacks targeting Duo came from various resources that are commonly used by developers. Examples include merge requests, commits, bug descriptions and comments, and source code. The researchers demonstrated how instructions embedded inside these sources can lead Duo astray.

“This vulnerability highlights the double-edged nature of AI assistants like GitLab Duo: when deeply integrated into development workflows, they inherit not just context—but risk,” Legit researcher Omer Mayraz wrote. “By embedding hidden instructions in seemingly harmless project content, we were able to manipulate Duo’s behavior, exfiltrate private source code, and demonstrate how AI responses can be leveraged for unintended and harmful outcomes.”



Source link

━ more like this

The White House app is just as weird and unnecessary as you’d expect

President Donald Trump may have a tendency to put his name on everything, but his administration decided to go with the more authoritative...

M5 MacBook Pro tests show Apple is pretty close to fixing its worst weakness

For years, Macs have had one glaring weakness: gaming. But with the new M5 MacBook Pro, Apple might finally be getting close to...

Sony is halting sales of memory cards and you have AI to blame for it

Sony has hit pause on a major part of its storage business, and not-so-surprisingly, AI is one of the reasons behind it. The...

Meta’s next AI glasses are reportedly designed with prescription lenses in mind

Two new models of Meta Ray-Ban AI glasses are on the way, and they're going to be catered towards those who use prescription...

Trump warns the US might not be there for NATO when needed – London Business News | Londonlovesbusiness.com

US President Donald Trump has issued a stark warning to allied nations regarding the potential consequences of their inaction amid the escalating crisis...
spot_img