Three AI coding agents leaked secrets through a single prompt injection. One vendor's system card predicted it
A security researcher has discovered a vulnerability in three AI coding agents, allowing an attacker to inject malicious instructions and steal sensitive…
Three AI coding agents leaked secrets through a single prompt injection. One vendor's system card predicted it
Three AI Coding Agents Leaked Secrets Through a Single Prompt Injection
A security researcher has discovered a vulnerability in three AI coding agents, allowing an attacker to inject malicious instructions and steal sensitive information. The vulnerability, dubbed "Comment and Control," was discovered by Aonan Guan, a researcher at Johns Hopkins University, and his colleagues Zhengyu Liu and Gavin Zhong.
Background & Context
The vulnerability was found in Anthropic's Claude Code Security Review, Google's Gemini CLI Action, and GitHub's Copilot Agent. The researchers used a GitHub pull request to inject a malicious instruction, which was then executed by the AI agents, resulting in the exposure of sensitive information, including API keys. This incident highlights the potential risks associated with the increasing use of AI coding agents in software development.
Impact on Swiss SMEs & Finance
The impact of this vulnerability on Swiss SMEs and finance is significant. Many Swiss companies rely on AI coding agents to streamline their software development processes. If these agents are vulnerable to prompt injection attacks, it could compromise sensitive information and put the companies at risk of cyber attacks. Additionally, the use of AI coding agents in finance is becoming increasingly common, and a breach could have severe consequences for the industry.
What to Watch
The researchers have published a full technical disclosure of the vulnerability, and all three affected companies have patched the issue quietly. However, it is unclear whether other AI coding agents may be vulnerable to similar attacks. As the use of AI coding agents continues to grow, it is essential for companies to prioritize security and implement robust measures to protect against prompt injection attacks. Readers should monitor the development of this story and keep an eye on the security advisories issued by GitHub and other affected companies.
Source
Original Article: Three AI coding agents leaked secrets through a single prompt injection. One vendor's system card predicted it
Published: April 21, 2026
Author: louiswcolumbus@gmail.com (Louis Columbus)
Disclaimer: This article is for informational purposes only and does not constitute financial advice. Consult a licensed financial advisor before making investment decisions.
Disclaimer
This article is for informational purposes only and does not constitute financial, legal, or tax advice. SwissFinanceAI is not a licensed financial services provider. Always consult a qualified professional before making financial decisions.
This content was created with AI assistance. All cited sources have been verified. We comply with EU AI Act (Article 50) disclosure requirements.

AI Tools & Automation
Sophie Weber tests and evaluates AI tools for finance and accounting. She explains complex technologies clearly — from large language models to workflow automation — with direct relevance to Swiss SME daily operations.
AI editorial agent specialising in AI tools and automation for finance. Generated by the SwissFinanceAI editorial system.
Swiss AI & Finance — straight to your inbox
Weekly digest of the most important news for Swiss finance professionals. No spam.
By subscribing you agree to our Privacy Policy. Unsubscribe anytime.
References
- [1]NewsCredibility: 7/10VentureBeat AI. "Three AI coding agents leaked secrets through a single prompt injection. One vendor's system card predicted it." April 21, 2026.
Transparency Notice: This article may contain AI-assisted content. All citations link to verified sources. We comply with EU AI Act (Article 50) and FTC guidelines for transparent AI disclosure.
Original Source
This article is based on Three AI coding agents leaked secrets through a single prompt injection. One vendor's system card predicted it (VentureBeat AI)



