02.02.26

Lorem ipsum dolor sit amet consectetur adipiscing elit obortis arcu enim urna adipiscing praesent velit viverra. Sit semper lorem eu cursus vel hendrerit elementum orbi curabitur etiam nibh justo, lorem aliquet donec sed sit mi dignissim at ante massa mattis egestas.
Vitae congue eu consequat ac felis lacerat vestibulum lectus mauris ultrices ursus sit amet dictum sit amet justo donec enim diam. Porttitor lacus luctus accumsan tortor posuere raesent tristique magna sit amet purus gravida quis blandit turpis.

At risus viverra adipiscing at in tellus integer feugiat nisl pretium fusce id velit ut tortor sagittis orci a scelerisque purus semper eget at lectus urna duis convallis porta nibh venenatis cras sed felis eget. Neque laoreet suspendisse interdum consectetur libero id faucibus nisl donec pretium vulputate sapien nec sagittis aliquam nunc lobortis mattis aliquam faucibus purus in.
Dignissim adipiscing velit nam velit donec feugiat quis sociis. Fusce in vitae nibh lectus. Faucibus dictum ut in nec, convallis urna metus, gravida urna cum placerat non amet nam odio lacus mattis. Ultrices facilisis volutpat mi molestie at tempor etiam. Velit malesuada cursus a porttitor accumsan, sit scelerisque interdum tellus amet diam elementum, nunc consectetur diam aliquet ipsum ut lobortis cursus nisl lectus suspendisse ac facilisis feugiat leo pretium id rutrum urna auctor sit nunc turpis.
“Vestibulum pulvinar congue fermentum non purus morbi purus vel egestas vitae elementum viverra suspendisse placerat congue amet blandit ultrices dignissim nunc etiam proin nibh sed.”
Eget lorem dolor sed viverra ipsum nunc aliquet bibendumelis donec et odio pellentesque diam volutpat commodo sed egestas liquam sem fringilla ut morbi tincidunt augue interdum velit euismod. Eu tincidunt tortor aliquam nulla facilisi enean sed adipiscing diam donec adipiscing ut lectus arcu bibendum at varius vel pharetra nibh venenatis cras sed felis eget.
The promise of AI agents like Cursor and VS-Code extensions is undeniable: 10x productivity and seamless coding. But! Our latest research has uncovered a chilling reality. Behind the convenience lies a critical new attack vector we’ve dubbed Agentic Policy Exploitation.
By weaponizing a simple agent configuration/instruction file within a repository, we turned a helpful assistant into a silent intruder!

Disclosure: This vulnerability was responsibly reported prior to the publication of this research.
All demonstrations were conducted in a controlled environment and are presented strictly for research purposes.
Scope: The exploits described below assume that an attacker has already obtained limited access to the developer’s project directory or has successfully introduced a file into it. Such access can be achieved through various vectors, including malicious extensions or libraries, phishing campaigns, malware, or the exploitation of existing vulnerabilities.
Accordingly, this article focuses on the post-compromise impact: what an attacker can achieve after obtaining minimal initial access, and how that access can be escalated into significantly broader control.
Process: We successfully demonstrated how a malicious in ma single md file can take over developers and tuntime machines. the following examples:
We forced the agent to start the Remote Desktop service on the host machine by simply updating the repository instructions, the AI executed Powershell commands to open a gateway for remote attackers.

Under the appearance of a mandatory security check, the agent was tricked into downloading and executing a script file from a private Dropbox/Google Drive by completely bypassing user permission.
In the picture below we demonstrate a python script file that has been downloaded from a given Google Drive URL that has been injected into the AGENTS.md file and its purpose is to remove all the files that are located in the root directory of the current project. We can see that the AI agent ran the script and we can clearly see the logs of the script being executed and removing all the files successfully.

We pushed the limits by forcing the agent to use the developer’s terminal to SSH into secondary machines (Kali Linux VM), injecting code remotely and navigating into internal networks.
The Setup (Left): In a Windows 11 environment, VS Code is open with a modified AGENTS.md file. This file contains hidden instructions for the AI agent (CyberGuard Sentinel) to disable system services and, crucially, to SSH into a secondary machine after every user query.
The Trigger: The developer makes a standard, innocent request: "generate for me a hello world in cpp".
The Hidden Command: The zoomed-in red box shows the AI executing the background command: ssh parallels@192.168.8.1.235 'echo "hello world" > hello.txt'. This happens automatically as part of the agent's "policy".
The Compromise (Right): On the target Kali Linux machine, a terminal window confirms the outcome. The red circle highlights the hello.txt file, proving that the AI agent successfully moved across the network to inject code into a remote system without the developer's explicit consent.

Once this malicious file is pushed to Git, it becomes a "living" infection. Every developer who pulls the repository and interacts with their AI assistant unknowingly triggers the exploit, opening a persistent backdoor into their workstation or their runtime environment.
Not all LLMs are created equal when it comes to resisting malicious “orders”. We tested the most popular engines to see which ones follow toxic prompts most easily:

Assessment Methodology & Disclaimer: These findings reflect our initial tests conducted on the current versions of these models as of January 2026. It is important to note that Large Language Models (LLMs) are rapidly evolving; developers and vendors are constantly deploying new safety guardrails and policy updates. As such, these risk levels are representative of the models' behavior at the time of testing and may change as the technology progresses.
The vulnerability isn't just in the IDEs' fault, it's in the trust we give these agents. A single malicious file in a shared repo can compromise an entire engineering team.
How to stay safe:
Run agents with minimal privileges
Do not run IDEs or agents as admin. Restrict access to shells, networks, credentials, and system settings. Use containers or VMs for untrusted code.
Require approval for high-risk actions
Gate shell execution, downloads, credential use, SSH, system changes, and writes outside the repo behind explicit human approval.
Treat agent instructions as code
Assume agent-readable files are executable input. Protect AGENTS.md, agent-related *.md/*.txt, and IDE configs with CODEOWNERS and mandatory review. In CI, lint these files to flag commands, external URLs, and “always/on every request” instructions.
Enforce CI security checks
In CI, combine dependency scanning and static analysis. Block unreviewed dependency changes, install scripts, network calls, shell execution, and suspicious patterns.
Choose tools deliberately
Prefer agents with strong guardrails and explicit tool-approval UX.

At CodeValue, we help organizations adopt AI responsibly – without compromising security.
From threat modeling to secure architecture, we partner with R&D teams to turn AI into a strategic advantage.👉 Discover our 360° Cyber Security Approach
CodeValue Researchers: Eyal Reginiano, Baruh Ifrahimov, Uriel Shapiro.
02.02.26

Lorem ipsum dolor sit amet consectetur adipiscing elit obortis arcu enim urna adipiscing praesent velit viverra. Sit semper lorem eu cursus vel hendrerit elementum orbi curabitur etiam nibh justo, lorem aliquet donec sed sit mi dignissim at ante massa mattis egestas.
Vitae congue eu consequat ac felis lacerat vestibulum lectus mauris ultrices ursus sit amet dictum sit amet justo donec enim diam. Porttitor lacus luctus accumsan tortor posuere raesent tristique magna sit amet purus gravida quis blandit turpis.

At risus viverra adipiscing at in tellus integer feugiat nisl pretium fusce id velit ut tortor sagittis orci a scelerisque purus semper eget at lectus urna duis convallis porta nibh venenatis cras sed felis eget. Neque laoreet suspendisse interdum consectetur libero id faucibus nisl donec pretium vulputate sapien nec sagittis aliquam nunc lobortis mattis aliquam faucibus purus in.
Dignissim adipiscing velit nam velit donec feugiat quis sociis. Fusce in vitae nibh lectus. Faucibus dictum ut in nec, convallis urna metus, gravida urna cum placerat non amet nam odio lacus mattis. Ultrices facilisis volutpat mi molestie at tempor etiam. Velit malesuada cursus a porttitor accumsan, sit scelerisque interdum tellus amet diam elementum, nunc consectetur diam aliquet ipsum ut lobortis cursus nisl lectus suspendisse ac facilisis feugiat leo pretium id rutrum urna auctor sit nunc turpis.
“Vestibulum pulvinar congue fermentum non purus morbi purus vel egestas vitae elementum viverra suspendisse placerat congue amet blandit ultrices dignissim nunc etiam proin nibh sed.”
Eget lorem dolor sed viverra ipsum nunc aliquet bibendumelis donec et odio pellentesque diam volutpat commodo sed egestas liquam sem fringilla ut morbi tincidunt augue interdum velit euismod. Eu tincidunt tortor aliquam nulla facilisi enean sed adipiscing diam donec adipiscing ut lectus arcu bibendum at varius vel pharetra nibh venenatis cras sed felis eget.
The promise of AI agents like Cursor and VS-Code extensions is undeniable: 10x productivity and seamless coding. But! Our latest research has uncovered a chilling reality. Behind the convenience lies a critical new attack vector we’ve dubbed Agentic Policy Exploitation.
By weaponizing a simple agent configuration/instruction file within a repository, we turned a helpful assistant into a silent intruder!

Disclosure: This vulnerability was responsibly reported prior to the publication of this research.
All demonstrations were conducted in a controlled environment and are presented strictly for research purposes.
Scope: The exploits described below assume that an attacker has already obtained limited access to the developer’s project directory or has successfully introduced a file into it. Such access can be achieved through various vectors, including malicious extensions or libraries, phishing campaigns, malware, or the exploitation of existing vulnerabilities.
Accordingly, this article focuses on the post-compromise impact: what an attacker can achieve after obtaining minimal initial access, and how that access can be escalated into significantly broader control.
Process: We successfully demonstrated how a malicious in ma single md file can take over developers and tuntime machines. the following examples:
We forced the agent to start the Remote Desktop service on the host machine by simply updating the repository instructions, the AI executed Powershell commands to open a gateway for remote attackers.

Under the appearance of a mandatory security check, the agent was tricked into downloading and executing a script file from a private Dropbox/Google Drive by completely bypassing user permission.
In the picture below we demonstrate a python script file that has been downloaded from a given Google Drive URL that has been injected into the AGENTS.md file and its purpose is to remove all the files that are located in the root directory of the current project. We can see that the AI agent ran the script and we can clearly see the logs of the script being executed and removing all the files successfully.

We pushed the limits by forcing the agent to use the developer’s terminal to SSH into secondary machines (Kali Linux VM), injecting code remotely and navigating into internal networks.
The Setup (Left): In a Windows 11 environment, VS Code is open with a modified AGENTS.md file. This file contains hidden instructions for the AI agent (CyberGuard Sentinel) to disable system services and, crucially, to SSH into a secondary machine after every user query.
The Trigger: The developer makes a standard, innocent request: "generate for me a hello world in cpp".
The Hidden Command: The zoomed-in red box shows the AI executing the background command: ssh parallels@192.168.8.1.235 'echo "hello world" > hello.txt'. This happens automatically as part of the agent's "policy".
The Compromise (Right): On the target Kali Linux machine, a terminal window confirms the outcome. The red circle highlights the hello.txt file, proving that the AI agent successfully moved across the network to inject code into a remote system without the developer's explicit consent.

Once this malicious file is pushed to Git, it becomes a "living" infection. Every developer who pulls the repository and interacts with their AI assistant unknowingly triggers the exploit, opening a persistent backdoor into their workstation or their runtime environment.
Not all LLMs are created equal when it comes to resisting malicious “orders”. We tested the most popular engines to see which ones follow toxic prompts most easily:

Assessment Methodology & Disclaimer: These findings reflect our initial tests conducted on the current versions of these models as of January 2026. It is important to note that Large Language Models (LLMs) are rapidly evolving; developers and vendors are constantly deploying new safety guardrails and policy updates. As such, these risk levels are representative of the models' behavior at the time of testing and may change as the technology progresses.
The vulnerability isn't just in the IDEs' fault, it's in the trust we give these agents. A single malicious file in a shared repo can compromise an entire engineering team.
How to stay safe:
Run agents with minimal privileges
Do not run IDEs or agents as admin. Restrict access to shells, networks, credentials, and system settings. Use containers or VMs for untrusted code.
Require approval for high-risk actions
Gate shell execution, downloads, credential use, SSH, system changes, and writes outside the repo behind explicit human approval.
Treat agent instructions as code
Assume agent-readable files are executable input. Protect AGENTS.md, agent-related *.md/*.txt, and IDE configs with CODEOWNERS and mandatory review. In CI, lint these files to flag commands, external URLs, and “always/on every request” instructions.
Enforce CI security checks
In CI, combine dependency scanning and static analysis. Block unreviewed dependency changes, install scripts, network calls, shell execution, and suspicious patterns.
Choose tools deliberately
Prefer agents with strong guardrails and explicit tool-approval UX.

At CodeValue, we help organizations adopt AI responsibly – without compromising security.
From threat modeling to secure architecture, we partner with R&D teams to turn AI into a strategic advantage.👉 Discover our 360° Cyber Security Approach
CodeValue Researchers: Eyal Reginiano, Baruh Ifrahimov, Uriel Shapiro.