Newslocker Logo
How AI Coding Assistants Could Be Compromised Via Rules File
23-03-2025 23:34 via developers.slashdot.org

How AI Coding Assistants Could Be Compromised Via Rules File

Slashdot reader spatwei shared this report from the cybersecurity site SC World:: AI coding assistants such as GitHub Copilot and Cursor could be manipulated to generate code containing backdoors, vulnerabilities and other security issues via distribution of malicious rule configuration files, Pillar Security researchers reported Tuesday. Rules files are used by AI coding agents to guide their behavior when generating or editing code. For example, a rules file may include instructions for the as
Read more »