Article - https://www.aiblade.net/p/is-github-copilot-poisoned-part-2
In my previous post, I looked at how code generation models could potentially be poisoned. The impacts could be devastating, and I created a small script to find evidence of this at play. However, my code was too slow, and I didn’t find any meaningful results.
In this post, I seek to improve upon my last experiment. I’ll investigate massive datasets of coding-related prompts, collect thousands of lines of AI-generated code, and analyse this code for evidence of malicious activity.
Try Notion For Free! https://affiliate.notion.so/pqesm7yjddbc
This is a public episode. If you would like to discuss this with other subscribers or get access to bonus episodes, visit www.aiblade.net