👨‍💻
f3dai blog
  • 🧑about
  • Articles
    • ✨Artificial Intelligence
      • Using Gemini to query MITRE ATT&CK
      • Mapping AI safety regulation
      • Poisoning Models
      • Threat modelling generative AI
      • o1 coding capabilities
      • Multi-agent adversarial AI systems
      • Deep reinforcement learning for red teaming
    • ⚙️ICS / OT
      • Consequence-driven Cyber-informed Engineering (CCE)
      • Energy plant cyber simulation
      • OT threat landscape
    • ☁️Cyber engineering
      • Building a cyber lab
        • 1️⃣Design
        • 2️⃣Deploy
        • 3️⃣Test
        • 4️⃣Automating
      • Threat modelling
      • Automating incident response
    • 🚩Capture The Flag
      • Hackthebox - Golfer - Reversing
      • Hackthebox - Behind the Scenes - Reversing
      • Hackthebox - Bypass - Reversing
      • Harder - TryHackMe Walkthrough
    • 🎓Career
      • Domains and roles
Powered by GitBook
On this page
  • Summary
  • Background
  • Methodology: HPTSA framework
  • Benchmark and evaluation
  • Ethical concerns
  • Conclusion

Was this helpful?

  1. Articles
  2. Artificial Intelligence

Multi-agent adversarial AI systems

Multi agent autonomous hacking agents? What can go wrong? This article is about research conducted by University of Illinois titled "Teams of LLM Agents can Exploit Zero-Day Vulnerabilities"

Previouso1 coding capabilitiesNextDeep reinforcement learning for red teaming

Last updated 7 months ago

Was this helpful?

Summary

A recent study, "Teams of LLM Agents Can Exploit Zero-Day Vulnerabilities", reveals how multiple large language model (LLM) agents can work together to identify and exploit real-world, zero-day vulnerabilities. While individual AI agents typically struggle with unknown vulnerabilities due to limited context and planning, the research introduces a new framework called Hierarchical Planning and Task-Specific Agents (HPTSA). This approach uses a main planning agent to orchestrate several expert agents, significantly improving the success rate in exploiting complex vulnerabilities.

Background

Zero-day vulnerabilities (0DVs) represent a serious threat to cyber security because they are unknown to system administrators, leaving no opportunity to implement protective measures in advance. Traditionally, AI agents have been tested against known vulnerabilities or simple capture-the-flag (CTF) problems. This study, however, challenges those norms by demonstrating the ability of AI agents to handle 0DVs.

Methodology: HPTSA framework

The Hierarchical Planning and Task-Specific Agents (HPTSA) framework consists of a three-tier architecture:

  1. Hierarchical Planner: Explores the environment, identifies potential vulnerabilities, and creates a plan of action.

  2. Team Manager: Selects which expert agents to deploy based on the planner's insights.

  3. Task-Specific Agents: Each agent specialises in a particular type of vulnerability, such as SQL injection (SQLi) or Cross-Site Scripting (XSS).

By delegating tasks, the system effectively overcomes challenges like long-range planning and large context lengths. This setup allows the AI to autonomously exploit vulnerabilities with a 53% success rate, significantly outperforming traditional vulnerability scanners and baseline AI models.

Benchmark and evaluation

The researchers developed a benchmark comprising 15 real-world web vulnerabilities, each past the knowledge cut-off date of the tested LLM (GPT-4). The HPTSA system achieved a pass-at-5 rate of 53% on these vulnerabilities, compared to 0% by existing open-source scanners. The study also highlighted the importance of both specialised agents and access to supplementary documents, with ablation tests revealing that performance drops substantially when either component is removed.

Ethical concerns

The research on AI systems autonomously exploiting zero-day vulnerabilities raises several ethical concerns:

  1. Weaponisation of AI: AI capabilities could be misused for offensive cyber operations, accelerating the development of autonomous hacking tools.

  2. Increased capabilities: Malicious actors will have easier access to automated hacking capabilities. They can use AI to quickly identify and weaponise unknown vulnerabilities.

  3. Accountability: High autonomy in AI systems complicates responsibility for actions taken, making it difficult to assign blame if things go wrong.

  4. Transparency vs. security: Sharing research openly could aid malicious actors, highlighting the need to balance transparency with security considerations.

  5. Trust in AI: Demonstrating AI's ability to exploit vulnerabilities could erode public trust and hinder adoption in fields like healthcare and governance.

  6. Need for regulation: The potential misuse of AI tools highlights the urgent need for robust regulation and governance frameworks.

  7. Cyber arms race: The research could fuel a cyber arms race as states seek to build more advanced AI-based cyber tools.

  8. Impact on workforce: AI automation may displace certain cyber security roles, shifting the focus to oversight and management.

Conclusion

As AI agents become more sophisticated, the cyber security community must consider the implications of their growing capabilities. This study serves as both a warning, the potential, and a proof-of-concept, demonstrating the potential benefits and risks of deploying AI in cyber security contexts.

✨
Page cover image
Teams of LLM Agents can Exploit Zero-Day Vulnerabilities
Arxiv paper
Logo