Hackphyr: A Local Fine-Tuned LLM Agent for Network Security Environments
Abstract
Large Language Models (LLMs) have shown remarkable potential across various domains, including cybersecurity. Using commercial cloud-based LLMs may be undesirable due to privacy concerns, costs, and network connectivity constraints. In this paper, we present Hackphyr, a locally fine-tuned LLM to be used as a red-team agent within network security environments. Our fine-tuned 7 billion parameter model can run on a single GPU card and achieves performance comparable with much larger and more powerful commercial models such as GPT-4. Hackphyr clearly outperforms other models, including GPT-3.5-turbo, and baselines, such as Q-learning agents in complex, previously unseen scenarios. To achieve this performance, we generated a new task-specific cybersecurity dataset to enhance the base model's capabilities. Finally, we conducted a comprehensive analysis of the agents' behaviors that provides insights into the planning abilities and potential shortcomings of such agents, contributing to the broader understanding of LLM-based agents in cybersecurity contexts
Community
Hackphyr is a fine-tuned model that works as a red team agent in the NetSecGame environment. The paper describes the fine-tuning process, the experiments in different network setups, and the comparisons with LLM agents based on GPT-4 and other models. It also contains a behavioral analysis of the agents' actions that shows that the best models behave in a very reasonable manner that resembles human practices.
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- LLM Honeypot: Leveraging Large Language Models as Advanced Interactive Honeypot Systems (2024)
- CIPHER: Cybersecurity Intelligent Penetration-testing Helper for Ethical Researcher (2024)
- Pathway to Secure and Trustworthy 6G for LLMs: Attacks, Defense, and Opportunities (2024)
- Detection Made Easy: Potentials of Large Language Models for Solidity Vulnerabilities (2024)
- CyberPal.AI: Empowering LLMs with Expert-Driven Cybersecurity Instructions (2024)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 1
Spaces citing this paper 0
No Space linking this paper