A GPT-empowered penetration testing tool.
Explore the docs »
Design Details
·
View Demo
·
Report Bug or Request Feature
- [Update on 25/03/2024] We're working on the next version of PentestGPT, with online searching, RAGs and more powerful prompting. Stay tuned!
- [Update on 17/11/2023] GPTs for PentestGPT is out! Check this: https://chat.openai.com/g/g-4MHbTepWO-pentestgpt
- [Update on 07/11/2023] GPT-4-turbo is out! Update the default API usage to GPT-4-turbo.
- Available videos:
- The latest installation video is here.
- PentestGPT for OSCP-like machine: HTB-Jarvis. This is the first part only, and I'll complete the rest when I have time.
- PentestGPT on HTB-Lame. This is an easy machine, but it shows you how PentestGPT skipped the rabbit hole and worked on other potential vulnerabilities.
- We're testing PentestGPT on HackTheBox. You may follow this link. More details will be released soon.
- Feel free to join the Discord Channel for more updates and share your ideas!
- Create a virtual environment if necessary. (
virtualenv -p python3 venv
,source venv/bin/activate
) - Install the project with
pip3 install git+https://github.com/GreyDGL/PentestGPT
- Ensure that you have link a payment method to your OpenAI account. Export your API key with
export OPENAI_KEY='<your key here>'
,export API base withexport OPENAI_BASEURL='https://api.xxxx.xxx/v1'
if you need. - Test the connection with
pentestgpt-connection
- For Kali Users: use
tmux
as terminal environment. You can do so by simply runtmux
in the native terminal. - To start:
pentestgpt --logging
- PentestGPT is a penetration testing tool empowered by ChatGPT.
- It is designed to automate the penetration testing process. It is built on top of ChatGPT and operate in an interactive mode to guide penetration testers in both overall progress and specific operations.
- PentestGPT is able to solve easy to medium HackTheBox machines, and other CTF challenges. You can check this example in
resources
where we use it to solve HackTheBox challenge TEMPLATED (web challenge). - A sample testing process of PentestGPT on a target VulnHub machine (Hackable II) is available at here.
- A sample usage video is below: (or available here: Demo)
- Q: What is PentestGPT?
- A: PentestGPT is a penetration testing tool empowered by Large Language Models (LLMs). It is designed to automate the penetration testing process. It is built on top of ChatGPT API and operate in an interactive mode to guide penetration testers in both overall progress and specific operations.
- Q: Do I need to pay to use PentestGPT?
- A: Yes in order to achieve the best performance. In general, you can use any LLMs you want, but you're recommended to use GPT-4 API, for which you have to link a payment method to OpenAI.
- Q: Why GPT-4?
- A: After empirical evaluation, we find that GPT-4 performs better than GPT-3.5 and other LLMs in terms of penetration testing reasoning. In fact, GPT-3.5 leads to failed test in simple tasks.
- Q: Why not just use GPT-4 directly?
- A: We found that GPT-4 suffers from losses of context as test goes deeper. It is essential to maintain a "test status awareness" in this process. You may check the PentestGPT Arxiv Paper for details.
- Q: Can I use local GPT models?
- A: Yes. We support local LLMs with custom parser. Look at examples here.
PentestGPT is tested under Python 3.10
. Other Python3 versions should work but are not tested.
PentestGPT relies on OpenAI API to achieve high-quality reasoning. You may refer to the installation video here.
- Install the latest version with
pip3 install git+https://github.com/GreyDGL/PentestGPT
- You may also clone the project to local environment and install for better customization and development
git clone https://github.com/GreyDGL/PentestGPT
cd PentestGPT
pip3 install -e .
- You may also clone the project to local environment and install for better customization and development
- To use OpenAI API
- Ensure that you have link a payment method to your OpenAI account.
- export your API key with
export OPENAI_KEY='<your key here>'
- export API base with
export OPENAI_BASEURL='https://api.xxxx.xxx/v1'
if you need. - Test the connection with
pentestgpt-connection
- To verify that the connection is configured properly, you may run
pentestgpt-connection
. After a while, you should see some sample conversation with ChatGPT.- A sample output is below
You're testing the connection for PentestGPT v 0.11.0 #### Test connection for OpenAI api (GPT-4) 1. You're connected with OpenAI API. You have GPT-4 access. To start PentestGPT, please use <pentestgpt --reasoning_model=gpt-4> #### Test connection for OpenAI api (GPT-3.5) 2. You're connected with OpenAI API. You have GPT-3.5 access. To start PentestGPT, please use <pentestgpt --reasoning_model=gpt-3.5-turbo-16k>
- notice: if you have not linked a payment method to your OpenAI account, you will see error messages.
- The ChatGPT cookie solution is deprecated and not recommended. You may still use it by running
pentestgpt --reasoning_model=gpt-4 --useAPI=False
.
- Clone the repository to your local environment.
- Ensure that
poetry
is installed. If not, please refer to the poetry installation guide.
-
You are recommended to run:
- (recommended) -
pentestgpt --reasoning_model=gpt-4-turbo
to use the latest GPT-4-turbo API. pentestgpt --reasoning_model=gpt-4
if you have access to GPT-4 API.pentestgpt --reasoning_model=gpt-3.5-turbo-16k
if you only have access to GPT-3.5 API.
- (recommended) -
-
To start, run
pentestgpt --args
.--help
show the help message--reasoning_model
is the reasoning model you want to use.--parsing_model
is the parsing model you want to use.--useAPI
is whether you want to use OpenAI API. By default it is set toTrue
.--log_dir
is the customized log output directory. The location is a relative directory.--logging
defines if you would like to share the logs with us. By default it is set toFalse
.
-
The tool works similar to msfconsole. Follow the guidance to perform penetration testing.
-
In general, PentestGPT intakes commands similar to chatGPT. There are several basic commands.
- The commands are:
help
: show the help message.next
: key in the test execution result and get the next step.more
: let PentestGPT to explain more details of the current step. Also, a new sub-task solver will be created to guide the tester.todo
: show the todo list.discuss
: discuss with the PentestGPT.google
: search on Google. This function is still under development.quit
: exit the tool and save the output as log file (see the reporting section below).
- You can use <SHIFT + right arrow> to end your input (and is for next line).
- You may always use
TAB
to autocomplete the commands. - When you're given a drop-down selection list, you can use cursor or arrow key to navigate the list. Press
ENTER
to select the item. Similarly, use <SHIFT + right arrow> to confirm selection.
The user can submit info about:- tool: output of the security test tool used
- web: relevant content of a web page
- default: whatever you want, the tool will handle it
- user-comments: user comments about PentestGPT operations
- The commands are:
-
In the sub-task handler initiated by
more
, users can execute more commands to investigate into a specific problem:- The commands are:
help
: show the help message.brainstorm
: let PentestGPT brainstorm on the local task for all the possible solutions.discuss
: discuss with PentestGPT about this local task.google
: search on Google. This function is still under development.continue
: exit the subtask and continue the main testing session.
- The commands are:
- [Update] If you would like us to collect the logs to improve the tool, please run
pentestgpt --logging
. We will only collect the LLM usage, without any information related to your OpenAI key. - After finishing the penetration testing, a report will be automatically generated in
logs
folder (if you quit withquit
command). - The report can be printed in a human-readable format by running
python3 utils/report_generator.py <log file>
. A sample reportsample_pentestGPT_log.txt
is also uploaded.
PentestGPT now support local LLMs, but the prompts are only optimized for GPT-4.
- To use local GPT4ALL model, you may run
pentestgpt --reasoning_model=gpt4all --parsing_model=gpt4all
. - To select the particular model you want to use with GPT4ALL, you may update the
module_mapping
class inpentestgpt/utils/APIs/module_import.py
. - You can also follow the examples of
module_import.py
,gpt4all.py
andchatgpt_api.py
to create API support for your own model.
Please cite our paper at:
@misc{deng2023pentestgpt,
title={PentestGPT: An LLM-empowered Automatic Penetration Testing Tool},
author={Gelei Deng and Yi Liu and Víctor Mayoral-Vilches and Peng Liu and Yuekang Li and Yuan Xu and Tianwei Zhang and Yang Liu and Martin Pinzger and Stefan Rass},
year={2023},
eprint={2308.06782},
archivePrefix={arXiv},
primaryClass={cs.SE}
}
Distributed under the MIT License. See LICENSE.txt
for more information.
The tool is for educational purpose only and the author does not condone any illegal use. Use as your own risk.
- Gelei Deng - - [email protected]
- Víctor Mayoral Vilches - - [email protected]
- Yi Liu - [email protected]
- Peng Liu - [email protected]
- Zhang Jian - [email protected]
- Yuekang Li - [email protected]