By default, BBOT saves its output in TXT, JSON, and CSV formats. The filenames are logged at the end of each scan:
Every BBOT scan gets a unique and mildly-entertaining name like demonic_jimmy
. Output for that scan, including scan stats and any web screenshots, etc., are saved to a folder by that name in ~/.bbot/scans
. The most recent 20 scans are kept, and older ones are removed. You can change the location of BBOT's output with --output
, and you can also pick a custom scan name with --name
.
If you reuse a scan name, it will append to its original output files and leverage the previous.
Multiple simultaneous output formats are possible because of output modules. Output modules are similar to normal modules except they are enabled with -om
.
The stdout
output module is what you see when you execute BBOT in the terminal. By default it looks the same as the txt
module, but it has options you can customize. You can filter by event type, choose the data format (text
, json
), and which fields you want to see:
Config Option | Type | Description | Default |
---|---|---|---|
modules.stdout.accept_dupes | bool | Whether to show duplicate events, default True | True |
modules.stdout.event_fields | list | Which event fields to display | [] |
modules.stdout.event_types | list | Which events to display, default all event types | [] |
modules.stdout.format | str | Which text format to display, choices: text,json | text |
modules.stdout.in_scope_only | bool | Whether to only show in-scope events | False |
txt
output is tab-delimited, so it's easy to grep:
# grep out only the DNS_NAMEs
cat ~/.bbot/scans/extreme_johnny/output.txt | grep '[DNS_NAME]' | cut -f2
evilcorp.com
www.evilcorp.com
mail.evilcorp.com
The csv
output module produces a CSV like this:
Event type | Event data | IP Address | Source Module | Scope Distance | Event Tags |
---|---|---|---|---|---|
DNS_NAME | evilcorp.com | 1.2.3.4 | TARGET | 0 | a-record,cdn-github,distance-0,domain,in-scope,mx-record,ns-record,resolved,soa-record,target,txt-record |
DNS_NAME | www.evilcorp.com | 2.3.4.5 | certspotter | 0 | a-record,aaaa-record,cdn-github,cname-record,distance-0,in-scope,resolved,subdomain |
URL | http://www.evilcorp.com | 2.3.4.5 | httpx | 0 | a-record,aaaa-record,cdn-github,cname-record,distance-0,in-scope,resolved,subdomain |
DNS_NAME | admin.evilcorp.com | 5.6.7.8 | otx | 0 | a-record,aaaa-record,cloud-azure,cname-record,distance-0,in-scope,resolved,subdomain |
If you manually enable the json
output module, it will go to stdout:
bbot -t evilcorp.com -om json | jq
You will then see events like this:
{
"type": "IP_ADDRESS",
"id": "IP_ADDRESS:13cd09c2adf0860a582240229cd7ad1dccdb5eb1",
"data": "1.2.3.4",
"scope_distance": 1,
"scan": "SCAN:64c0e076516ae7aa6502fd99489693d0d5ec26cc",
"timestamp": 1688518967.740472,
"resolved_hosts": ["1.2.3.4"],
"parent": "DNS_NAME:2da045542abbf86723f22383d04eb453e573723c",
"tags": ["distance-1", "ipv4", "internal"],
"module": "A",
"module_sequence": "A"
}
You can filter on the JSON output with jq
:
# pull out only the .data attribute of every DNS_NAME
$ jq -r 'select(.type=="DNS_NAME") | .data' ~/.bbot/scans/extreme_johnny/output.json
evilcorp.com
www.evilcorp.com
mail.evilcorp.com
BBOT supports output via webhooks to discord
, slack
, and teams
. To use them, you must specify a webhook URL either in the config:
config:
modules:
discord:
webhook_url: https://discord.com/api/webhooks/1234/deadbeef
...or on the command line:
bbot -t evilcorp.com -om discord -c modules.discord.webhook_url=https://discord.com/api/webhooks/1234/deadbeef
By default, only VULNERABILITY
and FINDING
events are sent, but this can be customized by setting event_types
in the config like so:
config:
modules:
discord:
event_types:
- VULNERABILITY
- FINDING
- STORAGE_BUCKET
...or on the command line:
bbot -t evilcorp.com -om discord -c modules.discord.event_types=["STORAGE_BUCKET","FINDING","VULNERABILITY"]
You can also filter on the severity of VULNERABILITY
events by setting min_severity
:
config:
modules:
discord:
min_severity: HIGH
The http
output module sends events in JSON format to a desired HTTP endpoint.
# POST scan results to localhost
bbot -t evilcorp.com -om http -c modules.http.url=http://localhost:8000
You can customize the HTTP method if needed. Authentication is also supported:
config:
modules:
http:
url: https://localhost:8000
method: PUT
# Authorization: Bearer
bearer: <bearer_token>
# OR
username: bob
password: P@ssw0rd
When outputting to Elastic, use the http
output module with the following settings (replace <your_index>
with your desired index, e.g. bbot
):
# send scan results directly to elasticsearch
bbot -t evilcorp.com -om http -c \
modules.http.url=http://localhost:8000/<your_index>/_doc \
modules.http.siem_friendly=true \
modules.http.username=elastic \
modules.http.password=changeme
Alternatively, via a preset:
config:
modules:
http:
url: http://localhost:8000/<your_index>/_doc
siem_friendly: true
username: elastic
password: changeme
The splunk
output module sends events in JSON format to a desired splunk instance via HEC.
You can customize this output with the following config options:
config:
modules:
splunk:
# The full URL with the URI `/services/collector/event`
url: https://localhost:8088/services/collector/event
# Generated from splunk webui
hectoken: xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxxxxx
# Defaults to `main` if not set
index: my-specific-index
# Defaults to `bbot` if not set
source: /my/source.json
The asset_inventory
module produces a CSV like this:
Host | Provider | IP(s) | Status | Open Ports |
---|---|---|---|---|
evilcorp.com | cdn-github | 1.2.3.4 | Active | 80,443 |
www.evilcorp.com | cdn-github | 2.3.4.5 | Active | 22,80,443 |
admin.evilcorp.com | cloud-azure | 5.6.7.8 | N/A |
The sqlite
output module produces a SQLite database containing all events, scans, and targets. By default, it will be saved in the scan directory as output.sqlite
.
# specifying a custom database path
bbot -t evilcorp.com -om sqlite -c modules.sqlite.database=/tmp/bbot.sqlite
The postgres
output module allows you to ingest events, scans, and targets into a Postgres database. By default, it will connect to the server on localhost
with a username of postgres
and password of bbotislife
. You can change this behavior in the config.
# specifying an alternate database
bbot -t evilcorp.com -om postgres -c modules.postgres.database=custom_bbot_db
config:
modules:
postgres:
host: psq.fsociety.local
database: custom_bbot_db
port: 5432
username: postgres
password: bbotislife
The mysql
output module allows you to ingest events, scans, and targets into a MySQL database. By default, it will connect to the server on localhost
with a username of root
and password of bbotislife
. You can change this behavior in the config.
# specifying an alternate database
bbot -t evilcorp.com -om mysql -c modules.mysql.database=custom_bbot_db
config:
modules:
mysql:
host: mysql.fsociety.local
database: custom_bbot_db
port: 3306
username: root
password: bbotislife
The subdomains
output module produces simple text file containing only in-scope and resolved subdomains:
evilcorp.com
www.evilcorp.com
mail.evilcorp.com
portal.evilcorp.com
Neo4j is the funnest (and prettiest) way to view and interact with BBOT data.
- You can get Neo4j up and running with a single docker command:
# start Neo4j in the background with docker
docker run -d -p 7687:7687 -p 7474:7474 -v "$(pwd)/neo4j/:/data/" -e NEO4J_AUTH=neo4j/bbotislife neo4j
- After that, run bbot with
-om neo4j
bbot -f subdomain-enum -t evilcorp.com -om neo4j
- Log in at http://localhost:7474 with
neo4j
/bbotislife
Neo4j uses the Cypher Query Language for its graph query language. Cypher uses common clauses to craft relational queries and present the desired data in multiple formats.
Cypher queries can be broken down into three required pieces; selection, filter, and presentation. The selection piece identifies what data that will be searched against - 90% of the time the "MATCH" clause will be enough but there are means to read from csv or json data files. In all of these examples the "MATCH" clause will be used. The filter piece helps to focus in on the required data and used the "WHERE" clause to accomplish this effort (most basic operators can be used). Finally, the presentation section identifies how the data should be presented back to the querier. While neo4j is a graph database, it can be used in a traditional table view.
A simple query to grab every URL event with ".com" in the BBOT data field would look like this:
MATCH (u:URL) WHERE u.data contains ".com" RETURN u
In this query the following can be identified:
- Within the MATCH statement "u" is a variable and can be any value needed by the user while the "URL" label is a direct relationship to the BBOT event type.
- The WHERE statement allows the query to filter on any of the BBOT event properties like data, tag, or even the label itself.
- The RETURN statement is a general presentation of the whole URL event but this can be narrowed down to present any of the specific properties of the BBOT event (
RETURN u.data, u.tags
).
The following are a few recommended queries to get started with:
// Get all "in-scope" DNS Nodes and return just data and tags properties
MATCH (n:DNS_NAME)
WHERE "in-scope" IN n.tags
RETURN n.data, n.tags
// Get the count of labels/BBOT events in the Neo4j Database
MATCH (n)
RETURN labels(n), count(n)
// Get a graph of open ports associated with each domain
MATCH z = ((n:DNS_NAME) --> (p:OPEN_TCP_PORT))
RETURN z
// Get all domains and IP addresses with open TCP ports
MATCH (n) --> (p:OPEN_TCP_PORT)
WHERE "in-scope" in n.tags and (n:DNS_NAME or n:IP_ADDRESS)
WITH *, TAIL(SPLIT(p.data, ':')) AS port
RETURN n.data, collect(distinct port)
// Clear the database
MATCH (n) DETACH DELETE n
This is not an exhaustive list of clauses, filters, or other means to use cypher and should be considered a starting point. To build more advanced queries consider reading Neo4j's Cypher documentation.
Additional note: these sample queries are dependent on the existence of the data in the target neo4j database.
The web_parameters
output module will utilize BBOT web parameter extraction capabilities, and output the resulting parameters to a file (web_parameters.txt, by default). Web parameter extraction is disabled by default, but will automatically be enabled when a module is included that consumes WEB_PARAMETER events (including the web_parameters
output module itself).
This can be useful for those who want to discover new common web parameters or those which may be associated with a specific target or organization. This could be very useful for further parameter bruteforcing, or even fed back into bbot via the paramminer modules. For example:
bbot -t evilcorp.com -m paramminer_getparams -c modules.paramminer_getparams.wordlist=/path/to/your/new/wordlist.txt