Where should I run kube-hunter?
Run kube-hunter on any machine (including your laptop), select Remote scanning and give the IP address or domain name of your Kubernetes cluster. This will give you an attackers-eye-view of your Kubernetes setup.
You can run kube-hunter directly on a machine in the cluster, and select the option to probe all the local network interfaces.
You can also run kube-hunter in a pod within the cluster. This gives an indication of how exposed your cluster would be in the event that one of your application pods is compromised (through a software vulnerability, for example).
By default, kube-hunter will open an interactive session, in which you will be able to select one of the following scan options. You can also specify the scan option manually from the command line. These are your options:
- Remote scanning To specify remote machines for hunting, select option 1 or use the
./kube-hunter.py --remote some.node.com
- Internal scanning To specify internal scanning, you can use the
--internaloption. (this will scan all of the machine's network interfaces) Example:
- Network scanning To specify a specific CIDR to scan, use the
./kube-hunter.py --cidr 192.168.0.0/24
Active hunting is an option in which kube-hunter will exploit vulnerabilities it finds, in order to explore for further vulnerabilities. The main difference between normal and active hunting is that a normal hunt will never change state of the cluster, while active hunting can potentially do state-changing operations on the cluster, which could be harmful.
By default, kube-hunter does not do active hunting. To active hunt a cluster, use the
./kube-hunter.py --remote some.domain.com --active
List of tests
You can see the list of tests with the
To see active hunting tests as well as passive:
./kube-hunter.py --list --active
To control logging, you can specify a log level, using the
./kube-hunter.py --active --log WARNINGAvailable log levels are:
- INFO (default)
./kube-hunter.py --cidr 192.168.0.0/24 --mappingThis will output all the Kubernetes nodes kube-hunter has found.
There are three methods for deploying kube-hunter:
You can run the kube-hunter python code directly on your machine.
You will need the following installed:
- python 2.7
Install module dependencies:
git clone firstname.lastname@example.org:aquasecurity/kube-hunter.git
pip install -r requirements.txt
In the case where you have python 3.x in the path as your default, and python2 refers to a python 2.7 executable, use "python2 -m pip install -r requirements.txt"
Aqua Security maintains a containerised version of kube-hunter at
aquasec/kube-hunter. This container includes this source code, plus an additional (closed source) reporting plugin for uploading results into a report that can be viewed at kube-hunter.aquasec.com. Please note that running the
aquasec/kube-huntercontainer and uploading reports data are subject to additional terms and conditions.
The Dockerfile in this repository allows you to build a containerised version without the reporting plugin.
If you run the kube-hunter container with the host network it will be able to probe all the interfaces on the host:
docker run -it --rm --network host aquasec/kube-hunter
Note for Docker for Mac/Windows: Be aware that the "host" for Docker for Mac or Windows is the VM which Docker runs containers within. Therefore specifying
--network hostallows kube-hunter access to the network interfaces of that VM, rather than those of your machine. By default kube-hunter runs in interactive mode. You can also specify the scanning option with the parameters described above e.g.
docker run --rm aquasec/kube-hunter --cidr 192.168.0.0/24
This option lets you discover what running a malicious container can do/discover on your cluster. This gives a perspective on what an attacker could do if they were able to compromise a pod, perhaps through a software vulnerability. This may reveal significantly more vulnerabilities.
job.yamlfile defines a Job that will run kube-hunter in a pod, using default Kubernetes pod access settings.
- Run the job with
kubectl createwith that yaml file.
- Find the pod name with
kubectl describe job kube-hunter
- View the test results with
kubectl logs <pod name>