A tool to find open S3 buckets and dump their contents💧
usage: s3scanner [-h] [--version] [--threads n] [--endpoint-url ENDPOINT_URL] [--endpoint-address-style {path,vhost}] [--insecure] {scan,dump} ... s3scanner: Audit unsecured S3 buckets by Dan Salmon - github.com/sa7mon, @bltjetpack optional arguments: -h, --help show this help message and exit --version Display the current version of this tool --threads n, -t n Number of threads to use. Default: 4 --endpoint-url ENDPOINT_URL, -u ENDPOINT_URL URL of S3-compliant API. Default: https://s3.amazonaws.com --endpoint-address-style {path,vhost}, -s {path,vhost} Address style to use for the endpoint. Default: path --insecure, -i Do not verify SSL mode: {scan,dump} (Must choose one) scan Scan bucket permissions dump Dump the contents of buckets
If you've found this tool useful, please consider donating to support its development. You can find sponsor options on the side of this repo page or in FUNDING.yml
Huge thank you to tines for being an ongoing sponsor of this project.
pip3 install s3scanner
or via Docker:
docker build . -t s3scanner:latest
docker run --rm s3scanner:latest scan --bucket my-buket
or from source:
git clone git@github.com:sa7mon/S3Scanner.git
cd S3Scanner
pip3 install -r requirements.txt
python3 -m S3Scanner
- ⚡️ Multi-threaded scanning
- 🔭 Supports tons of S3-compatible APIs
- 🕵️♀️ Scans all bucket permissions to find misconfigurations
- 💾 Dump bucket contents to a local folder
- 🐳 Docker support
- Scan AWS buckets listed in a file with 8 threads
$ s3scanner --threads 8 scan --buckets-file ./bucket-names.txt
- Scan a bucket in Digital Ocean Spaces
$ s3scanner --endpoint-url https://sfo2.digitaloceanspaces.com scan --bucket my-bucket
- Dump a single AWS bucket
$ s3scanner dump --bucket my-bucket-to-dump
- Scan a single Dreamhost Objects bucket which uses the vhost address style and an invalid SSL cert
$ s3scanner --endpoint-url https://objects.dreamhost.com --endpoint-address-style vhost --insecure scan --bucket my-bucket
S3Scanner
can scan and dump buckets in S3-compatible APIs services other than AWS by using the
--endpoint-url
argument. Depending on the service, you may also need the --endpoint-address-style
or --insecure
arguments as well.
Some services have different endpoints corresponding to different regions
Note: S3Scanner
currently only supports scanning for anonymous user permissions of non-AWS services
Service | Example Endpoint | Address Style | Insecure ? |
---|---|---|---|
DigitalOcean Spaces (SFO2 region) | https://sfo2.digitaloceanspaces.com | path | No |
Dreamhost | https://objects.dreamhost.com | vhost | Yes |
Linode Object Storage (eu-central-1 region) | https://eu-central-1.linodeobjects.com | vhost | No |
Scaleway Object Storage (nl-ams region) | https://s3.nl-ams.scw.cloud | path | No |
Wasabi Cloud Storage | http://s3.wasabisys.com/ | path | Yes |
📚 Current status of non-AWS APIs can be found in the project wiki
This tool will attempt to get all available information about a bucket, but it's up to you to interpret the results.
Possible permissions for buckets:
- Read - List and view all files
- Write - Write files to bucket
- Read ACP - Read all Access Control Policies attached to bucket
- Write ACP - Write Access Control Policies to bucket
- Full Control - All above permissions
Any or all of these permissions can be set for the 2 main user groups:
- Authenticated Users
- Public Users (those without AWS credentials set)
- Individual users/groups (out of scope of this tool)
What this means: Just because a bucket doesn't allow reading/writing ACLs doesn't mean you can't read/write files in the bucket. Conversely, you may be able to list ACLs but not read/write to the bucket
MIT