Skip to content

CloudScraper: Tool to enumerate targets in search of cloud resources. S3 Buckets, Azure Blobs, Digital Ocean Storage Space.

License

Notifications You must be signed in to change notification settings

khast3x/CloudScraper

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

18 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

logo

CloudScraper is a Tool to spider and scrape targets in search of cloud resources. Plug in a URL and it will spider and search the source of spidered pages for strings such as 's3.amazonaws.com', 'windows.net' and 'digitaloceanspaces'. AWS, Azure, Digital Ocean resources are currently supported.

Pre-Requisites

Non-Standard Python Libraries:

  • requests
  • argparse
  • beautifulsoup

Created with Python 3.6

General

This tool was inspired by a recent talk by Bryce Kunz. The talk Blue Cloud of Death: Red Teaming Azure takes us through some of the lesser known common information disclosures outside of the ever common S3 Buckets.

Usage:

usage: CloudScraper.py [-h] [-u URL] [-d DEPTH] [-l TARGETLIST]

    optional arguments:
      -h, --help     show this help message and exit
      -u URL         Target Scope
      -d DEPTH       Max Depth of links Default: 25
      -l TARGETLIST  Location of text file of Line Delimited targets

Example:

$ python3 CloudScraper.py -u https://rottentomatoes.com

ToDo

  • Multithread Functionality
  • Add key word customization

Various:

To add keywords, simply add to the list in the parser function.

Docker

To use CloudScraper as a container, simply pull the repository and build the image like so:

$ git clone https://github.com/jordanpotti/CloudScraper.git   
$ docker build -t cloudscraper CloudScraper/
$ docker run -it cloudscraper -u example.com

Contribute

Sharing is caring! Pull requests welcome, things like adding support for more detections, multithreading etc are highly desired :)

Why

So Bryce Kunz actually made a tool to do something similar but it used scrapy and I wanted to build something myself that didn't depend on Python2 or any scraping modules such as scrapy. I did end up using BeautifulSoup to parse for href links for spidering only. Hence, CloudScraper was born. The benefit of using raw regex's instead of parsing for href links, is that many times, these are not included in href links, they can be buried in JS or other various locations. CloudScraper grabs the entire page and uses a regex to look for links. This also has its flaws such as grabbing too much or too little but at least we know we are covering our bases :)

About

CloudScraper: Tool to enumerate targets in search of cloud resources. S3 Buckets, Azure Blobs, Digital Ocean Storage Space.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 91.1%
  • Dockerfile 8.9%