Skip to content

Crawls URL to get a better image of what is tied to a website.

Notifications You must be signed in to change notification settings

ghostwond3r/wildcrawl

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

50 Commits
 
 
 
 
 
 

Repository files navigation


Bash script that crawls a target URL and everything around it to get a better image of what is tied to a website.


Here's a summary of what the script does:

  • Crawls the target URL using Hakrawler .
  • Removes duplicate links.
  • Extracts all files of certain types (e.g., PDF, DOC, ZIP, JPG) and saves them to a separate file.
  • Extracts each domain from the link crawled and removes duplicates.
  • Filters out certain domains (e.g., Facebook, Twitter, LinkedIn).
  • Extract records (AAAA, CNAME, NS, etc)
  • Extract DNS by fetching the certificate.
  • Crawl again but this time using Lynx on the domain tied to the main one.
  • Gets main IPs of all domains.
  • Extracts the title of each domain and IP.
  • Filters out any results that have a title starting with "404".
  • Search emails from each domain.
  • Retrieves server banner.

At the end, the tool gives you 6 options;

  • Injection
  • HTTP-ENUM
  • Port Scan
  • Fuzzing
  • Nikto
  • Access control


At the end you will have these files saved in /scan_1, /scan_2, etc.

  • crawl.txt
  • domains.txt
  • files.txt
  • injection_point.txt
  • ips.txt
  • report.txt
  • social_links.txt
  • temp_banner.txt
  • unique_urls.txt

image



Installation and Usage

git clone https://github.com/NeverWonderLand/wildcrawl.git
cd wildcrawl
chmod +x install.sh
./install.sh
chmod +x wildcrawl.sh

Then start with:
./wildcrawl.sh

Then type URL target (with http:// or https://)

e.g. https://example.com

About

Crawls URL to get a better image of what is tied to a website.

Resources

Stars

Watchers

Forks

Packages

No packages published

Languages