Web recon tool (find temporary files, parse robots.txt, search folders, google dorks and search domains hosted on same server)
Automatic GIT/SVN clone (using dvcs-ripper) if .git or .svn folder if found.
./crawlic.py -u http://site.tld/ -t rtf
-t : techniques to use for scanning (default rtfgd):
- r : robots.txt
- t : temporary files (~, .bak, etc)
- f : folders
- g : google dorks
- d : reverse dns search (via bing)
[*] Starting robots.txt search on http://site.tld
[!] /hidden_file.php
[*] Starting folder search on http://site.tld
[!] http://site.tld/admin/
[!] http://site.tld/private/
[*] Starting temp file search on http://site.tld
[!] http://site.tld/index.php~
[*] Crawling finished
Edit user_agent.lst, one user agent per line
Edit folders.lst, one directory per line
Edit extensions.lst, one file extension per line
Edit dorks.lst, one dork per line
Edit google_dorks, one dork per line, use %s as target url
- Python 2.x
- git/svn ripper needs LWP.pm library (Original dvcs ripper: https://github.com/kost/dvcs-ripper)
"THE BEER-WARE LICENSE" (Revision 42): Ganapati wrote this file. As long as you retain this notice you can do whatever you want with this stuff. If we meet some day, and you think this stuff is worth it, you can buy me a beer in return.