crawpy/crawpy.py

271 lines
11 KiB
Python
Executable File
Raw Permalink Blame History

This file contains invisible Unicode characters

This file contains invisible Unicode characters that are indistinguishable to humans but may be processed differently by a computer. If you think that this is intentional, you can safely ignore this warning. Use the Escape button to reveal them.

#!/usr/bin/python3
import sys
import argparse
from colorama import *
import src.config
import os
from src.RequestEngine import RequestEngine
from src.Banner import Banner
import asyncio
import concurrent.futures
import subprocess
import random
base_path = os.path.abspath(os.path.dirname(__file__))
init()
GREEN = Fore.GREEN
RED = Fore.RED
RESET = Fore.RESET
BLUE = Fore.BLUE
YELLOW = Fore.YELLOW
MAGENTA = Fore.MAGENTA
user_agents = [
"Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/103.0.0.0 Safari/537.36",
"Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/87.0.0.0 Safari/537.36",
"Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/82.0.0.0 Safari/537.36",
"Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:102.0) Gecko/20100101 Firefox/102.0",
"Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:102.0) Gecko/20100101 Firefox/92.0",
"Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/103.0.5060.114 Safari/537.36 Edg/103.0.1264.62"
"Mozilla/5.0 (Macintosh; Intel Mac OS X 12_4) AppleWebKit/605.1.15 (KHTML, like Gecko) Version/15.4 Safari/605.1.15",
]
def new_session(url):
# we need to strip -lt and -l flags from it
sys.stdout.write(f"{MAGENTA}[*] Fuzzing: {url}/FUZZ\n{RESET}")
args = " ".join(sys.argv)
tmp = args.split('-lt ', maxsplit=1)[-1].split(maxsplit=1)[0]
args = args.replace(f" -lt {tmp}","")
tmp = args.split('-l ', maxsplit=1)[-1].split(maxsplit=1)[0]
args = args.replace(f" -l {tmp}","")
args += f" -u {url}/FUZZ"
args = args.split(" ")
if '-s' in args:
proc = subprocess.Popen(args, stdout=subprocess.PIPE, stderr=subprocess.DEVNULL)
stdout, stderr = proc.communicate()
else:
proc = subprocess.Popen(args)
stdout, stderr = proc.communicate()
def flush_input():
try:
import msvcrt
while msvcrt.kbhit():
msvcrt.getch()
except ImportError:
import termios #for linux/unix
termios.tcflush(sys.stdin, termios.TCIOFLUSH)
if __name__ == "__main__":
parser = argparse.ArgumentParser()
# Mandatory options
parser.add_argument("-u", "--url", dest="url",help="URL")
parser.add_argument("-w", "--wordlist", dest="wordlist", help="Wordlist")
# Optional options
parser.add_argument("-t", "--threads", dest="threads", help="Size of the semaphore pool",default=50)
parser.add_argument("-rc", "--recursive-codes", dest="recursive_codes", default=[], help="Recursive codes to scan recursively\nExample: 301,302,307")
parser.add_argument("-rp", "--recursive-paths", dest="recursive_paths", default=[], help="Recursive paths to scan recursively, please note that only given recursive paths will be scanned initially\nExample: admin,support,js,backup")
parser.add_argument("-rd", "--recursive-depth", dest="recursive_depth",default=-1, type=int, help="Recursive scan depth\nExample: 2")
parser.add_argument("-e", "--extension", dest="extensions",default=[],help="Add extensions at the end. Seperate them with comas \n Example: -x .php,.html,.txt")
parser.add_argument("-to", "--timeout", dest="timeout",default=5, help="Timeouts, I suggest you to not use this option because it is procudes lots of erros now which I was not able to solve why")
parser.add_argument("-follow", "--follow-redirects", dest="follow_redirects",action='store_true', help="Follow redirects")
# Filter options
parser.add_argument("-ac","--auto-calibrate", dest="auto_calibrate", action='store_true', help="Automatically calibre filter stuff")
parser.add_argument("-fc","--filter-code", dest="filter_code", default=[], help="Filter status code")
parser.add_argument("-fs","--filter-size", dest="filter_size", default=[], help="Filter size")
parser.add_argument("-fw","--filter-word", dest="filter_word", default=[], help="Filter words")
parser.add_argument("-fl","--filter-line", dest="filter_line", default=[], help="Filter line")
parser.add_argument("-k", "--ignore-ssl", dest="verify_ssl",action='store_false',help="Ignore untrusted SSL certificate")
parser.add_argument("-m","--max-retry", dest="max_retry", default=3,help="Max retry")
parser.add_argument("-H","--headers", dest="headers", action='append', help="Headers, you can set the flag multiple times.For example: -H \"X-Forwarded-For: 127.0.0.1\", -H \"Host: foobar\" ")
parser.add_argument("-o","--output", dest="output_file", default=f"{base_path}/reports/", help="Output folder")
parser.add_argument("-gr","--generate-report", dest="generate_report", action="store_true", help="If you want crawpy to generate a report, default path is crawpy/reports/<url>.txt" )
# list related args
parser.add_argument("-l","--list", dest="url_list", help="Takes a list of urls as input and runs crawpy on via multiprocessing\n-l ./urls.txt")
parser.add_argument("-lt","--list-threads", type=int, default=10, dest="list_threads", help="Number of threads for running crawpy parallely when running with list of urls")
parser.add_argument("-s","--silent", action='store_true', dest="silent", help="Make crawpy not produce output")
parser.add_argument("-X","--http-method", dest="http_method", default="GET",help="HTTP request method")
parser.add_argument("-p","--proxy", dest="proxy_server",help="Proxy server, ex: 'http://127.0.0.1:8080'")
args = parser.parse_args()
if len(sys.argv) < 2:
parser.print_usage()
sys.exit(0)
Banner().greet()
if args.url_list != None:
url_file = open(args.url_list,"r").read()
urls = url_file.split()
sys.stdout.write(f"{YELLOW}[!] URL List supplied: {args.url_list}\n{RESET}")
sys.stdout.write(f"{YELLOW}[!] Yielding url list on {args.list_threads} threads\n{RESET}")
sys.stdout.write(f"{YELLOW}[!] Wordlist: {args.wordlist}\n{RESET}")
sys.stdout.write(f"{YELLOW}[!] Semaphore pool for each session: {args.threads}\n{RESET}")
with concurrent.futures.ThreadPoolExecutor(max_workers=args.list_threads) as executor:
res = executor.map(new_session, urls)
#os.execve()
sys.exit()
conf = src.config.config
# initialize
conf['base_url'] = args.url
conf['wordlist'] = args.wordlist
conf['threads'] = int(args.threads)
conf['http_method'] = args.http_method
conf['timeout'] = args.timeout
conf['verify_ssl'] = args.verify_ssl
conf['follow_redirects'] = args.follow_redirects
conf['recursive_depth'] = args.recursive_depth
conf['headers']['User-Agent'] = random.choice(user_agents)
if len(args.recursive_paths)>0:
conf['recursive_paths'] = args.recursive_paths.split(",")
conf['auto_calibrate'] = args.auto_calibrate
conf['proxy_server'] = args.proxy_server
# output
if args.generate_report:
conf['generate_report_enabled'] = True
if "reports/" in args.output_file:
_url = args.url.replace("://","_").replace(".","_").replace("FUZZ","").replace("/","_") # i know this is ugly but it works
conf['output_file_txt'] = open(f"{args.output_file}{_url}.txt","w")
conf['output_file_html'] = open(f"{args.output_file}{_url}.html","w")
else:
_url = args.url.replace("://","_").replace(".","_").replace("FUZZ","").replace("/","_") # i know this is ugly but it works
conf['output_file_txt'] = open(f"{args.output_file}/{_url}.txt","w")
conf['output_file_html'] = open(f"{args.output_file}/{_url}.html","w")
if args.headers != None:
for header in args.headers:
conf['headers'][header.split(':')[0]] = header.split(':')[1]
if len(args.filter_word) > 0:
conf['filter_word'] = [int(_) for _ in args.filter_word.split(",")]
if len(args.filter_code) > 0:
conf['filter_code'] = [int(_) for _ in args.filter_code.split(",")]
if len(args.filter_size) > 0:
conf['filter_size'] = [int(_) for _ in args.filter_size.split(",")]
if len(args.filter_line) > 0:
conf['filter_line'] = [int(_) for _ in args.filter_line.split(",")]
# extension
if len(args.extensions) > 0:
conf['extensions'] = args.extensions.split(",")
if len(args.recursive_codes) > 0:
conf['is_recursive'] = True
conf['recursive_codes'] = [int(_) for _ in args.recursive_codes.split(",")]
conf['filter_code_backup'] = conf['filter_code'].copy()
conf['filter_size_backup'] = conf['filter_size'].copy()
conf['filter_word_backup'] = conf['filter_word'].copy()
if args.recursive_depth == -1:
sys.stdout.write(f"{RED}[!] You must enter a recursion depth! {RESET}\n")
sys.exit(1)
#info section
sys.stdout.write(f"{BLUE}[*] Url: {conf['base_url']} \n{RESET}")
sys.stdout.write(f"{BLUE}[*] Wordlist: {conf['wordlist']}\n{RESET}")
sys.stdout.write(f"{BLUE}[*] Semaphore pool: {conf['threads']}\n{RESET}")
sys.stdout.write(f"{BLUE}[*] Extensions: {conf['extensions']}\n{RESET}")
sys.stdout.write(f"{BLUE}[*] Follow redirects: {conf['follow_redirects']}\n{RESET}")
sys.stdout.write(f"{BLUE}[*] HTTP Method {conf['http_method']}\n{RESET}")
sys.stdout.write(f"{BLUE}[*] Headers {conf['headers']}\n{RESET}")
# Matchers
if len(conf['filter_word']) > 0:
sys.stdout.write(f"{YELLOW}[*] Filter words: {conf['filter_word']}\n{RESET}")
if len(conf['filter_code']) > 0:
sys.stdout.write(f"{YELLOW}[*] Filter codes: {conf['filter_code']}\n{RESET}")
if len(conf['filter_size']) > 0:
sys.stdout.write(f"{YELLOW}[*] Filter size: {conf['filter_size']}\n{RESET}")
if len(conf['filter_line']) > 0:
sys.stdout.write(f"{YELLOW}[*] Filter line: {conf['filter_line']}\n{RESET}")
if conf['proxy_server']:
sys.stdout.write(f"{YELLOW}[*] Proxy server: {conf['proxy_server']}{RESET}\n")
if conf['is_recursive']:
sys.stdout.write(f"{YELLOW}[!] Recursive scan enabled with depth {conf['recursive_depth']}{RESET}\n")
if conf['generate_report_enabled']:
sys.stdout.write(f"{YELLOW}[!] Generate report enabled, writing: {conf['output_file_txt'].name}{RESET}\n")
sys.stdout.write(f"{YELLOW}[!] Generate report enabled, writing: {conf['output_file_html'].name}{RESET}\n")
if os.name != 'nt':
os.system("stty -echo")
requester = RequestEngine(conf)
loop = asyncio.get_event_loop()
#loop.set_debug(1)
sys.stdout.write(f"{BLUE}-------------\n{RESET}")
try:
loop.run_until_complete(requester.run())
except KeyboardInterrupt:
sys.stdout.write(f"{RED}[!] Keyboard interrupt recieved, exiting ...{RESET}\n")
try:
conf['output_file_txt'].close() # try to close the file on interrupts
conf['output_file_html'].close() # try to close the file on interrupts
except:
pass
pass
except:
pass
if os.name != 'nt':
os.system("stty echo")
if conf['generate_report_enabled']:
conf['output_file_txt'].close()
template_html = open(f"{base_path}/reports/template.html").read()
conf['output_file_html'].write(template_html.replace("placeholder",conf['html_report']))
conf['output_file_html'].close()
flush_input()