Skip to main content

Command line URL fuzzer

Project description

# PyIntruder
Simple Command Line URL Fuzzer

./ -h
Usage: ./ [options] <base url> <payload list>
(Use '$' as variable in url that will be swapped out with each payload)

Example:$.pdf payloads.txt

-h, --help show this help message and exit
-r, --redir Allow HTTP redirects
-s, --save Save HTTP response content to files
-o OUT, --out=OUT Directory to save HTTP responses

# Description
This script allows a user to quickly test many similar URLs and analyze responses. This can act as a simplified alternative to Burp Suite's "Intruder" tool (which heavily rate-limits requests in the free version......).

# Use Case

As an example, say you observe the following URL:
When accessing the URL, your browser redirects you to a page which automatically downloads a file (this could be any type of file - pdf, doc, exe, mp3, etc.). This is a common method of allowing users of a website to download content. In this particular example, the URL above seems to beg the question: "I wonder what I might find at ''? ...or at ''?"

This program automates the process of attempting to browse to each of these potentially-interesting URLs by automatically cycling through a list of custom "payloads". A user can create a list of payloads (say, for example, a list of numbers from 1 through 100) and try each payload in a particular position within the URL (use the dollar-sign character to tell the program where to swap out your payloads within the URL).

./$ payloads.txt
In the above command, where "payloads.txt" is a text file containing a list of numbers 1 - 100 (one number per line), a user can quickly determine which URLs lead somewhere interesting by comparing HTTP status code, Content-Length, or response time:

sample output:
root@kali:~# ./$ payloads.txt
Status Length Time Host
200 0 110.536
200 0 112.312
302 0 104.266


200 0 137.111
302 0 120.607
302 0 108.553

In this case, it looks like the interesting URLs are the ones that return a 302 HTTP status code (redirect). If all URLs are redirecting and you cant find any other distinguishing factors, try using the "-r" option to enable redirection. The redirected results will often contain more interesting/varying content-lengths. The program defaults to disabling the following of redirects. The reason for this is that it is usually much faster and a little less noisy/intrusive, which is good when running an initial scan.

In order to download whatever files might be available at each of these links, you can run a command like this:
./ -rs -o /path/to/save/files$ payloads-refined.txt

- The "r" option tells the program to follow redirects
- The "s" option tells the program to save HTTP responses
- The "o" option tells the program where you want to save the responses on your local machine (this option is optional; by default, if "s" is used without "o", it will save files to the current directory)
- "payloads-refined.txt" is your refined list of payloads. This can be useful in a case like this if you want to weed out a bunch of URLs that you found out don't go anywhere interesting.

If it's not already installed, make sure to [install Requests]( (try running "pip install requests").

Project details

Release history Release notifications

History Node


This version
History Node


History Node


History Node


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Filename, size & hash SHA256 hash help File type Python version Upload date
PyIntruder-0.1.3-py2-none-any.whl (7.9 kB) Copy SHA256 hash SHA256 Wheel py2 Jan 11, 2017

Supported by

Elastic Elastic Search Pingdom Pingdom Monitoring Google Google BigQuery Sentry Sentry Error logging CloudAMQP CloudAMQP RabbitMQ AWS AWS Cloud computing Fastly Fastly CDN DigiCert DigiCert EV certificate StatusPage StatusPage Status page