Generates a random robots.txt deny list to throw script kiddies off the scent.
OSX or Debian-based Linux
Python 3.6+
Check that /usr/share/dict/words exists for your os and shuf
or gshuf
commands work.
If not, in ubuntu install sudo apt install wamerican coreutils
or in OSX brew install coreutils
Change the global variable numberofwords
in each .py file to the number of deny paths to generate in the robots.txt file.
To run on OSX, simply python3 gen-robots-osx.py
or python3.6 gen-robots-osx.py
if you have multiple pythons installed.
To run on Debian/Ubuntu, simply python3 gen-robots-ubuntu.py
or python3.6 gen-robots-ubuntu.py
if you have multiple pythons installed.
Regardless of version, a robots.txt
file will be generated in the same folder where you execute the script from.