Skip to content

Robofinder retrieves historical #robots.txt files from #Archive.org, allowing you to uncover previously disallowed directories and paths for any domain—essential for deepening your #OSINT and #recon process.

License

Notifications You must be signed in to change notification settings

Spix0r/robofinder

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

27 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Robofinder

Robofinder is a powerful Python script designed to search for and retrieve historical robots.txt files from Archive.org for any given website. This tool is ideal for security researchers, web archivists, and penetration testers to uncover previously accessible paths or directories that were listed in a site's robots.txt.

Features

  • Fetch historical robots.txt files from Archive.org.
  • Extract and display old paths or directories that were once disallowed or listed.
  • Save results to a specified output file.
  • Silent Mode for unobtrusive execution.
  • Multi-threading support for faster processing.
  • Option to concatenate extracted paths with the base URL for easy access.
  • Debug mode for detailed execution logs.
  • Extract old parameters from robots.txt files.

Installation

Using pipx

Install Robofinder quickly and securely using pipx:

pipx install git+https://github.com/Spix0r/robofinder.git

Manual Installation

To install manually:

git clone https://github.com/Spix0r/robofinder.git
cd robofinder
pip install -r requirements.txt

Usage

Basic Command

If installed via pipx:

robofinder -u https://example.com

For manual installation:

python3 robofinder.py -u https://example.com

Options and Examples

  • Save output to a file:

    robofinder -u https://example.com -o results.txt
  • Silent Mode (minimal output to console):

    robofinder -u https://example.com -s
  • Concatenate paths with the base URL:

    robofinder -u https://example.com -c
  • Extract Paramters:

    robofinder -u https://example.com -p
  • Enable Debug Mode:

    robofinder -u https://example.com --debug
  • Multi-threading (default: 10 threads):

    robofinder -u https://example.com -t 10

Advanced Usage

Combine options for tailored execution:

robofinder -u https://example.com -t 10 -c -o results.txt -s

Example Output

Running Robofinder on example.com with 10 threads, silent mode, and saving just the paramters to the results.txt:

robofinder -u https://example.com -t 10 -o results.txt -s -p

Contributing

Contributions are highly welcome! If you have ideas for new features, optimizations, or bug fixes, feel free to submit a Pull Request or open an issue on the GitHub repository.


About

Robofinder retrieves historical #robots.txt files from #Archive.org, allowing you to uncover previously disallowed directories and paths for any domain—essential for deepening your #OSINT and #recon process.

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages