hakrawler icon indicating copy to clipboard operation
hakrawler copied to clipboard

Simple, fast web crawler designed for easy, quick discovery of endpoints and assets within a web application

Results 18 hakrawler issues
Sort by recently updated
recently updated
newest added

When more than one depth is used ("-d 2,3,4,...,n"), the information from where the found URLs are not available. A feature that provides data on where the found URLs are.

What I mean is For example if there is some website named xyz.com and lets say im getting js files.... but what i wanna see is the routes of the...

![imagem](https://user-images.githubusercontent.com/7319156/183070560-bfdce96e-a027-4850-9142-82cfa157e385.png) Command: ``` hakrawler -d 10 -t 100 -subs -u ``` The wordlist used was gerated from httpx and have less then 600 targets. Today was the fifth time that...

echo https://www.google.com | docker run --rm -i hackluke/hakrawler -subs Error parsing URL: parse "https://www.google.com ": invalid character " " in host name No way to make it works :(

`url.Parse` does not return an error most of the time because it treats random strings as relative URLs.

Hi Hakluke, This pull request aims to add a condition to help separating the links with GET parameters and normal URL links in the result. I added a small condition...

Hello. First, I appreciate your script and hard work for this project that we can all use. I think its great. I have been struggling to crawl a website. Read...

Hello. Hakrawler is great, but unfortunately it doesnt support SOCKS or HTTPS proxy. Please add support of SOCKS/HTTPS.