Wordhound doesn't save any data
#1
It's crawling good and finishes after a time but the the output file I determined isn't there. It saves a file with the webpage name in the Wordhound folder but it's empty. What I tried:
-running with sudo
-setting the output files manually (to the wordhound folder itself and in other folders)
-leaving the option
-different webpages


Basically I want to extract as many words as possible from a webpage. Are there alternatives probably?

(Not directly Hashcat related but I take the risk to ask here.) Smile
#2
https://bitbucket.org/mattinfosec/wordhound/issues