08-16-2019, 10:55 PM
(08-16-2019, 10:49 PM)slawson Wrote: It accepted it at 3GB file size.
I do have a question though. If I do a benchmark on my system it shows 43,000,000 H/s. When I processed my 3GB file the speed went down to about 3,000 H/s. I then broke down my file to 1Million Lines and I am running at about 4,000,000 H/s. Are there any parameters that I can set to fix that or is it just the way it works? I used -w 3 -O, but it didn't seem to help.
Hashcat is slow when it cracks a lot of hashes.
You should use mdxfind on the total list first, it can handle a lot of cracks a lot better. Then you should use mdsplit to find the remainder and run those on hashcat.