Divide the workload to multiple computers - Printable Version +- hashcat Forum ( https://hashcat.net/forum)+-- Forum: Deprecated; Ancient Versions ( https://hashcat.net/forum/forum-46.html)+--- Forum: Feature Requests ( https://hashcat.net/forum/forum-7.html)+--- Thread: Divide the workload to multiple computers ( /thread-31.html) |

Divide the workload to multiple computers - richardsguy - 05-17-2010
How about a way to divide the workload to multiple computers so that if something takes 10 days to do, then it will reduce the time by 3 fold. Something for bruteforce implementation. For example, suppose a hash has 100 steps, so u have an option to divide it into multiple stages i.e 3 stages, 4, 5 etc. So on one computer you run stage 1, while at the same time, on computer 2, you run stage 2, etc RE: Divide the workload to multiple computers - atom - 05-17-2010
(05-17-2010, 04:44 AM)richardsguy Wrote: How about a way to divide the workload to multiple computers so that if something takes 10 days to do, then it will reduce the time by 3 fold. Something for bruteforce implementation. For example, suppose a hash has 100 steps, so u have an option to divide it into multiple stages i.e 3 stages, 4, 5 etc. So on one computer you run stage 1, while at the same time, on computer 2, you run stage 2, etc this is already supported by hashcat and oclHashcat by using -s and -l parameter in combination. example: wordlist contains 10000 words and you have 4 pcs (all of same speed): -s 0 -l 2500 -s 2500 -l 2500 -s 5000 -l 2500 -s 7500 -l 2500 -- atom |