bulk extracting hashes
I was wondering if there scripts that could be run recursively against a directory of files (doc, ppt, xls, docx, pdf, zip, 7z, rar, etc..) that would call the appropriate office2john.py (for example) to extract hashes for a large number of files.  They then could be passed to hashid or similar to make sorted lists to be passed into hashcat.
You could easily write a bash script. Do something along the lines of:
find ~/officedocs -name '*.pdf' -or -name '*.doc' -or -name '*.7z' | xargs office2john.py ARG ARG

Just add each kind of office document that you want to search in the find options, then pipe it all over to xargs office2john.py
Or pass it to a case statement so you can execute the proper application e.g. 7z2john rar2john zip2john etc.

Hope this helps get you in the right direction.