|
Message-ID: <5212882e871a2ece774122145215dcf4@smtp.hushmail.com> Date: Sat, 30 Jan 2016 04:44:52 +0100 From: magnum <john.magnum@...hmail.com> To: john-users@...ts.openwall.com Subject: Re: MPI with Spot Instances? On 2016-01-30 04:25, japhar81 wrote: > They basically just disappear, whatever they were in the middle of -- i > guess my question is, will the resume re-run whatever jobs those nodes were > in the middle of and didn't report back? And if one of them happens to have > hit a match, will that get saved somehow too? A cracked password is very unlikely to not end up in the pot file. The beauty of Solar's design is that if a session dies before it wrote a recently cracked password to the pot file, it will also not have written the corresponding unit of work to the session file. So a resume will almost certainly re-crack the password. Jumbo's MPI functionality is very KISS minded and doesn't rely on "reporting back" anything to anyone at all. Each node runs in its own daft world not giving a dang about the others. Each node writes its own session file just as any non-MPI session would. In fact, the code paths are *very* near 100% the same as when running --node=x/y in a single process except the "x" and "y" is filled in automagically. Worst-case scenario is supposedly that a resume will do a bit of redundant work. This is obviously by design - better safe than sorry. The default "Save" timer in john.conf is 60 seconds for Jumbo so you will hopefully not lose more than that. Some modes (eg. single w/ many salts and PRINCE regardless of salts) may be much worse than that though, to the point that a stop/resume once an hour may end up never proceeding past this hour at all. magnum > On Fri, Jan 29, 2016 at 10:20 PM, magnum <john.magnum@...hmail.com> wrote: > >> On 2016-01-29 18:12, japhar81 wrote: >> >>> Ok, so corollary question -- does the session stuff work with MPI? i.e. >>> lets say I start the spot instances externally, and mpiexec jtr with some >>> flavor of --session (on a box that wont die). If those nodes die >>> mid-process, will that be recorded in the session to enable a resume later >>> when I spin new nodes and start mpiexec again? >>> >> >> Sure (as far as I can imagine how spot instances work). Session file >> integrity is very well tested. >> >> magnum >> >> >> On Wed, Jan 27, 2016 at 4:03 PM, magnum <john.magnum@...hmail.com> wrote: >>> >>> On 2016-01-27 17:25, japhar81 wrote: >>>> >>>> I've been playing around with MPI clustering JtR for a while, and I've >>>>> managed to get it running smoothly on static nodes. What I'd like to do >>>>> next is create an auto-scaling group in AWS, using spot instances. What >>>>> this basically means is nodes will come and go, with their hostnames/IPs >>>>> changing at random.. I can not figure out how I would run JtR in that >>>>> instance -- since it requires a node list in a file on startup to >>>>> mpirun. >>>>> >>>>> If it matters, I'm looking to do a brute-force using the ASCII mode. Has >>>>> anyone found a way to do a dynamic cluster that adds/removes nodes at >>>>> random? Is this even possible? >>>>> >>>>> >>>> I'm not aware of any existing work that would do this. A solution using >>>> JtR as-is, but with some yet-to-be-implemented master issuing jobs, could >>>> involve looking at the existing "-node=x/y" as describing "pieces" >>>> instead >>>> of "nodes". So instead of saying -node=2/8 as in "you are node 2 of 8" >>>> you'd say -node=4321/100000 as in "do piece 4321 of 100000". Then you'd >>>> submit pieces to active nodes. Obviously you'd have to handle dying nodes >>>> that never reported back their given piece, and re-issue those pieces. >>>> >>>> magnum >>>> >>>> >>>> >>> >> >
Powered by blists - more mailing lists
Confused about mailing lists and their use? Read about mailing lists on Wikipedia and check out these guidelines on proper formatting of your messages.