@cancel In the few times I did it in the past I used wget --mirror with a few tweaked parameters for directory traversal and domain-spanning.
@blindcoder It only seems to download .html, images, css, etc.
@cancel @blindcoder are the binary files hosted on the same domain as the html+images+css? I think wget needs explicit options to allow getting from multiple domains in recursive mode, and it might also need options to limit recursion depth in that case to avoid downloading the whole internet...
@cancel @mathr I think wget also respects rel=nofollow but I don't know how to turn that off...