Not really relevant bc it's kinda nitpicky but amazingly it's not even 4gb, it's more like 2gb if you use the float16 version (which has no quality degredations). Quite amazing that so many images fit in that small a package.
hmm -- I tried converting to float16 just using a naive model.half() call and saw some quality degradation in my images compared to just autocasting parts of the model to float16 while leaving others at float32. Curious if anyone else has had the same experience.
Might be that there's some degredation but I think it's pretty close. Anyway I'm using their 'official' fp16 version which they might be doing some extra magic on idk. I.e. via
My best guess is that I just did something wrong lol. Autocast seems to convert most of the model to fp16 anyways and that works great, so I'll just keep using that!