BangkokPadang@alien.topBtoLocalLLaMA@poweruser.forum•100B, 220B, and 600B models on huggingface!English
1·
11 months agoHonestly, a 4bit quantized version of the 220B model should run on a 192GB M2 Studio, assuming these models could even work with a current transformer/loader.
If people started doing this with any regularity, nVidia would intentionally bork the drivers.