https://huggingface.co/deepnight-research
I’m not affiliated with this group at all, I was just randomly looking for any new big merges and found these.
100B model: https://huggingface.co/deepnight-research/saily_100B
220B model: https://huggingface.co/deepnight-research/Saily_220B
600B model: https://huggingface.co/deepnight-research/ai1
They have some big claims about the capabilities of their models, but the two best ones are unavailable to download. Maybe we can help convince them to release them publicly?
the devs mentioned that the 600B model takes about 1,3TB space alone…
Give it 5 years with the Mac Studio. Next year 256gb, will go up real quick.
Honestly, a 4bit quantized version of the 220B model should run on a 192GB M2 Studio, assuming these models could even work with a current transformer/loader.
Make it 0.01bpm quantized and you will fit in good ol’ 3090.