Caffeine_Monster@alien.topBtoMachine Learning@academy.garden•[D] The Status of Open Source Code LLMsEnglish
1·
1 year agoI haven’t got round to trying the xwin coder models, but the precursor 70b chat model was extremely impressive when compared against both chat GPT 3.5 and 4.
You could build an infiniband cluster. The 3090 would give you most bang for buck. Though it’s a lot more work than trading out for A100s, and the extra hardware will cost. You can get 9 GPUs on an single epyc server mobo and still have good bandwidth. So we are talking about manually sourcing and building 10 boxes.
But unless you are training stuff and have cheap electricity a cluster probably doesn’t make sense. No idea why you would need ~1800GB vram.