• 0 Posts
  • 7 Comments
Joined 1 year ago
cake
Cake day: November 9th, 2023

help-circle



  • NEW FEATURE: Context Shifting (A.K.A. EvenSmarterContext) - This feature utilizes KV cache shifting to automatically remove old tokens from context and add new ones without requiring any reprocessing. So long as you use no memory/fixed memory and don’t use world info, you should be able to avoid almost all reprocessing between consecutive generations even at max context. This does not consume any additional context space, making it superior to SmartContext.

    do you tried this koblodcpp new feature?


  • Newbie question, but is there a way to have 4*A100 40G cards run as one, with 160G VRAM in total?

    Yes, that will work but lose some performance

    edit) If this is possible, can I run 8*3090 24G cards as one also?

    Yes and no, yes you can do this but no unless you need run a 176GB model, more gpu with a model only lose performance but not increase it

    for example, if I run a 13b gptq 4bit model on 1*3090 it get 45t/s, if i run it on 2*3090 it will slow down to 30t/s but 3*3090 is same

    and also, I dont have any idea why people always saying about nvlink, 2*pciex16 3090 give same speed compare with 2*pciex1, not sure if there will change for higher number but it just not help at all for 2*3090

    so basically the real matter is what model you want to run, not what and how many GPU you can use, you need run your model with minimum number gpu to get best performance, if 3090 have 36Gvram but not only 24Gvram, 36G*2 will way more faster than 24G * 3 even the total Vram is same