• 2 Posts
  • 14 Comments
Joined 11 months ago
cake
Cake day: October 31st, 2023

help-circle


  • That is absolutely impressive, but:

    1. is light quantization that bad? Couldn’t you run 99% of the same model for half the cost? Is running unquantized just a flex/exercise/bragging right?
    2. Would quantized run faster? Slower? The same?
    3. Isn’t Falcon-180B kinda… meh? I mean it’s pretty smart from size alone, but the lack of fine tuning by the community means it’s kind of like running LLaMA-70b by itself.
    4. Would one of those new crazy good Threadrippers beat the GPUs? lol



  • The thing is, as far as I’m aware, “sound generation” is always a separate TTS thing cobbled together, and even “vision” is a separate thing that describes the image for the AI.

    This 13b model is probably still state of the art in the vision department for open models, a few crop up now and again but they didn’t surprise me much.
    https://llava-vl.github.io/

    If you need to recognize audio, check Whisper, or Faster-Whisper, or anything developed from that. If you need to generate voice, check Bark, maybe Silero, RVC, etc.

    You probably won’t find it all wrapped into one neat package like ChatGPT+ right now, but I’d love to be proven wrong.








  • Actually GBNF is this re-branding, BNF is the proper name (the G is Georgi Gerganov’s). There’s also a reason why languages compile to assembly but that doesn’t mean it’s user-friendly. Or Abstract Syntax Trees. There’s stuff that pretty much only applies to compilers, doesn’t mean it’s a good general-purpose solution.

    Though I must imagine implementing BNF is orders of magnitude easier than implementing the monster that is extended regular expressions.