• 0 Posts
  • 12 Comments
Joined 1 year ago
cake
Cake day: October 30th, 2023

help-circle



  • The problem with 70B is that it is incrementally better than smaller models, but is still nowhere near competitive with GPT-4, so it is stuck in no man’s land.

    Once we finally get an open source model or architecture that can spar even with GPT-4, let alone 5, there will be much more interest in large models.

    Regarding Falcon Chat 180B, it’s no better in my tests and for my use cases than fine tuned Llama 2 70B, which is a shame. It makes me think that there is something fundamentally wrong with Falcon, besides the laughably small context window.