Communick News
  • Communities
  • Create Post
  • heart
    Support Lemmy
  • search
    Search
  • Login
  • Sign Up
mcmoose1900@alien.topB to LocalLLaMA@poweruser.forumEnglish · 2 years ago

Nouse-Capybara-34B 200K

huggingface.co

external-link
message-square
12
link
fedilink
1
external-link

Nouse-Capybara-34B 200K

huggingface.co

mcmoose1900@alien.topB to LocalLLaMA@poweruser.forumEnglish · 2 years ago
message-square
12
link
fedilink
NousResearch/Nous-Capybara-34B · Hugging Face
huggingface.co
external-link
We’re on a journey to advance and democratize artificial intelligence through open source and open science.
alert-triangle
You must log in or register to comment.
  • Combinatorilliance@alien.topB
    link
    fedilink
    English
    arrow-up
    1
    ·
    2 years ago

    I believe these are TheBloke’s GGUF quants if anyone’s interested: https://huggingface.co/TheBloke/Nous-Capybara-34B-GGUF

    • WolframRavenwolf@alien.topB
      link
      fedilink
      English
      arrow-up
      1
      ·
      2 years ago

      Also note this important issue that affects this and all other Yi-based models:

      BOS token as 1 seriously hurts these GGUF Yi models

      • a_beautiful_rhind@alien.topB
        link
        fedilink
        English
        arrow-up
        1
        ·
        2 years ago

        So we can just skip BOS token on all these models?

        • ambient_temp_xeno@alien.topB
          link
          fedilink
          English
          arrow-up
          1
          ·
          2 years ago

          I did the gguf-py/scripts/gguf-set-metadata.py some-yi-model.gguf tokenizer.ggml.bos_token_id 144

          and it’s changed the outputs a lot from yesterday.

  • metalman123@alien.topB
    link
    fedilink
    English
    arrow-up
    1
    ·
    2 years ago

    Can’t wait to see the benchmarks on these things.

  • toothpastespiders@alien.topB
    link
    fedilink
    English
    arrow-up
    1
    ·
    2 years ago

    Dang, after that 34b drought it’s like suddenly stumbling onto the great lakes right now.

  • vasileer@alien.topB
    link
    fedilink
    arrow-up
    1
    ·
    2 years ago

    200K context!!

    • mcmoose1900@alien.topOPB
      link
      fedilink
      English
      arrow-up
      1
      ·
      2 years ago

      Precisely 47K fits in 24GB at 4bpw.

      I have not tried 3.5, but I think it could be much more.

  • mcmoose1900@alien.topOPB
    link
    fedilink
    English
    arrow-up
    1
    ·
    2 years ago

    Based on the 200K Context Yi 34B.

    https://i.redd.it/t914lpuyb80c1.gif

    • AdOne8437@alien.topB
      link
      fedilink
      English
      arrow-up
      1
      ·
      2 years ago

      if it is based in yi, should it not have the yi-licence instead of mit?

      • mcmoose1900@alien.topOPB
        link
        fedilink
        English
        arrow-up
        1
        ·
        2 years ago

        Yes.

        But its ML land! Everyone violates licenses anyway :P

        • candre23@alien.topB
          link
          fedilink
          arrow-up
          1
          ·
          2 years ago

          Accurate.

LocalLLaMA@poweruser.forum

localllama@poweruser.forum

Subscribe from Remote Instance

Create a post
You are not logged in. However you can subscribe from another Fediverse account, for example Lemmy or Mastodon. To do this, paste the following into the search field of your instance: !localllama@poweruser.forum

Community to discuss about Llama, the family of large language models created by Meta AI.

Visibility: Public
globe

This community can be federated to other instances and be posted/commented in by their users.

  • 4 users / day
  • 1 user / week
  • 4 users / month
  • 4 users / 6 months
  • 1 local subscriber
  • 4 subscribers
  • 1.02K Posts
  • 5.81K Comments
  • Modlog
  • mods:
  • communick@poweruser.forum
  • BE: 0.19.11
  • Modlog
  • Instances
  • Docs
  • Code
  • join-lemmy.org