• bot-333@alien.topB
    link
    fedilink
    English
    arrow-up
    1
    ·
    1 year ago

    “New RLAIF Finetuned 7b Model” Interesting. “beats Openchat 3.5” Nice! “and comes close to GPT-4” Bruh.

    • Evening_Ad6637@alien.topB
      link
      fedilink
      English
      arrow-up
      1
      ·
      1 year ago

      heheh i can’t read that any more… i really have become very prejudiced when comes to that… to be honest, when it comes to any comparison with GPT-4.

      People have really to understand that even GPT-4 has been aligned, lobotomized and it has been massively downgraded in terms of its perfomance – due to security reasons (what is understandable for me), but anyway this thing still is an absolute beast. if we consider all the restrictions GPT-4 has to undergo, all the smartness at openAI, all the ressources at microsoft and so on, we have to realize that currently nothing is really comparable to GPT-4. Especially not 7B models.

      • noeda@alien.topB
        link
        fedilink
        English
        arrow-up
        1
        ·
        1 year ago

        I’ve seen the “… beats GPT-4” enough times that now whenever I see a title that suggests a tiny model can compete with GPT-4 I see it as a negative signal; that the authors are bullshitting through some benchmarks or some other shenanigans.

        It’s annoying because the models might be legitimately good models for being open and within their weight class but now you’ve put my brain in BS detecting mode and I can’t trust you’ve done good faith measurement anymore.

        • Evening_Ad6637@alien.topB
          link
          fedilink
          English
          arrow-up
          1
          ·
          1 year ago

          Yeah I dont think authors are intentionally bullshitting or intentionally doing “benchmark cosmetics”, but maybe it’s more lack of knowledge on whats going on in terms of (most of) benchmarks and their the image that has become ruined in the meantime.

          • Competitive_Ad_5515@alien.topB
            link
            fedilink
            English
            arrow-up
            1
            ·
            1 year ago

            Sure, but name-dropping the biggest name in the game and comparing yourself favourably to it is a big swing. It’s either a naive at best marketing claim or it’s untrue.

        • bot-333@alien.topB
          link
          fedilink
          English
          arrow-up
          1
          ·
          1 year ago

          There are SO many models “bullshitting through some benchmarks or some other shenanigans” that I’m cooking my own benchmark system LOL.

    • PrometheusZer0@alien.topB
      link
      fedilink
      English
      arrow-up
      1
      ·
      1 year ago

      Here’s what I’m using:

      FROM starling-lm-7b-alpha.Q5_K_M.gguf

      PARAMETER stop <|end_of_turn|>

      PARAMETER stop <|im_sep|>

      TEMPLATE """

      GPT4 User: {{.Prompt}}<|end_of_turn|>GPT4 Assistant:

      """

  • noeda@alien.topB
    link
    fedilink
    English
    arrow-up
    1
    ·
    1 year ago

    The first image posted; looks like it’s not even close to GPT-4?

    • Real-Elk-6109@alien.topB
      link
      fedilink
      English
      arrow-up
      1
      ·
      1 year ago

      Considering “close” as a relative word, it came closer than other open-source models. But you have a point too.

  • alexthai7@alien.topB
    link
    fedilink
    English
    arrow-up
    1
    ·
    1 year ago

    Do someone know why it writes the line feed code all the time in its answer ? <0 x 0 A>

    Besides this, I find the model amazing.

  • thereisonlythedance@alien.topB
    link
    fedilink
    English
    arrow-up
    1
    ·
    1 year ago

    I was sceptical, but darn it’s good. Mistral is a fantastic base and with this technique these guys have pushed it another step closer. A lot of the answers I’m getting are on on par with old GPT-4 (pre-turbo, turbo in the API is a step up on old GPT-4 IMO).

    • Sweet_Protection_163@alien.topB
      link
      fedilink
      English
      arrow-up
      1
      ·
      1 year ago

      I can’t wait for the trustworthy closed sourced benchmarks. Can’t believe I’m saying that… but it’s honestly what we need.

      • liqui_date_me@alien.topB
        link
        fedilink
        English
        arrow-up
        1
        ·
        1 year ago

        Wonder if that’s a good startup idea? Something that can benchmark language models and charges a fee for doing so

  • pseudonerv@alien.topB
    link
    fedilink
    English
    arrow-up
    1
    ·
    1 year ago

    Form huggingface model card,

    Starling-RM-7B-alpha is a reward model trained from Llama2-7B-Chat.

    From their webpage, https://starling.cs.berkeley.edu

    Our reward model is fine-tuned from Llama2-7B-Chat

    Yet, the model config.json

    "max_position_embeddings": 8192,
    "model_type": "mistral",
    "num_attention_heads": 32,
    "num_hidden_layers": 32,
    "num_key_value_heads": 8,
    "rms_norm_eps": 1e-05,
    "rope_theta": 10000.0,
    "sliding_window": 4096,
    

    SO? Whoever is doing the PR has no f***ing idea what their student labors are actually doing.

    • Warm_Shelter1866@alien.topB
      link
      fedilink
      English
      arrow-up
      1
      ·
      1 year ago

      What does it mean that an LLM is a reward model ? , I always thought of rewards only in the RL field . And how would the reward model be used during finetuning?

  • georgejrjrjr@alien.topB
    link
    fedilink
    English
    arrow-up
    1
    ·
    1 year ago

    If there is something somehow inherently superior about having a separate reward model, that should be teased out.

    It would be nice to see stronger baselines / ablations for this reason. I realize it’s nigh impossible to keep up with the unrelenting pace of advances, so I don’t fault the authors here. That said, if there isn’t a compelling reason to keep the separate preference model, community people-hours will probably be best spent sticking with DPO/IPO to avoid the hyper-parameter tuning rabbit hole.

    My guess: the way things are going, we’ll soon see a rough consensus emerge around a sane default DPO or Identity-PO recipe for fine-tunes (the same way we’ve seen gradual convergence around decoder-only transformer + rotational positional embeddings + group query attention + FlashAttention 2) to be applied absent a compelling reason to use a different reward signal.

    No matter what, preference datasets like this are helpful. Pity about the license being claimed here, it’s hard to imagine it would hold up, but the specter is a bit of a hindrance.

  • Thistleknot@alien.topB
    link
    fedilink
    English
    arrow-up
    1
    ·
    1 year ago

    rm is the reward model… not the same as the lm model. I tried the lm, wasn’t impressed. Gpt-3.5 did better for summarizing quotes. It was good, but I honestly think open hermes and or synthia 1.3b do better

  • OC2608@alien.topB
    link
    fedilink
    English
    arrow-up
    1
    ·
    1 year ago

    How to earn VC money 101: “Beats GPT-4!”

    And voila! you’re rich now.