• Shoddy_Vegetable_115@alien.topB
      link
      fedilink
      English
      arrow-up
      1
      ·
      1 year ago

      Exactly. It didn’t hallucinate even once in my tests. I used RAG and it gave me perfect to-the-point answers. But I know most people want more verbose outputs it’s just that it’s good for factual retrieval use cases.

      • julylu@alien.topB
        link
        fedilink
        English
        arrow-up
        1
        ·
        1 year ago

        Maybe for RAG, short answer is less possible for hallucination?I will test more. thanks

      • Intel@alien.topB
        link
        fedilink
        English
        arrow-up
        1
        ·
        1 year ago

        This is a fine-tuned/instruction-tuned model. Explicit system prompts or instructions like “generate a long, detailed answer” can make the model generate longer responses. 🙂

        –Kaokao, AI SW Engineer @ Intel