• 0 Posts
  • 3 Comments
Joined 1 year ago
cake
Cake day: October 31st, 2023

help-circle
  • It’s not really about fairness though, it’s about knowing where things stand.

    I’ve used GPT 4 a lot so I have a rough idea of what it can do in general, but I’ve almost no experience with local LLMs. That’s something I’ve only played a little with recently after seeing the advances in the past year.

    So, I don’t really see it as a question that disparages local LLMs, so I don’t see fairness as an issue - it’s not a competition to me.



  • word2vec is a library used to create embeddings.

    Embeddings are vector representations (a list of numbers) that represents the meaning of some text.

    RAG is retrieval augmented generation. It is a method to get better answers from GPT by giving it relevant pieces of information along with the question.

    RAG is done by:

    1. taking a long text splitting it into pieces
    2. creating embeddings for each piece and storing those
    3. when someone asks a question, create an embedding for the question
    4. find the most similar embeddings of sections from the long text say the top 3
    5. send those 3 pieces along with the question to GPT for the answer

    A vector DB is a way to store those section embeddings and to also search for the most relevant sections. You do not need to use a vector DB to perform RAG, but it can help. Particularly if you want to store and use RAG with a very large amount of information.