I came across this new finetuned model based on Openchat 3.5 which is apparently trained used Reinforcement Learning from AI Feedback (RLAIF).
https://huggingface.co/berkeley-nest/Starling-LM-7B-alpha
Check out this tweet: https://twitter.com/bindureddy/status/1729253715549602071
Form huggingface model card,
From their webpage, https://starling.cs.berkeley.edu
Yet, the model config.json
SO? Whoever is doing the PR has no f***ing idea what their student labors are actually doing.
yeah I was put off by the lack of mention on the base model
What does it mean that an LLM is a reward model ? , I always thought of rewards only in the RL field . And how would the reward model be used during finetuning?