• 0 Posts
  • 4 Comments
Joined 11 months ago
cake
Cake day: October 31st, 2023

help-circle
  • Could you please share a citation for the mentioned research papers?

    I’m interested in seeing this as well.

    He probably means that, although scaling might still deliver better loss reduction, this won’t necessarily cash out to better performance “on the ground”.

    Subjectively, GPT4 does feel like a smaller step than GPT3 and GPT2 were. Those had crazy novel abilities that the previous one lacked, like GPT3’s in-context learning. GPT4 displays no new abilities.* Yes, it’s smarter, but everything it does was possible, to some limited degree, with GPT3. Maybe this just reflects test saturation. GPT4 performs so well that there’s nowhere trivial left to go. But returns do seem to be diminishing.

    (*You might think of multimodality, but they had to hack that into GPT4. It didn’t naturally emerge with scale, like, say, math ability.)