Honestly, I mostly just follow hugging face’s blog and articles. I know there are some latest fancy attention improvements, alternatives for RLHF, GPU whatever optimization, etc, but I’m not going to implement those myself. If it’s not in hugging face’s ecosystem, then I most likely wouldn’t use it in my daily work/production code anyway.
Yeah, reading all their latest releases is already taking me a lot of time so I just mostly stop there. They also don’t have a lot of documentations for their latest stuffs, so it takes a bit time to figure things out. I think their packages will settle down to a more stable state after a year or two, after the NLP hype cooldowns a bit.