u/ehartford
It’s source available, not open source. They don’t even allow commercial usage.
Important: researcher only, non commercial license.
I think Karpatgy’s nanogpt series is a great alternative to this.
https://youtube.com/playlist?list=PLAqhIrjkxbuWI23v9cThsA9GvCAUhRvKZ&feature=shared
OpenAI hires ml experts AND software engineers. For the first few years at OpenAI, he was much more of the second as his knowledge of ml was very minimal.
Are your finetunes full training or lora/qlora?
Do they plan on offering Mi200 or MI300 to the public?
Can’t wait to try out a GPU with more than 80gb vram
How about T4 GPU or something like 3090 from runpod? the 3090 costs around 0.5$ per hour which is around 350 dollars per month and it gives you 24 GB which should be enough for t4
Is Goliath that good? Is it that better than all of the Llama2-70B tunes that’s worth the hardware investments needed for running it?
I wouldn’t call it “Arabic”. It doesn’t handle Arabic text well.
Oh Today I discovered there’s a programming language called zig.
Can I dm you for career related questions?
I really like this pattern!
Someone starts a cool ml project in python. Another person re-creates it more efficiently in c++.
Then someone else rebuilds it in rust.
Well, two areas where Llama can improve are:
Yes, every ML practitioner knows him but probably the prompt engineer aren’t aware of his massive contributions to the field.
In no specific order:
Zephyr B OpenHermes2.5 OpenChat3.5