If you don’t know what those are, refer to these two reddit posts about Marx 3B V3 and Akins 3B, the unquantized model weights are available at HuggingFace. Link to Marx 3B V3 and Akins 3B.

As the StableLM support for llama.cpp has just been recently, u/The-Bloke(Thank you so much!) quantized my StableLM models to GGUF as a lot of people are wanting to try the model in GGUF, you can find the GGUF conversion for Marx 3B V3 and Akins 3B. Again, credit to u/The-Bloke for quantizing the model, thank you!

By the way, I don’t know what dataset to finetune on right now. If you know a good dataset, let me know and I will look into it. Though I could probably only finetune on datasets below 5k conversations, maybe 10k.

  • ttkciar@alien.topB
    link
    fedilink
    English
    arrow-up
    1
    ·
    1 year ago

    Yaay! :-) just in time for the weekend! I’ll give them a whirl :-)

    Thanks for the heads-up!

    As for datasets, I’ve been thinking that HelixNet might be instrumental in generating high-quality synthetic datasets (as were used to train Microsoft’s phi), but I haven’t had a chance to mess with that idea yet. Sorry I don’t have anything concrete to suggest.