- 4 Posts
- 15 Comments
DreamGenX@alien.topBto LocalLLaMA@poweruser.forum•Is there a 7B model capable to extract keywords from a text and return them as an array?English1·2 years agoOn top of what other said, make sure to include a few shot examples in your prompt, and consider using constrained decoding (ensuring you get valid json of whatever schema you provide, see pointers on how to do it with llama.cpp).
For few shotting chat models, append fake previous turns, like:
System: User: Assistant: ... User: Assistant: User:
DreamGenX@alien.topBto LocalLLaMA@poweruser.forum•Training on the rephrased test set is all you need: 13B models can reach GPT-4 performance in benchmarks with no contamination detectable by traditional methodsEnglish1·2 years agoIt’s inevitable people will game the system when it’s so easy, and the payoff can be huge. Not so long ago people could still get huge VC checks for showing off GitHub stars or benchmark numbers.
DreamGenX@alien.topBto LocalLLaMA@poweruser.forum•A fun day evaluating LLM Chat GUIs/Servers in Docker. Here's what I learned...English1·2 years agoCurious to hear what other UIs people use and for what purpose / what they like about each (like Oogabooga, or Kobold).
DreamGenX@alien.topBto LocalLLaMA@poweruser.forum•What’s recommended hosting for open source LLMs?English1·2 years agoI can recommend vLLM. Also offers OpenAI compatible API service, if you want that.
DreamGenX@alien.topOPBto LocalLLaMA@poweruser.forum•DreamGen Opus — Uncensored model for story telling and chat / RPEnglish1·2 years agoThank you so much for the kind feedback! If you have found some cool prompts, come share them with others on our discord.
DreamGenX@alien.topOPBto LocalLLaMA@poweruser.forum•DreamGen Opus — Uncensored model for story telling and chat / RPEnglish1·2 years agoI hope it will be something tasty! :)
DreamGenX@alien.topOPBto LocalLLaMA@poweruser.forum•DreamGen Opus — Uncensored model for story telling and chat / RPEnglish1·2 years agoThe training data had example of up to 4096 tokens. The model should also work beyond that, but I did not do a deep analysis of degradation.
DreamGenX@alien.topOPBto LocalLLaMA@poweruser.forum•DreamGen Opus — Uncensored model for story telling and chat / RPEnglish1·2 years agoI agree, I hope I can make things cheaper with better utilization. You have to consider that a single GPU is not used 100% the time, so there’s a lot of waste. And due to lack of scale, I also do not get any special pricing on the GPUs. The more users, the closer the utilization will be to 100%, and the better GPU pricing. (For instance, I heard that on Google Cloud, enterprise customers can negotiate the on-demand GPU price down to the regular spot price for some of the GPUs)
DreamGenX@alien.topOPBto LocalLLaMA@poweruser.forum•DreamGen Opus — Uncensored model for story telling and chat / RPEnglish1·2 years agoWow, amazing, thanks for giving it a try GGUF and other quants are coming, so your computer should have an easier time soon! :)
What’s the maximum possible dead babies score? :D
DreamGenX@alien.topOPBto LocalLLaMA@poweruser.forum•DreamGen Opus — Uncensored model for story telling and chat / RP1·2 years agoThank you!
DreamGenX@alien.topOPBto LocalLLaMA@poweruser.forum•DreamGen Opus — Uncensored model for story telling and chat / RPEnglish1·2 years agoGreat news, the great /u/TheBloke is working on this!
DreamGenX@alien.topBto LocalLLaMA@poweruser.forum•OpenAI API update (mostly for coders)English1·2 years agoI have been using the Python API client 1.0 preview version (which was just released) for some time with vLLM OpenAI compatible server and it worked well – at least I did not notice any issues.
DreamGenX@alien.topOPBto LocalLLaMA@poweruser.forum•DreamGen Opus — Uncensored model for story telling and chat / RPEnglish1·2 years agoThere was a bug on the website where the first time the “Continue” would not work if you did not refresh, should work now even though the editor is quite janky still, sorry for that :(
(can’t wait for AI to take over React from me :P)
Found a live stream on YouTube, for anyone interested: https://www.youtube.com/watch?v=o35EY8I9PXU