ninjasaid13@alien.topB to LocalLLaMA@poweruser.forumEnglish · 1 year agoMultiLoRA: Democratizing LoRA for Better Multi-Task Learningarxiv.orgexternal-linkmessage-square1fedilinkarrow-up11arrow-down10
arrow-up11arrow-down1external-linkMultiLoRA: Democratizing LoRA for Better Multi-Task Learningarxiv.orgninjasaid13@alien.topB to LocalLLaMA@poweruser.forumEnglish · 1 year agomessage-square1fedilink
minus-squareFullOf_Bad_Ideas@alien.topBlinkfedilinkEnglisharrow-up1·1 year ago In our benchmark, training LLaMA-7B with sequences of 1024 tokens with n = 5 would use more VRAM than full parameter fine-tuning This is a deal breaker. I am hopeful for LoftQ integration into training frameworks, it has more potential. https://arxiv.org/abs/2310.08659
This is a deal breaker.
I am hopeful for LoftQ integration into training frameworks, it has more potential. https://arxiv.org/abs/2310.08659