Hi everyone, we’ve been working on benchmarking different open-source LLMs. We measure, in particular, on the performance of these models once finetued (via QLoRA) on classic NLP downstream tasks like summarization and classification. We also put particular emphasis on benchmarking inference time/cost for these models once deployed.
We’ve just ran our study on the new Zephyr-7B-beta model, a DPO-tuned version of Mistral-7B.
We first tested out-of-the-box performance of Zephyr for summarization under zero-shot and few-shot (for classification, we couldn’t do few-shot because of context length, and we haven’t tried zero-shot since most other open source models gave subpar results).
Then we tested the performance after QLoRA fine-tuning and saw substantial performance boost (as it should). Afterwards we experimented levers we can pull to increase model performance (using NEFTune and/or finetuning on all modules as opposed to attention modules only).
Summarization
Dataset Used: Newsgroup
Rank: 64
Method | Zephyr-7B-β Zero-Shot | Zephyr-7B-β Few-Shot | Fine-Tuning + QLoRA | Fine-Tuning + QLoRA + NEFTune | Fine-Tuning + QLoRA + Full Module Tuning | Fine-Tuning + QLoRA + NEFTune + Full Module Tuning |
---|---|---|---|---|---|---|
ROUGE-1 (in %) | 33.93 | 35.99 | 52.84 | 52.97 | 53.50 | 53.05 |
ROUGE-2 (in %) | 11.21 | 12.97 | 27.75 | 28.44 | 29.66 | 29.23 |
- We see that zero-shot and few-shot performance is already pretty good out-of-the box
- QLoRA was able to refine the syntactic style and pithiness of outputs to match that of the training set
- NEFTune did not improve summarization performance noticeably
- Tuning all modules (as opposed to attention modules) yielded slightly better results
Classification
Dataset Used: Samsum
Rank: 8
Training samples (fraction) | Zephyr-7B-β | Zephyr-7B-β w/ NEFTune | Zephyr-7B-β w/ Full Module Tuning | Zephyr-7B-β w/ NEFTune + Full Module Tuning |
---|---|---|---|---|
266 (2.5%) | 46.05 | 49.61 | 65.36 | 67.23 |
533 (5%) | 55.66 | 60.33 | 72.26 | 72.94 |
1066 (10%) | 66.48 | 64.65 | 73.29 | 72.82 |
2666 (25%) | 66.73 | 68.04 | 74.27 | 75.85 |
5332 (50%) | 69.54 | 72.10 | 74.83 | 74.40 |
10664 (100%) | 74.90 | 72.93 | 77.76 | 77.86 |
- NEFTune boosted performance in low-data regimes
- Tuning all modules has achived ~10x sample efficiency and better better performance at 100% traning fraction
are the finetuned models published somewhere?