BitDistiller: Unleashing the Potential of Sub-4-Bit LLMs via Self-Distillation
Paper
•
2402.10631
•
Published
•
2
| PPL | arc_easy | arc_challenge | piqa | winogrande | hellaswag | mmlu | QA Avg |
|---|---|---|---|---|---|---|---|
| 2794.56 | 26.14 ± 0.90 | 20.56 ± 1.18 | 51.85 ± 1.17 | 49.72 ± 1.41 | 25.77 ± 0.44 | - | 34.81 |
Training method based on BitDistiller Paper
Base model
TinyLlama/TinyLlama_v1.1