Too big

#1
by ShunAonuma - opened

Why it's a 9 GB model after 4-bit quantization?
The 4-bit GGUF versions (like Q4_K_M) are about 5 GB.

Sign up or log in to comment