ZeroGPU-LLM-Inference / quantize_to_awq_colab.ipynb
Alikestocode's picture
Lower Gemma AWQ group size to 16
f8c20fd
raw
history blame
32.9 kB
Open in Colab