Gemma3NPC-filtered-v2-Q4-GGUF

The quantized GGUF version of Gemma3NPC-filtered-v2-float16

Another one of those test models

we were chatting and just decided: "What would happen if we have a higher learning rate and run 3 epochs🤓"

So here it is, the second generation of filtered Gemma3NPC, with bare minimum effort of typing 6 characters and a little patience(3 hours).

Again, our training notebook is on Github.


Training Parameters

Parameter Gemma3NPC-Filtered v2
Learning rate 2e-5 4e-5
Warmup Steps 150 100
Gradient clipping 0.5 1.0

Training Loss

This time, we accidentally stopped the training when it reach step 200, so when we resumed, the training started from scratch but seems to have used the last checkpoint.

Here is a graph of the training loss, saved after after 5 steps.

chart


Next Steps

Our top priority is now the gathering of more datasets, such as SODA and some real video game data.

We might try to switch to a new model(Qwen?), as the Gemma3n license is a little restrictive.

New methods other than SFT to improve performance, like GRPO.

Downloads last month
32
GGUF
Model size
7B params
Architecture
gemma3n
Hardware compatibility
Log In to view the estimation

4-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for chimbiwide/Gemma3NPC-filtered-v2-Q4-GGUF

Quantized
(2)
this model

Collection including chimbiwide/Gemma3NPC-filtered-v2-Q4-GGUF