Issue with finetuning Gemma 3 with "train_on_responses_only" #2714
Unanswered
MostAardvark224
asked this question in
Q&A
Replies: 1 comment
-
Fix (kind of): I used my same dataset on Qwen3-32B and it worked just fine, so if you're having the same issue try switching models. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Hey all, I'm new to unsloth and was wondering if anyone could help me solve an issue with finetuning Gemma 3.
Here's my code: (for context most of this is from the unsloth colab notebook on finetuning Gemma 3, I just adapted it for my own dataset).
When I print out the row, this is what it looks like:
'<start_of_turn>user\n my prompt xyz <end_of_turn>\n<start_of_turn>model\n{"model completion as JSON object"}<end_of_turn>\n'
which is what I think the Gemma 3 chat template is supposed to look like (it's just missing the token.
I then initialize my SFTTrainer
I've looked all around and can't really find any solutions. I think the issue likely has something to do with my dataset because if I use the "Finetome-100k" dataset that was used in the original notebook it works just fine. I just can't pinpoint where the error is coming from exactly.
Any help would be MUCH appreciated. Please ask further questions if more specifics are required.
Beta Was this translation helpful? Give feedback.
All reactions