>>106621985
>>106621964
>>106621592
>>106621054
>>106621637
>>106621789
Need to head to bed now but here's the FP16 GGUF if anyone wants to try it.
gofile.io/d/rri4pw
I highly recommend using the Gemma2 prompt template (hopefully whatever front end you using are we supports that. If you're using a CLI engine, US specially need to make sure you use that properly)
<start_of_turn>user
{your_system_prompt}
{user_message_1}<end_of_turn>
<start_of_turn>model
{model_response_1}<end_of_turn>
<start_of_turn>user
{user_message_2}<end_of_turn>
<start_of_turn>model
I recommend setting a limit on how many tokens it can generate (128 seems to be the sweet spot) and are reasonable repeat_last_n setting (or whatever your inference engine's equivalent to that is)
Gn, sweet dreams /wait/ :)
ps here's the dataset I used: https://gofile.io/d/qcdvPV