zlacker

[return to "Mistral 7B Fine-Tune Optimized"]
1. nickth+Cb[view] [source] 2023-12-20 20:55:18
>>tosh+(OP)
Anytime I see a claim that our 7b models are better than gpt-4 I basically stop reading. If you are going to make that claim, give me several easily digestible examples of this taking place.
◧◩
2. thorum+Xl[view] [source] 2023-12-20 21:55:50
>>nickth+Cb
Anecdotally, I finetuned Mistral 7B for a specific (and slightly unusual) natural language processing task just a few days ago. GPT-4 can do the task, but it needs a long complex prompt and only gets it right about 80-90% of the time - the finetuned model performs significantly better with fewer tokens. (In fact it does so well that I suspect I could get good results with an even smaller model.)
◧◩◪
3. skelet+021[view] [source] 2023-12-21 03:44:14
>>thorum+Xl
Can you please point me in the direction of the guide you used for fine tuning? Did you use QLoRA?
[go to top]