zlacker

[parent] [thread] 5 comments
1. thorum+(OP)[view] [source] 2023-12-20 21:55:50
Anecdotally, I finetuned Mistral 7B for a specific (and slightly unusual) natural language processing task just a few days ago. GPT-4 can do the task, but it needs a long complex prompt and only gets it right about 80-90% of the time - the finetuned model performs significantly better with fewer tokens. (In fact it does so well that I suspect I could get good results with an even smaller model.)
replies(2): >>oceanp+01 >>skelet+3G
2. oceanp+01[view] [source] 2023-12-20 22:01:24
>>thorum+(OP)
I have a fine tuned version of Mistral doing a really simple task and spitting out some JSON. I'm getting equivalent performance to GPT-4 on that specialized task. It's lower latency, it's outputting more tokens/sec., more reliable, private, and completely free.

I don't think we will have an Open Source GPT4 for a long time so this is sorta clickbait, but for the small, specialized tasks, tuned on high quality data, we are already in the "Linux" era of OSS models. They can do real, practical work.

replies(3): >>YetAno+3h >>tomrod+mi >>cced+EO
◧◩
3. YetAno+3h[view] [source] [discussion] 2023-12-20 23:50:50
>>oceanp+01
> completely free

Not according to my calculation. For low request rate it is likely more expensive than GPT4.

◧◩
4. tomrod+mi[view] [source] [discussion] 2023-12-20 23:59:13
>>oceanp+01
Been my thought for awhile now.

Can you recommend where I can learn more about hardware requirements for running Mistral/Mixtral?

5. skelet+3G[view] [source] 2023-12-21 03:44:14
>>thorum+(OP)
Can you please point me in the direction of the guide you used for fine tuning? Did you use QLoRA?
◧◩
6. cced+EO[view] [source] [discussion] 2023-12-21 05:31:22
>>oceanp+01
How are you guys fine tuning?
[go to top]