zlacker

[parent] [thread] 7 comments
1. Whitne+(OP)[view] [source] 2023-12-27 18:44:11
llama.cop Performs better than what?

3.5 Turbo, one of the 4.0 models, API or App?

replies(1): >>buggle+N
2. buggle+N[view] [source] 2023-12-27 18:47:32
>>Whitne+(OP)
JSON mode and function-calling with a JSON schema in the OpenAI API.
replies(1): >>Whitne+V9
◧◩
3. Whitne+V9[view] [source] [discussion] 2023-12-27 19:39:06
>>buggle+N
Right, but which model?

It makes a huge difference.

replies(1): >>buggle+Ha
◧◩◪
4. buggle+Ha[view] [source] [discussion] 2023-12-27 19:42:43
>>Whitne+V9
I’ve been using OpenChat 3.5 1210 most recently. Before that, Mistral-OpenOrca. Both return JSON more consistently than gpt-3.5-turbo.
replies(1): >>airstr+Tc
◧◩◪◨
5. airstr+Tc[view] [source] [discussion] 2023-12-27 19:54:52
>>buggle+Ha
gpt-3.5-turbo is not the benchmark
replies(1): >>buggle+lo
◧◩◪◨⬒
6. buggle+lo[view] [source] [discussion] 2023-12-27 20:49:24
>>airstr+Tc
I don’t know what point you’re trying to make. They also return JSON more consistently than gpt-4, but I don’t use that because it’s overkill and expensive for my text extraction tasks.
replies(1): >>Whitne+5y
◧◩◪◨⬒⬓
7. Whitne+5y[view] [source] [discussion] 2023-12-27 21:48:41
>>buggle+lo
Because people have different interests and want to hear your results for different reasons.

Some want to consider results relative to cost, and some are interested only in how it compares to SOTA.

replies(1): >>buggle+xJ
◧◩◪◨⬒⬓⬔
8. buggle+xJ[view] [source] [discussion] 2023-12-27 22:56:32
>>Whitne+5y
I mean, sure, but the parent should also just explicitly state what it is they were asking or claiming. I’ve answered every question asked. Making vague declarations about something not being “the benchmark,” while not stating what you think “the benchmark” should be, is unhelpful.
[go to top]