zlacker

[parent] [thread] 3 comments
1. tarrud+(OP)[view] [source] 2024-10-16 18:42:54
Isn't 3b the kind of size you'd expect to be able to run on the edge? What is the point of using 3b via API when you can use larger and more capable models?
replies(1): >>little+Aa
2. little+Aa[view] [source] 2024-10-16 19:38:39
>>tarrud+(OP)
GP misunderstood: 3b will be available for running on edge devices, but you must sign a deal with Mistral to get access to the weights to run.

I don't think that can work without a significant lobbying push towards models running on the edge but who knows (especially since they have a former French Minister in the founding team).

replies(1): >>ed+Od
◧◩
3. ed+Od[view] [source] [discussion] 2024-10-16 19:59:04
>>little+Aa
> GP misunderstood

I don’t think it’s fair to claim the weights are available if you need to hammer out a custom agreement with mistral’s sales team first.

If they had a self-serve process, or some sort of shink-wrapped deal up to say 500k users, that would be great. But bespoke contracts are rarely cheap or easy to get. This comes from my experience building a bunch of custom infra for Flux1-dev, only to find I wasn’t big enough for a custom agreement, because, duh, the service doesn’t exist yet. Mistral is not BFL, but sales teams don’t like speculating on usage numbers for a product that hasn’t been released yet. Which is a bummer considering most innovation happens at a small scale initially.

replies(1): >>little+Qa1
◧◩◪
4. little+Qa1[view] [source] [discussion] 2024-10-17 06:05:03
>>ed+Od
I'm not defending Mistral here, I don't think it's a good idea I just wanted to to not out that there is no paradox as if the 3b model was API-only.
[go to top]