zlacker

[return to "Qwen3-Coder-Next"]
1. alexel+t9[view] [source] 2026-02-03 16:39:12
>>daniel+(OP)
Is this going to need 1x or 2x of those RTX PRO 6000s to allow for a decent KV for an active context length of 64-100k?

It's one thing running the model without any context, but coding agents build it up close to the max and that slows down generation massively in my experience.

◧◩
2. segmon+sr[view] [source] 2026-02-03 17:51:52
>>alexel+t9
1 6000 should be fine, Q6_K_XL gguf will be almost on par with the raw weights and should let you have 128k-256k context.
[go to top]