Damus
redshift profile picture
redshift
@redshift
Kimi K2.6 seems lobotomized. Everyone is using 4 bit quantization and offering at prices comparable to 8 bit quantization. Users aren't sophisticated enough to choose the right providers yet.

Routstrd chose @Private Provider Routstr AI as it has the best price, they proxy to @PayPerQ, which in-turn proxies to OpenRouter (correct me if i'm wrong) and OpenRouter just sorts by price/some custom algorithm.

And we don't even know if the providers are lying or not.
Kimi K2.6 is a good model for medium level tasks. It shouldn't be this bad. (Btw GPT 5.5 is great).

Any ideas on how we can handle this better?



5โค๏ธ1๐Ÿ‘€1๐Ÿ’œ1
PayPerQ · 2w
Since we proxy to openrouter for this particular model, you can use any parameters that OR accepts and force 8 bit quantization if you like. We could be better about advertising this feature though. Iโ€™m not sure if the standard Flow is 4 or 8 bit though. It may still be 8.