Gemini and Codex already scored higher on benchmarks than Opus 4.6 and they recently added a $100 tier with limited 2x limits, that's their answer and it seems people have caught on.
> that's their answer and it seems people have caught on.
There's nothing to catch on to. OpenAI have been shouting "come to us!! We are 10x cheaper than Anthropic, you can use any harness" and people don't come in droves. Because the product is noticeably worse.
> and people don't come in droves. Because the product is noticeably worse.
As of Oct 2025, it appears that openai markets share is 15x that of anthropic: 60% vs 3.5% [1].
As of April 2026, openai has 900 million weekly users [2] while anthropic has 300 million monthly users [1].
As of March 2026, openai app downloads were 2.2 million per day, while anthropic app downloads were 340,000. openai mobile users were 248 million per day, while anthropic mobile users were 9.4 million. In Feb 2026, chatgpt had 5.4 billion web visits, while claude had 290 million web visits. [3]
It seems to me that openai operates at a much higher scale than anthropic. Since you used droves as a proxy for product quality, by that standard anthropic has a much more inferior product. :)
What happens if you flood the market with a bunch of implausible bets like "sun won't rise tomorrow"? Sure, you might try to filter that out with some sort of "seasoning" period (ie. don't buy new markets), but then that means more time for arbitrageurs to correctly price the market, depriving you of any price advantage you might have had.
This locks up your money in the meantime, right? If so, considering the fed funds rate is 3.64% (and you can probably get higher rates on stablecoins), a huge chunk of those "winnings" is going to be eaten up by the opportunity cost of the money.
You forget that Polymarket is just a casino, and the house always wins.
For example, recent events show that any bet can be selectively disputed by arbitrary reason ("we found insiders", "we found this immoral/illegal", etc.).
Except that the mere existence of the market with the question posed for people to consider, probably activates the availability heuristic[1], causing people to overestimate the likelihood.
I don't want a nudge. I want a clear RED WARNING with "You've gone away from your computer a bit too long and chatted too much at the coffee machine. You're better off starting a new context!"
I think after the TTL expires the session should be autocompacted and the user should given a choice to continue with compacted version or be hit with the full read cost of continuing with their large but expired context. At the moment users are blind what is going on.
Why is nobody even asking why that should be an issue? No other text editor shits the bed that way. The whole point of the computer is that it patiently waits for my input.
Some claim that some of the recent smaller local models are as good as Sonnet 4.5 of last year and the bigger high-end models can be as almost as good as Claude, Gemini and Codex today, but some say they're benchmaxed and not representative.
To try things out you can use llama.cpp with Vulkan or even CPU and a small model like Gemma 4 26B-A4B or Gemma 4 31B or Qwen 3.5 35-A3B or Qwen3.5 27B. Some of the smaller quants fit within 16GB of GPU memory. The default people usually go with now is Q4_K_XL, a 4-bit quant for decent performance and size.
Get a second hand 3090/4090 or buy a new Intel Arc Pro B70. Use MoE models and offload to RAM for best bang for your buck. For speed try to find a model that fits entirely within VRAM. If you want to use multiple GPUs you might want to switch to vLLM or something else.
reply