I think there is a pattern it will always be nerfed the few weeks before launching a new model. Probably because they are throwing a bunch of compute at the new model.
Yeah maybe that but atleast let us know about this Or have dynamic limits? Nerfing breaks trust.
Though I am not sure if they actually nerf it intentionally. Haven't heard from any credible source. I did experience in my workflow though.
I don't use AI studio for my work. I used Antigravity/Gemini CLI and 3 pro was great for few weeks and now it's worse than 3 flash or any smaller model from competitor which are rated lower on benchmarks
IME, they definitely nerf models. gemini-2.5-pro-exp-03-25 through AI Studio was amazing at release and steadily degraded. The quality started tanking around the time they hid CoT.
Tried Antigravity for 2 queries and my model quota limit breached. Model definitely felt better than GPT 5.1 (my current daily driver). I am continuing to use Gemini 3 Pro on Cursor to evaluate further.
It’s a full fledged tool like DataDog which is opensource and can be selfhosted. You can replace honeycomb with it.
What features do you think it’s missing compared to honeycomb?