Hacker News new | past | comments | ask | show | jobs | submit login

I'm going to have to heavily disagree. Gemini 2.5 Pro has super impressive performance on large context problems. I routinely drive it up to 4-500k tokens in my coding agent. It's the only model where that much context produces even remotely useful results.

I think it also crushes most of the benchmarks for long context performance. I believe on MRCR (multi round coreference resolution) it beats pretty much any other model's performance at 128k at 1M tokens (o3 may have changed this).






I find that it consistently breaks around that exact range you specified. In the sense that reliability falls off a cliff, even though I've used it successfully close to the 1M token limit.

At 500k+ I will define a task and it will suddenly panic and go back to a previous task that we just fully completed.


Totally agreed on this. The context size is what made me switch to Gemini. Compared to Gemini, Claude's context window length is a joke.

Particularly for indie projects, you can essentially dump the entire code into it and with pro reasoning model, it's all handled pretty well.


Yet somehow chatting with Gemini in the web interface, it forgets everything after 3 messages, while GPT (almost) always feels natural in long back-and-forths. It’s been like this for at least a year.

My experience has been different. I worked with it to disgnose two different problems. On the last one I counted questions and answers. It was 15.

OOI what coding agent are you managing to get to work nicely with G2.5 Pro?

I mostly use Roo Code inside visual studio. The modes are awesome for managing context length within a discrete unit of work.

Is that a codebase you're giving it?



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: