Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

> My bet is their model doesn't realistically compare to any of the frontier models.

I've been using composer-1 in Cursor for a few weeks and also switching back and forth between it, Gemini Flash 3, Claude Opus 4.5, Claude Sonnet 4.5 and GPT 5.2.

And you're right it's not comparable. It's about the same quality of code output of the aforementioned models but about 4x as fast. Which enables a qualitatively different workflow for me where instead of me spending a bunch of time waiting on the model, the model is waiting on me to catch up with its outputs. After using composer-1, it feels painful to switch back to other models.

I work in a larg(ish) enterprise codebase. I spend a lot of time asking it questions about the codebase and then making small incremental changes. So it works very well for my particular workflow.

Other people use CLI and remote agents and that sort of thing and that's not really my workflow so other models might work better for other people.





Does it have some huge context window? Or is it really good at grep?

The Copilot version of this is just fucking terrible at suggesting anything remotely useful about our codebase.

I've had reasonable success just sticking single giant functions into context and asking Sonnet 4.5 targeted questions (is anything in this function modifying X, does this function appear to be doing Y) as a shortcut for reading through the whole thing or scattershot text search.

When I try to give it a whole file I actually hit single-query token limits.

But that's very "opt-in" on my part, and different from how I understand Cursor to work.


It is really good at grep and will make multiple grep calls in parallel.

And when I open it in the parent directory of a bunch of repos in our codebase, it can very quickly trace data flow through a bunch of different services. It will tell me all the files the data goes through.

It's context window is "only" 200k tokens. When it gets near 200k, it compresses the conversation and starts a new conversation..... which mostly works but sometimes it has a bit of amnesia if you have a really long running conversation on something.


> It is really good at grep and will make multiple grep calls in parallel.

How does that work? Multiple agents grepping simultaneously?


When other models would grep, then read results, then use search, then read results, then read 100 lines from a file, then read results, Composer 1 is trained to grep AND search AND read in one round trip It may read 15 files, and then make small edits in all 15 files at once

Presumably if it knows it needs to perform multiple searches in order to gather information (e.g. searching for redundant implementations of an algorithm, plus calls to the codebase's canonical implementation) it should be able to run those searches in parallel grep calls.

I'm trying to figure that one out.

LLMs are inherently single-threaded in how they ingest and produce info. So, as far as I can gather from the description, either it spawns sub-agents, or it has a tool dedicated for the job.


Probably something closer to ripgrep, if not actually ripgrep.



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: