The bigger issue is the advantage Anthropic, Google and OpenAI have in developing and deploying their own models. It wasn't that long ago that Cursor was reading 50 lines of code at a time to save on token costs. Anthropic just came out and yolo'd the context window because they could afford to, and it blew everything else away.
Cursor could release a cli tomorrow but it wouldn't help them compete when Anthropic and Google can always be multiples cheaper