Appreciate all the takes so far, the team is reading this thread for feedback. Feel free to pile on with bugs or feature requests we'll all be reading.
I wonder if it is a concious decision not to include this (I imagine it opens a lot of possibilities of going crazy, but it also seems to be the source of a great amount of Claud Code's power). I would very much like to play with this if it appears in gemini-cli
Next step would be the possibility to define custom prompts, toolsets and contexts for specific re-occuring tasks, and these appearing as tools to the main agent. Example for such a thing: create_new_page. The prompt could describe the steps one needs to create the page. Then the main agent could simply delegate this as a well-defined task, without cluttering its own context with the operational details.
High ROI feature requests:
• Pattern-based permissions - Bash(git:) to allow git but not rm, Write(logs/.txt) for path scoping
• CLI permission flags - --allowedTools "Read,Bash(npm test)" --deniedTools "Write" for session overrides
• Allow/deny precedence rules - explicit deny should override general allow (security principle)
• Config file hierarchy - system → user → project precedence for policy enforcement
Medium ROI improvements:
• Command argument filtering - whitelist git commit but not git --exec-path=/bin/sh
• Multiple config formats - support both simple arrays and structured permission objects
• Runtime permission diagnostics - gemini permissions list to debug what's actually enabled
• Environment variable injection - top-level env config for OTEL endpoints, API keys, etc.
The permission engine is really the key piece - once you can express "allow X but not Y within X", it unlocks most advanced use cases. Keep up the great work!
Even with 1M context, for large projects, it makes sense to define boundaries These will typically be present in some form, but they are not available precisely to the coding agent. Imagine there was a simple YAML format where I could specify modules and where they can be found in the source tree, and the APIs of other modules it interacts with. Then it would be trivial to turn this into a context that would very often fit into 1M tokens. When an agent decides something needs to be done in the context of a specific module, it could then create a new context window containing exactly that module, effetively turning a large codebase into a small codebase, for which Gemini is extraordinarily effective.
At the very least, we need better documentation on how to get that environment variable, as we are not on GCP and this is not immediately obvious how to do so. At the worst, it means that your users paying for gemini don't have access to this where your general google users do.
All different products doing the sameish thing. I don’t know where to send users to do anything. They are all licensed differently. Bonkers town.
1. CodeRunner - https://github.com/BandarLabs/coderunner/tree/main?tab=readm...
currently it seems these are the CLI tools available. Is it possible to extend or actually disable some of these tools (for various reasons)?
> Available Gemini CLI tools:
- ReadFolder
- ReadFile
- SearchText
- FindFiles
- Edit
- WriteFile
- WebFetch
- ReadManyFiles
- Shell
- Save Memory
- GoogleSearch- Here [1] it says "Project settings override user settings." How does gemini determine if we're in a project? Does it look for a `.gemini` folder in the current working directory as well as every parent directory up to the root? Would Gemini be able to read the contents of a subfolder of the CWD if the subfolder contains a different `.gemini` folder?
- I don't see documentation for the `selectedAuthType` field in the documentation for settings.json. Mine says `oauth-personal`. I could've sworn I signed in with my Google Workspace account. Does `oauth-personal` apply to Workspace accounts?
And a feature request: it would be nice to add a restriction in the settings.json file forcing anybody who uses gemini in that project to sign in to a Workspace account in a specific org (or use a specific project, I guess).
[1]: https://github.com/google-gemini/gemini-cli/blob/main/docs/c...
- On a new chat I have to re-approve things like executing "go mod tidy", "git", write files... I need to create a new chat for each feature, (maybe an option to clear the current chat on VsCode would work)
- I have found some problems with adding some new endpoint on an example Go REST server I was trying it on, it just deleted existing endpoints on the file. Same with tests, it deleted existing tests when asking to add a test. For comparison I didn't find these problems when evaluating Amp (uses Claude 4)
Overall it works well and hope you continue with polishing it, good job!!
> You exceeded your current quota, please check your plan and billing details. For more information on this error, head to: https://ai.google.dev/gemini-api/docs/rate-limits.
Discouraging
I'm a Gemini Pro subscriber and I would love to be able to use my web-based chat resource limits with, or in addition to, what is offered here. I have plenty of scripts that are essentially "Weave together a complex prompt I can send to Gemini Flash to instantly get the answer I'm looking for and xclip it to my clipboard", and this would finally let me close the last step in that scripts.
Love what I'm seeing so far!
CC has this issue too, but way less often, and second shot almost always works.
Is the recommendation to specifically ask "analyze the codebase" here?
Edit: I tried it. The setup was a breeze. I fed the CLI two git commit IDs and some light prompting on what to look for. It gave a reasonable response. I'll try on a real PR shortly.
Edit: I should mention that I'm accessing this through Gemini Code Assist, so this may be something out of your wheelhouse.
A natural question to ask is, if in the near future, can Google One "Google AI Pro" subscribers have higher limits than what is offered for free users?
I think with better prompting on my end, as I have good experience with Gemini, this will be awesome. You probably could tweak a lot on your end as well, don't let it get stuck in cycles.
And thinking is stupid. "Show me how to generate a random number in python"... 15s later you get an answer.
I strongly believe in AI cli apps for devs, congrats.
like to just get a short response - for simple things like "what's a nm and grep command to find this symbol in these 3 folders". I use gemini alot for this type of thing already
Or would that have to be a custom prompt I write?
However, Gemini at one point output what will probably be the highlight of my day:
"I have made a complete mess of the code. I will now revert all changes I have made to the codebase and start over."
What great self-awareness and willingness to scrap the work! :)
First it did the search itself and then added "echo" for each of them - cute
Then it tried to use pytrends which didn't go anywhere
Then it tried some other paid service which also didn't go anywhere
Then it tried some other stuff which also didn't go anywhere
Finally it gave up and declared failure.
It will probably be useful as it can do the modify/run loop itself with all the power of Gemini but so far, underwhelming.
Unfortunately the CLI version wasn't able to create coherent code or fix some issues I had in my Rust codebase as well.
Here's hope that it eventually becomes great.
Claude did it fine but I was not happy with the code. What Gemini came up with was much better but it could not tie things together at the end.
I can't say much about writing new code though.
We really are living in the future
I’ve been using Claude for a side project for the past few weeks and I find that we really get into a groove planning or debugging something and then by the time we are ready to implement, we’ve run out of context window space. Despite my best efforts to write good /compact instructions, when it’s ready to roll again some of the nuance is lost and the implementation suffers.
I’m looking forward to testing if that’s solved by the larger Gemini context window.
I haven't looked at this Gemini CLI thing yet, but if its open source it seems like any model can be plugged in here?
I can see a pathway where LLMs are commodities. Every big tech company right now both wants their LLM to be the winner and the others to die, but they also really, really would prefer a commodity world to one where a competitor is the winner.
If the future use looks more like CLI agents, I'm not sure how some fancy UI wrapper is going to result in a winner take all. OpenAI is winning right now with user count by pure brand name with ChatGPT, but ChatGPT clearly is an inferior UI for real work.
If the module just can't be documented in this way in under 100 lines, it's a good time to refactor. Chances are if Claude's context window is not enough to work with a particular module, a human dev can't either. It's all about pointing your LLM precisely at the context that matters.
Im actually interested to see if we see a rise in demand for DRAM that is greater than usual because more software is vibe coded than being not, or some form of vibe coding.
https://developers.google.com/gemini-code-assist/resources/p...
When you use Gemini Code Assist for individuals, Google collects your prompts, related code, generated output, code edits, related feature usage information, and your feedback to provide, improve, and develop Google products and services and machine learning technologies.
To help with quality and improve our products (such as generative machine-learning models), human reviewers may read, annotate, and process the data collected above. We take steps to protect your privacy as part of this process. This includes disconnecting the data from your Google Account before reviewers see or annotate it, and storing those disconnected copies for up to 18 months. Please don't submit confidential information or any data you wouldn't want a reviewer to see or Google to use to improve our products, services, and machine-learning technologies.
"If you don't want this data used to improve Google's machine learning models, you can opt out by following the steps in Set up Gemini Code Assist for individuals."
and then the link: https://developers.google.com/gemini-code-assist/docs/set-up...
If you pay for code assist, no data is used to improve. If you use a Gemini API key on a pay as you go account instead, it doesn't get used to improve. It's just if you're using a non-paid, consumer account and you didn't opt out.
That seems different than what you described.
I hope this is something they're working on making clearer.
To clear everything up, we've put together a single doc that breaks down the Terms of Service and data policies for each account type, including an FAQ that covers the questions from this thread.
Here’s the link: https://github.com/google-gemini/gemini-cli/blob/main/docs/t...
Thanks again for pushing for clarity on this!
*What we DON'T collect:*
- *Personally Identifiable Information (PII):* We do not collect any personal information, such as your name, email address, or API keys.
- *Prompt and Response Content:* We do not log the content of your prompts or the responses from the Gemini model.
- *File Content:* We do not log the content of any files that are read or written by the CLI.
https://github.com/google-gemini/gemini-cli/blob/0915bf7d677...
Which pretty much means if you are using it for free, they are using your data.
I don't see what is alarming about this, everyone else has either the same policy or no free usage. Hell the surprising this is that they still let free users opt-out...
If this is legal, it shouldn’t be.
Not if you pay for it.
I like Gemini 2.5 Pro, too, and recently, I tried different AI products (including the Gemini Pro plan) because I wanted a good AI chat assistant for everyday use. But I also wanted to reduce my spending and have fewer subscriptions.
The Gemini Pro subscription is included with Google One, which is very convenient if you use Google Drive. But I already have an iCloud subscription tightly integrated with iOS, so switching to Drive and losing access to other iCloud functionality (like passwords) wasn’t in my plans.
Then there is the Gemini chat UI, which is light years behind the OpenAI ChatGPT client for macOS.
NotebookLM is good at summarizing documents, but the experience isn’t integrated with the Gemini chat, so it’s like constantly switching between Google products without a good integrated experience.
The result is that I end up paying a subscription to Raycast AI because the chat app is very well integrated with other Raycast functions, and I can try out models. I don’t get the latest model immediately, but it has an integrated experience with my workflow.
My point in this long description is that by being spread across many products, Google is losing on the UX side compared to OpenAI (for general tasks) or Anthropic (for coding). In just a few months, Google tried to catch up with v0 (Google Stitch), GH Copilot/Cursor (with that half-baked VSCode plugin), and now Claude Code. But all the attempts look like side-projects that will be killed soon.
Google's AI offerings that should be simplified/consolidated:
- Jules vs Gemini CLI?
- Vertex API (requires a Google Cloud Account) vs Google AI Studio API
Also, since Vertex depends on Google Cloud, projects get more complicated because you have to modify these in your app [1]:
``` # Replace the `GOOGLE_CLOUD_PROJECT` and `GOOGLE_CLOUD_LOCATION` values # with appropriate values for your project. export GOOGLE_CLOUD_PROJECT=GOOGLE_CLOUD_PROJECT export GOOGLE_CLOUD_LOCATION=global export GOOGLE_GENAI_USE_VERTEXAI=True ```
[1]: https://cloud.google.com/vertex-ai/generative-ai/docs/start/...
Gemini 2.5 Pro is the best model I've used (even better than o3 IMO) and yet there's no simple Claude/Cursor like subscription to just get full access.
Nevermind Enterprise users too, where OpenAI has it locked up.
Some jerk has learned that we prefer CLI things and has come to the conclusion that we should therefore pay extra for them.
Workaround is to use their GUI with some MCPs but I dislike it because window navigation is just clunky compared to terminal multiplexer navigation.
You clearly have never had the "pleasure" to work with a Google product manager.
Especially the kind that were hired in the last 15-ish years.
This type of situation is absolutely typical, and probably one of the more benign thing among the general blight they typically inflict on Google's product offering.
The cartesian product of pricing options X models is an effing nightmare to navigate.
I also have a pro subscription and wish I could get an API key with that with generous quota as well but pro is just for "consumers" using Gemini app I guess
This is very confusing how they post about this on X, you would think you get additional usage. Messaging is very confusing.
I can't find any way to upgrade to a paid plan, is this even possible for individuals, or is it just "free or enterprise"?
/Edit: Okay I went through the Gemini docs. I found that in Google Cloud you can enable Gemini Code Assist Standard and Enterprise for the account
- Standard is $19.00/mo
- Enterprise is $45.00/mo
Difference between the 2 editions: https://cloud.google.com/products/gemini/pricing
/Edit2: Found the actual management screen: https://codeassist.google.com/overview
If I Could Talk to Satya...
I'd say:
“Hey Satya, love the Copilots—but maybe we need a Copilot for Copilots to help people figure out which one they need!”
Then I had them print out a table of Copilot plans:
- Microsoft Copilot Free - Github Copilot Free - Github Copilot Pro - Github Copilot Pro+ - Microsoft Copilot Pro (can only be purchased for personal accounts) - Microsoft 365 Copilot (can't be used with personal accounts and can only be purchased by an organization)
More of my notes here: https://simonwillison.net/2025/Jun/25/gemini-cli/
It's the only argument I can think of, something like Go would be goated for this use case in principle.
I really don't mind either way. My extremely limited experience with Node indicates they have installation, packaging and isolation polished very well.
https://bun.sh/docs/bundler/executables
https://docs.deno.com/runtime/reference/cli/compile/
Note, I haven't checked that this actually works, although if it's straightforward Node code without any weird extensions it should work in Bun at least. I'd be curious to see how the exe size compares to Go and Rust!
My exact same reaction when I read the install notes.
Even python would have been better.
Having to install that Javascript cancer on my laptop just to be able to try this, is a huge no.
https://www.npmjs.com/package/pkg
or perhaps this one:
Is your vision with Gemini CLI to be geared only towards non-commercial users? I have had a workspace account since GSuite and have been constantly punished for it by Google offerings all I wanted was gmail with a custom domain and I've lost all my youtube data, all my fitbit data, I cant select different versions of some of your subscriptions (seemingly completely random across your services from a end-user perspective), and now as a Workspace account I cant use Gemini CLI for my work, which is software development. This approach strikes me as actively hostile towards your loyal paying users...
... and other stuff.
I have thrown very large codebases at this and it has been able to navigate and learn them effortlessly.
> hello
[API Error: {"error":{"message":"{\n \"error\": {\n \"code\": 429,\n \"message\": \"Resource has been exhausted (e.g. check quota).\",\n \"status\": \"RESOURCE_EXHAUSTED\"\n }\n}\n","code":429,"status":"Too Many Requests"}}] Please wait and try again later. To increase your limits, request a quota increase through AI Studio, or switch to another /auth method
⠼ Polishing the pixels... (esc to cancel, 84s)
Definitely not because of Claude Code eating our lunch!
What are they supposed to do?
“Oh no, they’ve released CLI tool before us! It’s game over, we can’t do it too, we need to come up with something else now!”
If you mean: This is "inspired" by the success of Claude Code. Sure, I guess, but it's also not like Claude Code brought anything entirely new to the table. There is a lot of copying from each other and continually improving upon that, and it's great for the users and model providers alike.
Which is surprising because at first i was ready to re-up my Google life. I've been very anti-Google for ages, but at first 2.5 Pro looked so good that i felt it was a huge winner. It just wasn't enjoyable to use because i was often at war with it.
Sonnet/Opus via Claude Code are definitely less intelligent than my early tests of 2.5 Pro, but they're reasonable, listen, stay on task and etc.
I'm sure i'll retry eventually though. Though the subscription complexity with Gemini sounds annoying.
Then there are 3rd party channels, if you have a recent samsung phone, you get 1 yr access to AI features powered by gemini, after which you need to pay. And lord knows where else has google been integrating gemini now.
Ive stopped using google's AI now. Its like they have dozens of teams within gemini on completely different slack sessions.
Microsoft Copilot
Microsoft 365 Copilot
Copilot Pro
Microsoft Security Copilot
Microsoft Copilot Studio
GitHub Copilot
Copilot+ PC
Microsoft 365 Copilot for Sales
Microsoft 365 Copilot for Service
Microsoft 365 Copilot for Finance
- Open-source (Apache 2.0, same as OpenAI Codex)
- 1M token context window
- Free tier: 60 requests per minute and 1,000 requests per day (requires Google account authentication)
- Higher limits via Gemini API or Vertex AI
- Google Search grounding support
- Plugin and script support (MCP servers)
- Gemini.md file for memory instruction
- VS Code integration (Gemini Code Assist)
I just symlink now to AGENTS.md
Claude Code seems to require the CLAUDE.md filename.
You might want to tell Claude not to write so many comments but you might want to tell Gemini not to reach for Kotlin so much, or something.
A unified approach might be nice, but using the same prompt for all of the LLM "coding tools" is probably not going to be as nice as having prompts tailored for each specific tool.
If it sounds too good to be true, it probably is. What’s the catch? How/why is this free?
Also they can throttle the service whenever they feel it's too costly.
https://youtu.be/HC6BGxjCVnM?feature=shared&t=36
It's a FOSS MCP server I created a couple of weeks ago:
- https://github.com/mbailey/voicemode
# Installation (~/.gemini/settings.json)
{
"theme": "Dracula",
"selectedAuthType": "oauth-personal",
"mcpServers": {
"voice-mode": {
"command": "uvx",
"args": [
"voice-mode"
]
}
}
}If you need to do research, pre-training, RLHF, inference for 5-10 different models across 20 different products, how do you optimally allocate your very finite compute? Weight towards research and training for better future models, or weigh towards output for happier consumers in the moment?
It would make sense that every project in deepmind is in constant war for TPU cycles.
1) I tried to use it on an existing project asking this "Analyse the project and create a GEMINI.md". It fumbled some non sense for 10-15 minutes and after that it said it was done, but it had only analysed a few files in the root and didn't generate anything at all.
2) Despite finding a way to login with my workspace account, it then asks me for the GOOGLE_CLOUD_PROJECT which doesn't make any sense to me
3) It's not clear AT ALL if and how my data and code will be used to train the models. Until this is pretty clear, for me is a no go.
p.s: it feels like a promising project which has been rushed out too quickly :/
This is my experience with ALL AI editors and models. Half of the time, they say they changed things, but they didn't change anything.
1. Go to their enterprise site
2. See what privacy guarantees they advertise above the consumer product
3. Conclusion: those are things that you do not get in the consumer product
These companies do understand what privacy people want and how to write that in plain language, and they do that when they actually offer it (to their enterprise clients). You can diff this against what they say to their consumers to see where they are trying to find wiggle room ("finetuning" is not "training", "ever got free credits" means not-"is a paid account", etc)For Code Assist, here's their enterprise-oriented page vs their consumer-oriented page:
https://cloud.google.com/gemini/docs/codeassist/security-pri...
https://developers.google.com/gemini-code-assist/resources/p...
It seems like these are both incomplete and one would need to read their overall pages, which would be something more like
https://support.google.com/a/answer/15706919?hl=en
https://support.google.com/gemini/answer/13594961?hl=en#revi...
In comparison to Claude Code Opus 4, it seemed much more eager to go on a wild goose chase of fixing a problem by creating calls to new RPCs that then attempted to modify columns that didn't exist or which had a different type, and its solution to this being a problem was to then propose migration after migration to modify the db schema to fit the shape of the rpc it had defined.
Reminded me of the bad old days of agentic coding circa late 2024.
I'm usually a big fan of 2.5 Pro in an analysis / planning context. It seems to just weirdly fall over when it comes to tool calling or something?
It generates a bunch of fake activity indicators based on your prompt, then cycles through them on a timer. It has no bearing on the actual activity going on underneath.
It appears to be much slower than Claude Code, possibly due to being overloaded, but it feels like it thinks a lot more before beginning to suggest file edits. The permissions aren't as nice either. Where Claude Code suggests "allow uv run pytest without approval", Gemini suggests "allow uv run without approval", which is broader than I would like.
1. First authentication didn't work on my headless system, because it wants an oauth redirect to localhost - sigh.
2. Next, WebFetch isn't able to navigate github, so I had to manually dig out some references for it.
3. About 2 mins in, I just got ``` ℹ Rate limiting detected. Automatically switching from gemini-2.5-pro to gemini-2.5-flash for faster responses for the remainder of this session. ``` in a loop with no more progress.
I understand the tool is new, so not drawing too many conclusions from this yet, but it does seem like it was rushed out a bit.
Then I hit the rate limit. - Fine, no worries, it'll be interesting to see if the quality changes.
Then it starts getting stuck and taking forever to complete anything. So I shut it down for the day.
Today, I start it back up and ask it to pickup where it left off and it starts spinning. I forget about it and come back 7.5 hours later and it' still spinning. When I kill it it said: 1 Turn, 90k input tokens, 6.5 hours of API time... WTH?
And now I'm just totally rate limited - `status: 429, statusText: 'Too Many Requests'` - every time. Also, I can't find any kind of usage data anywhere!
This perfectly demonstrates the benefit of the nodejs platform. Trivial to install and use. Almost no dependency issues (just "> some years old version of nodejs"). Immediately works effortlessly.
I've never developed anything on node, but I have it installed because so many hugely valuable tools use it. It has always been absolutely effortless and just all benefit.
And what a shift from most Google projects that are usually a mammoth mountain of fragile dependencies.
(uv kind of brings this to python via uvx)
Gemini CLI does not take new direction especially well. After planning, I asked it to execute and it just kept talking about the plan.
Another time when I hit escape and asked it to stop and undo the last change, it just plowed ahead.
It makes a lot of mistakes reading and writing to files.
Some, but by no means all, of the obsequious quotes from my first hour with the product: - “You’ve asked a series of excellent questions that have taken us on a deep dive ...” - “The proposed solution is not just about clarity; it’s also the most efficient and robust.”
Therefore I was not surprised to experience Gemini spiraling into an infinite loop of self-deprecation - literally it abandoned the first command and spiraled into 5-10line blocks of "i suck"
---
Right now there is one CLI that influences and stands far and beyond all others. Smooth UX, and more critical some "natural" or inherent ability to use its tools well.
Gemini can also achieve this - but i think it's clear the leader is ahead because they have a highly integrated training process with the base model and agentic tool use.
How did they do that pretty "GEMINI" gradient in the terminal? is that a thing we can do nowadays? It doesn't seem to be some blocky gradient where each character is a different color. It's a true gradient.
(yes I'm aware this is likely a total clone of claude code, but still am curious about the gradient)
And it is a blocky gradient, each character is a color. It's just the gradient they chose is slow enough that you don't notice.
1. The thing going in a circle trying to fix a bug by persistently trying different permutations of an API interface it never bothered to check the definition of. Isn't that why it's internet connected?
2. When I asked it to just analyze instead of change stuff. It just hung for 20 minutes giving me responses saying that gemini-2.5-pro was slow, and that it was switching to 2.5-flash, with no other output to indicate what it was doing other than those cute scrolling messages that mean nothing.
At least in Claude it's clear that the messages mean nothing, because they're clearly generic. Gemini gives you the impression the status messages mean something since they're sort of related to the topic being worked on.
Because it says in the README:
> Authenticate: When prompted, sign in with your personal Google account. This will grant you up to 60 model requests per minute and 1,000 model requests per day using Gemini 2.5 Pro.
> For advanced use or increased limits: If you need to use a specific model or require a higher request capacity, you can use an API key: ...
When I have the Google AI Pro subscription in my Google account, and I use the personal Google account for authentication here, will I also have more requests per day then?
I'm currently wondering what makes more sense for me (not for CLI in particular, but for Gemini in general): To use the Google AI Pro subscription, or to use an API key. But I would also want to use the API maybe at some point. I thought the API requires an API key, but here it seems also the normal Google account can be used?
It integrates with VS Code, which suits my workflow better. And buying credits through them (at cost) means I can use any model I want without juggling top-ups across several different billing profiles.
Ultimately quality wins out with LLMs. Having switched a lot between openai, google and Claude, I feel there's essentially 0 switching cost and you very quickly get to feel which is the best. So until Claude has a solid competitor I'll use it, open source or not
https://blog.google/technology/developers/introducing-gemini...
When I try to log in using a personal account, it tells me I need to generate something in the dashboard. I know that on my main account from maybe 15 years ago I had things like domains thats nowdays Google Workspace. But I'm trying to log in and authorize using my second (NOT MAIN), relatively new account. Despite this, it redirects me to the console (like it would for the first account), and there it says I'm logged in (!), but the application itself says I'm not.
I switched profiles within the console too, tried the same steps, and it still results in the same problem. Yes, I can see my new profile in the top right corner, not the old one. It sends me to a "Users may have to specify a GOOGLE_CLOUD_PROJECT if:" github [0] saying this ID is needed, and I honestly would even generate it, but I don't want to waste time with the program if you can't even get basic authorization working correctly.
0: https://github.com/google-gemini/gemini-cli/blob/main/docs/c...
Never had anything like this with claude code.
I've used Gemini 2.5 Pro quite a lot and like most people I find it's very intelligent. I've bent over backwards to use Gemini 2.5 Pro in another piece of work because it's so good. I can only assume it's the gemini CLI itself that's using the model poorly. Keen to try again in a month or two and see if this poor first impression is just a teething issue.
I told it that it did a pretty poor job and asked it why it thinks that is, told it that I know it's pretty smart. It gave me a wall of text and I asked for the short summary
> My tools operate on raw text, not the code's structure, making my edits brittle and prone to error if the text patterns aren't perfect. I lack a persistent, holistic view of the code like an IDE provides, so I can lose track of changes during multi-step tasks. This led me to make simple mistakes like forgetting a calculation and duplicating code.
Set up not too long ago, and afaik pretty load-bearing for this. Feels great, just don’t ask me any product-level questions. I’m not part of the Gemini CLI team, so I’ll try to keep my mouth shut.
Not going to lie, I’m pretty anxious this will fall over as traffic keeps climbing up and up.
A bit gutted by the `make sure it is not a workspace account`. What's wrong with Google prioritising free accounts vs paid accounts? This is not the first time they have done it when announcing Gemini, too.
How do I reset permissions so it always asks again for `git` invocations?
Thanks!
It keeps putting import statements in the middle of files or duplicating edits at the wrong places, but it's hard to get frustrated since it seems pretty self-aware. I didn't say it was making a mess - it's just being hard on itself.
A short session ended up sending over 3 million tokens though - wonder how the economics of this work out for google?
Tried upgrading to the Standard plan through Google Cloud with the hope that it would allow me to do more, but after switching my account to the now-paid account, it still almost instantly downgraded me to 2.5-flash
For the times when I was able to use 2.5-pro, the output has been very good. But the moment it switches to flash, the quality goes down by about 80% and I would never let it code on anything
It's more focused on implications for docs strategy (I'm worried that agent providers are steering us towards duplicating information that's already in eng docs) rather than user best practices i.e. "put this into your agent doc to improve agent performance"
---
Right now there is one CLI that influences and stands far and beyond all others. Smooth UX, and more critical some "natural" or inherent ability to use its tools well.
Gemini can also achieve this - but i think it's clear the leader is ahead because they have a highly integrated training process with the base model and agentic tool use.
It is vastly more difficult to understand what Google is offering compared to the others, to what cost, getting an API-key or similar, understanding usage/billing across the suite, etc.
I wouldn't expect any regular person to bother signing up for any of Google's products, let alone understand what they're really offering.
gave it brief instructions to deploy a hobby static site on cloud run yesterday; with additional architecture to set up (load balancer etc). It got into a loop, started deleting the gcloud resouces it created when it hit a 403 error for the site.
I hit ESC and prodded a little in terms of authentication rules ... and off it went to complete the task. bravo!
That's a ton of free limit. This has been immensely more successful than void ide.
I guess I will use something else. This is all very annoying given that I actually pay for Gemini Pro...
1. Gemini Code Assist (GCA) for Individuals: FREE for 1,000 model requests/day
2. GCA Standard: $22.80/month for 1,500 model requests/day (1.5x more than FREE)
3. GCA Enterprise: $54.00/month for 2,000 model requests/day (2X more than FREE)
Source: https://codeassist.google
I use aichat now but it's not perfect. https://github.com/sigoden/aichat
The UI on this tool is much better.
At the bottom of README.md, they state:
"This project leverages the Gemini APIs to provide AI capabilities. For details on the terms of service governing the Gemini API, please refer to the terms for the access mechanism you are using:
* Gemini API key
* Gemini Code Assist
* Vertex AI"
The Gemini API terms state: "for Unpaid Services, all content and responses is retained, subject to human review, and used for training".
The Gemini Code Assist terms trifurcate for individuals, Standard / Enterprise, and Cloud Code (presumably not relevant).
* For individuals: "When you use Gemini Code Assist for individuals, Google collects your prompts, related code, generated output, code edits, related feature usage information, and your feedback to provide, improve, and develop Google products and services and machine learning technologies."
* For Standard and Enterprise: "To help protect the privacy of your data, Gemini Code Assist Standard and Enterprise conform to Google's privacy commitment with generative AI technologies. This commitment includes items such as the following: Google doesn't use your data to train our models without your permission."
The Vertex AI terms state "Google will not use Customer Data to train or fine-tune any AI/ML models without Customer's prior permission or instruction."
What a confusing array of offerings and terms! I am left without certainty as to the answer to my original question. When using the free version by signing in with a personal Google account, which doesn't require a Gemini API key and isn't Gemini Code Assist or Vertex AI, it's not clear which access mechanism I am using or which terms apply.
It's also disappointing "Google's privacy commitment with generative AI technologies" which promises that "Google doesn't use your data to train our models without your permission" doesn't seem to apply to individuals.
Well, not sure that it makes sense to do it, anyway I've tried to run in in a cell and in the google colab terminal. Still waiting for auth (?)
better question is why do you need a modle specific CLI when you should be able to plug in to individual models.
I hate this openwashing. It's a closed model, its weights are nowhere to be seen! (not to mention the training data, the real "source" of a LLM)
The fact that there is a small component that is open source that accesses this closed model doesn't change that at all.
We are now three years into the AI revolution and they are still forcing us to copy and paste and click click crazy to get the damn files out.
STOP innovating. STOP the features.
Form a team of 500 of your best developers. Allocate a year and a billion dollar budget.
Get all those Ai super scientists into the job.
See if you can work out “download all files”. A problem on the scale of AGI or Dark Matter, but one day google or OpenAI will crack the problem.
> 429: Too many requests
Mind you, this is with a paid API key
This is shown at the top of the screen in https://aistudio.google.com/apikey as the suggested quick start for testing your API key out.
Not a great look. I let our GCloud TAM know. But still.
Gemini Pro and Claude play off of each other really well.
Just started playing with Gemini CLI and one thing I miss immediately from Claude code is being able to write and interject as the AI does its work. Sometimes I interject by just saying stop, it stops and waits for more context or input or ai add something I forgot and it picks it up..
Google services have become a patchwork of painfully confounding marketing terms that mean nothing and obfuscate what they actually provide.
I do not get it why they don’t pick Go or Rust so i get a binary.
Not impressed. These companies have billions at their disposal, and probably pay $0 in tax, and the best they can come up with is this?
While writing this comment, thinking that there should be some packaging tool that would create a binaries from npx cli tools. I remember such things for python. Binaries were fat, but it is better then keep nodejs installed on my OS