Right now:
- You can't interrupt Claude (you press stop and he keeps going!)
- At best it stops but just keeps spinning
- The UI disconnects intermittently
- It disconnects if you switch to other parts of Claude
- It can get stuck in plan mode
- Introspection is poor
- You see XML in the output instead of things like buttons
- One session at a time
- Sessions at times don't load
- Everytime you navigate away from Code you need to wait for your session to reappear
I'm sure I'm missing a few things.
I thought coding was a solved problem Boris?
Claude Code (the product, not the underlying model) has been one of the buggiest, least polished products I have ever used. And it's not exactly rocket science to begin with. Maybe they should try writing slightly less than 100% of their code with AI?
You get a buggy electron app and they get billions in valuation.
Clearly no one values quality anymore. 1000% yolo
My current solution uses Tailscale with Termius on iOS. It's a pretty robust solution so far, except for the actual difficulty of reading/working on a mobile screen. But for the most part, input controls work.
My one gripe with Termius is that I can't put text directly into stdin using the default iOS voice-to-text feature baked into the keyboard.
[1] https://elliotbonneville.com/phone-to-mac-persistent-termina...
[2] https://elliotbonneville.com/claude-code-is-all-you-need/
Wrote a daemon + mobile app (similar to Happy, but fixed a lot of the problems) and baked in Tailscale support.
Will open source it soon and should have an official release in the next few weeks: https://getroutie.com/
It also feels kind of nice to just fire off an email and let it do it's thing.
They bring down production because the version string was changed incorrectly to add an extra date. That would have been picked up in even the most basic testing since the app couldn't even start.
https://news.ycombinator.com/item?id=46532075
The fix (not even a PR or commit message to explain) https://github.com/anthropics/claude-code/commit/63eefe157ac...
No root cause analysis either https://github.com/anthropics/claude-code/issues/16682#issue...
Sounds like a problem AI can easily solve!
First of all /remote-control in the terminal just printed a long url. Even though they advertise we can control it from the mobile app (apparently it should show a QR code but doesn't). I fire up the mobile app but the session is nowhere to be seen. I try typing the long random URL in the mobile browser, but it simply throws me to the app, but not the session. I read random reddit threads and they say the session will be under "Code", not "Chats", but for that you have to connect github to the Claude app (??, I just want to connect to the terminal Claude on my PC, not github). Ok I do it.
Now even though the session is idle on the pc, the app shows it as working... I try tapping the stop button, nothing happens. I also can't type anything into it. Ok I try starting a prompt on the pc. It starts the work on the PC, but on the mobile app I get a permission dialog... Where I can deny or allow the thing that actually already started on the pc because I already gave permission for that on the PC. And many more. Super buggy.
I wonder if they let Claude write the tests for their new features... That's a huge pitfall. You can think it works and Claude assures you all is fine but when you start it everything falls apart because there are lots of tests but none actually test the actual things.
I had to downgrade to an earlier release because an update introduced a regression where they weren't handling all of their own event types.
This is normal behavior on desktop sometimes its in the middle of something? I also assume there's some latency
> - At best it stops but just keeps spinning
Latency issues then?
> - It can get stuck in plan mode
I've had this happen from the desktop, and using Claude Code from mobile before remote control, I assume this has nothing to do with remote control but a partial outage of sorts with Claude Code sometimes?
I don't work for Anthropic, just basing off my anecdotal experience.
Frequently chews through lots of expensive Opus tokens, then it just stops with no communication about why or what's next.
No way to tell what it's done, what's remaining to complete.
Only choice is to re-run everything and eat the cost of the wasted time and tokens.
We’ve been building in this space for a while, and the issues listed here are exactly the hard parts: session connectivity, reconnection logic, multi-session UX, and keeping state in-sync across devices. Especially when it comes to long running tasks and the edge cases that show up in real use.
- - -
get tailscale (free) and join on both devices
install tmux
get an ios/android terminal (echo / termius)
enable "remote login" if on mac (disable on public wifi)
mosh/ssh into computer
now you can do tmux then claude / codex / w/e on either device and reconnect freely via tmux ls and tmux attach -t <id>
- - -
You can name tmux and resume by name via tmux new -s <feature> and tmux attach -t <feature>
Been testing it today with Claude Code and it seems to work quite well switching between my laptop and phone.
[1]: https://zellij.dev and https://github.com/zellij-org/zellij
[2]: https://github.com/crigler/dtach and https://dtach.sourceforge.net
Based on my experience many people don't know this is a thing you can do.
How do you deal with multiple concurrent sessions of CC with this setup?
How important is mosh? I wasn't able to get it set up the last time I tried... ran into a bunch of issues.
[terminal.shell]
args = ["-l", "-c", "tmux attach || tmux"]
program = "/opt/homebrew/bin/zsh"
tmux supports tabs so you can have multiple Claude Code sessions running concurrently. You do need to learn a few tmux keyboard shortcuts to use it effectively (e.g. opening/closing/switching tab).Depends- how good is your signal? Mosh has a great property that it buffers everything locally so there's no lag even if your connection sucks.
On ssh, every keystroke is a roundtrip
Could even use cc to check in on and/or "send-keys"
What wasn't working about mosh? Just install mosh and use mosh to connect
It's this.
Don't have a Dropbox moment ;) [1]
You can test it right now if you want with the included free models.
(I am actually using zellij on the remote and using various CLIs more than I am using only opencode on the web. I was using wezterm mux until about a week ago but the current state of the terminal is not very good for this scenario. It seems like almost all the CLIs are choking because of nodejs ink library)
A fork named Happier looks promising, but is alpha-stage and is also a mystery-meat vibe-coded security roulette.
I also remember base cable without any movies was around $60 or something and with some movie channels is >$100. And that's not inflation adjusted. You can easily get 3 or 4 of the top services for $100 today.
Finally claiming there are more ads on these services is a joke. There was ~20m for every 30m of programming, meaning 1/3 of the time you're watching commercials. And not just any commercials, the same commercials over and over. There was even a case of shows being sped up on cable to show more commercials.
I get it, everyone wants everything seamlessly and for next to nothing, but claiming that 90s cable was even comparable is absurd.
https://www.digitaltrends.com/home-theater/how-networks-spee...
You can get all 7 of the major streaming subs for less without even shopping out deals. That is 100s of times the volume and quality of content that was delivered on cable for far less. It is so much content realistically that no one I have ever met has subscribed to all of them at once.
The argument really is empty. The fragmentized experience is annoying, but it isn't more expensive...And it DEFINITELY has fewer ads.
I literally see no ads on my streaming subscription for close to a tenth of the price of cable.
The results are enough for me and I'm not doing things that allow me to differentiate the output between ChatGPT, Claude and, the others.
The agents are more like the radio in my car, whenever I want music, I switch channel until I find something good enough.
If I'm really in need of something special, I'll use Spotify on my phone.
And sometimes, I just drive with the radio off.
There's a comparison of the approaches as I see them here https://yepanywhere.com/subscription-access-approaches
I was starting to think I've really fallen behind or something. I feel relief AND horror.
And in modern stacks, it almost necessitates a man in the middle - tailscale is common but it's still a central provider. So is it really the most inefficient way possible?
The daily “what broke and changed now” with claude code is wearing me out fast.
Seriously, if you listen to Dario talk, it's non-stop how there's a tsunami coming, people have no idea of what's ahead, how general intelligence or superintelligence is right around the corner. But also, this is super dangerous, and only Anthropic can save us from a doomsday scenario.
Coding with AI requires immense restraint and strong scope limits.
There's an open issue on github for it:
Why does the remote control needs that? For what?
I rather use the common developer tools like termux or mosh etc. on a phone if I need that functionality.
Also, I felt the need to use it far more when I was on Pro vs a Max plan. On Pro when you hit the usage windows it's nice to be able to kick claude back into gear without scheduling your life around getting back to the terminal to type "continue".
- Some of us don't do full yolo mode all the time, then tool approvals or code reviews are required, nice to do a quick review and decide if you need to go back to your computer or not
- Letting claude spin or handle a long-running task outside of normal work hours and being able to check in intermittently to see if something crashed
The remote control feature is cool but the real unlock for me was voice. Typing on a phone is a terrible interface for coding conversations. Speaking is surprisingly natural for things like "check the test output" or "what did that agent do while I was away."
The tmux crowd in this thread is right that SSH + tmux gets you 90% of the way there. But adding voice on top changes the interaction model. You stop treating it like a terminal and start treating it like a collaborator.
Here is a demo of it controlling my smart lights: https://www.youtube.com/watch?v=HFmp9HFv50s
Not sure if we have any LLM-tooling for the latter, seems to be more about how you use the tools we have available, but they're all pulling us to be "do first, think later" so unless you're careful, they'll just assume you want to do more and think less, hence all the vibeslop floating around.
Given the number of CC users I know who spend significant time on creating/iterating designs and specs before moving to the coding phase, I can tell you, your assumption is wrong. Check how different people actually use it before projecting your views.
I would argue that rapidly iterating reveals more about the problem, even for the most thoughtful of us. It's not like you check your own reasoning at the door when you just dive head first into something.
So even if it comes at the expense of long term maintainability - everyone should have this in their toolbox.
i think it still pulls to do then think because you cant tell what the agent understood of what you asked it to do from that first think, until its actually produced something.
Claude Code and similar agents help me execute experiments, prototypes and full designs based on ideas that I have been refining in my head for years, but never had the time or resources to implement.
They also help get me past design paralysis driven by overthinking.
Perhaps the difference between acceleration and slop is the experience to know what to keep, what to throw away, and what to keep refining.
I wonder if is anyone working on an AI framework that encourages us to keep our eye on the big picture, then walk away when a reasonable amount of work is done for the day.
Yes, individuals are creating cool mobile coding solutions and Anthropic doesn't want to get left behind. I know I'm working my ass off at work right now because LLM coding makes it fun, but I also often don't prioritize what I'm doing for the big picture because I just try every thing that comes into my inbox, in order, because it's so fast to do with Claude Code.
We all sense it!: <https://newsroom.haas.berkeley.edu/ai-promised-to-free-up-wo...> <https://ghuntley.com/teleport/> <https://steve-yegge.medium.com/the-ai-vampire-eda6e4f07163>
My favorite way to vibe code is by voice while in the hot tub. Rest AND focus AND build.
The one feature drawback of tailscale/tmux/termius is no file upload. And ergonomics, ability to view files/diffs easily, though that's subjective.
With e.g. tmux you'll piggyback on decades of SSH development.
If you'd have to switch to a different tool to do your coding that's not vendor lock in.
For example, maybe I have an idea for a feature and I want to spin up a new branch and have agents work on that. But then I get stuck or bored (I'm talking personal usage), so decide to park it. But maybe after a few days I have a shower thought and want to resume it.
The current method of listing sessions and resuming them can work, but you need to find the right session. If there is something that shows all the branches, a docs overview of what that feature it, and the current progress it would make this workflow a lot more effective. Plus I switch LLMs when I hit rate limits.
I'm probably going to just build it myself, but wondering if anyone has something that does this already.
But if you don’t want to, I’ve been building basically this https://github.com/Robdel12/OrbitDock
Native macOS and iOS apps, backed by a rust binary that I put anywhere and connect to. Right now I’m just LAN but eventually will tailscale.
Works with claude and codex. Both passively watching an active CLI session for both and you can take over those sessions if needed and interact in the app
Project is at http://github.com/vincent-163/claude-code-multi/. Can be installed easily with nodejs.
Please provide feedbacks and suggestions!
Typing "/clear" in the terminal clears it, but the Claude iOS app just outputs raw XML instead and doesn't actually do anything:
<command-name>/clear</command-name>
<command-message>clear</command-message>
<command-args></command-args>
<local-command-stdout></local-command-stdout>I built Crabigator[1] and it's a wrapper around `claude` and `codex`, so its ready for coding on the go on start and already streaming. Plus, crabigator shows many parallel windows, separated by repo/project/machine, so you can manage multiple agents seamlessly.
Excited to see how this matures so people without that inclination can also be constantly pestered by the nagging idea that someone, somewhere is being more productive than them :)
I often get ideas while I'm in bed or outside away from my computer, and was thinking that the ability to code on your computer from your phone, through AI, would be such a killer app.
My favorite use case would be asking the AI to review code and going over its findings/suggestions while I'm away from the computer or trying to fall asleep.
Gives me full sandboxing with bypass permissions, tmux, and cherry-pick level control over what gets pulled back out into my work dir.
Mix in tailscale and I can control it from anywhere, on any device, with full transferrability using established and battle proven tooling.
You all can pretend the software dev cycle hasn't changed... get real.
I was using the Claude app on my iPhone and claude code on a MacBook pro. PR is merged, still on the bike :)
Imagine if tomorrow they make a 10x smarter AI, but they say: you can only use if you upload your source code to us and you can’t see anymore the source code.
So you either stay on lower end models or you give up and use a 10x model.
I only see one issue: will be very difficult for them to “guard” the source code and don’t let you access.
I wonder if they would take away your ability to prompt, maybe only letting you run agentically.
For the vibe'y workflows, this would easily solve parallel long running work without skipping permissions: schedule 10 different tasks and go for a run. Occasionally review what the hallucination machine wants to do, smash yes a few times, occasionally tell it not to be silly, have a nice run. Essentially, solving remote development, though perhaps not quite in the way how people usually think of it.
> Limitations
> One remote session at a time: each Claude Code session supports one remote connection.
Hmm. Give it 1-12 months.
Claude Code only supports logging out the current session via /logout
There's no logout all sessions equivalent unlike the web UI.
jfc no
So your hook -> HA -> push notification? And then you just tap to approve?
I've been building something similar, basically a way to run your full dev environment on your Mac and connect to it from iOS. terminal, files, AI agent all talking to the same session. the tricky part is honestly just keeping state in sync when you switch devices.
This is not to say engineers are getting replaced — but, certainly, they are changing their work. And, sure, maybe _some_ of them are being replaced. Not most of the ones I know, though. They are essential to orchestrate, curate, maintain, and drive all of this.
(Now, do they want to orchestrate it? Whole different story...)
And of course, OpenClaw is built to be a very generalist agent with a chat interface - same effective outcome as remotely controlling an AI harness, but not exactly what everyone wants.
Seeing how the labs tend to copy the best functionality in any FOSS developments, I decided to wait - happy I did, here's the official functionality for this that is much more trustworthy.