Rendered at 07:40:49 GMT+0000 (Coordinated Universal Time) with Cloudflare Workers.
elicohen1000 12 minutes ago [-]
Been using Claude for pair programming since we're just two founders building MediTailor. It's wild - I can now prototype features that would have required hiring a full-time dev six months ago. The bottleneck shifted from "can we build this" to "should we build this" which is a much better problem to have.
edf13 51 minutes ago [-]
Nice - I do something similar in a semi manual way.
I do find Codex very good at reviewing work marked as completed by Claude, especially when I get Claude to write up its work with a why,where & how doc.
It’s very rare Claude has fully completed the task successfully and Codex doesn’t find issues.
axldelafosse 43 minutes ago [-]
I created the first version of loop after getting tired of doing this manually!
The vibes are great. But there’s a need for more science on this multi agent thing.
axldelafosse 3 hours ago [-]
I agree! Right now it is leveraging the Codex App Server, which is open-source and very well implemented, but using Claude Code Channels is probably a bit hacky.
The good thing is that it establishes a direct connection so it's already much better than having one agent spawn the other and wait for its output, or read/write to a shared .md file -- but it would be cool to make it work for all agent harnesses.
I prefer claude for generation / creativity, codex for bull-headed, accurate complaining and audit. Very rarely claude just doesn't "get it" and it makes sense to have codex direct edit. But generally I think it's happiest and best used complaining.
bradfox2 4 hours ago [-]
Multi turn review of code written by cc reviewed by codex works pretty well. Been one of the only ways to be able to deliver larger scoped features without constant bugs. I've seen them do 10-15 rounds of fix and review until complete.
Also implemented this as a gh action, works well for sentry to gh to auto triage to fix pr.
_ink_ 46 minutes ago [-]
How do you do this? Are you just switching between clis? Or is there a tool that uses the models in that way?
encoderer 3 hours ago [-]
Yes I’ve had a lot of success with this too. I found with prompt tightening I seldom do more than 5 rounds now, but it also does an explicit plan step with plan review.
Currently I’m authoring with codex and reviewing with opus.
Even with the same model (--self-review), that makes a huge difference, and immediately highlights how bad the first iterations of an LLM output can be.
I do find Codex very good at reviewing work marked as completed by Claude, especially when I get Claude to write up its work with a why,where & how doc.
It’s very rare Claude has fully completed the task successfully and Codex doesn’t find issues.
The good thing is that it establishes a direct connection so it's already much better than having one agent spawn the other and wait for its output, or read/write to a shared .md file -- but it would be cool to make it work for all agent harnesses.
Open to ideas! The repo is open-source.
Also implemented this as a gh action, works well for sentry to gh to auto triage to fix pr.
Currently I’m authoring with codex and reviewing with opus.
Even with the same model (--self-review), that makes a huge difference, and immediately highlights how bad the first iterations of an LLM output can be.