zlacker

The Codex App

submitted by meetpa+(OP) on 2026-02-02 18:02:48 | 803 points 629 comments
[view article] [source] [go to bottom]

NOTE: showing posts with links only show all posts
4. strong+J4[view] [source] 2026-02-02 18:25:02
>>meetpa+(OP)
Genuinely excited to try this out. I've started using Codex much more heavily in the past two months and honestly, it's been shockingly good. Not perfect mind you, but it keeps impressing me with what it's able to "get". It often gets stuff wrong, and at times runs with faulty assumptions, but overall it's no worse than having average L3-L4 engs at your disposal.

That being said, the app is stuck at the launch screen, with "Loading projects..." taking forever...

Edit: A lot of links to documentation aren't working yet. E.g.: https://developers.openai.com/codex/guides/environments. My current setup involves having a bunch of different environments in their own VMs using Tart and using VS Code Remote for each of them. I'm not married to that setup, but I'm curious how it handles multiple environments.

Edit 2: Link is working now. Looks like I might have to tweak my setup to have port offsets instead of running VMs.

5. samuel+T4[view] [source] 2026-02-02 18:25:58
>>meetpa+(OP)
It's basically what Emdash (https://www.emdash.sh/), Conductor (https://www.conductor.build/) & CO have been building but as first class product from OpenAI.

Begs the question if Anthropic will follow up with a first-class Claude Code "multi agent" (git worktree) app themselves.

18. xGrill+X5[view] [source] 2026-02-02 18:30:20
>>meetpa+(OP)
Is this not just a skinned version of Goose: https://block.github.io/goose/
◧◩
19. FanaHO+56[view] [source] [discussion] 2026-02-02 18:31:11
>>samuel+T4
https://code.claude.com/docs/en/desktop
21. asdev+m6[view] [source] 2026-02-02 18:31:48
>>meetpa+(OP)
Built an open source lightweight version of this that works with any cli agent: https://github.com/built-by-as/FleetCode
◧◩◪
56. embiri+ja[view] [source] [discussion] 2026-02-02 18:52:51
>>romain+Z9
Only thing i'd add re windows is it's taking us some time to get really solid sandboxing working on Windows, where there are fewer OS-level primitives for it. There's some more at https://developers.openai.com/codex/windows and we'd love help with testing and feedback to make it robust.
◧◩
58. davidc+ra[view] [source] [discussion] 2026-02-02 18:53:46
>>justke+j9
Shameless plug, but you can sandbox codex cli without a container using my macOS app: https://multitui.com
◧◩
63. embiri+2b[view] [source] [discussion] 2026-02-02 18:56:23
>>justke+j9
(I work on Codex) We have a robust sandbox for macOS and Linux. Not quite yet for Windows, but working on that! Docs: https://developers.openai.com/codex/security
70. e1g+ac[view] [source] 2026-02-02 19:01:47
>>meetpa+(OP)
Wow, this is nearly an exact copy of Codex Monitor[1]: voice mode, project + threads/agents, git panel, PR button, terminal drawer, IDE integrations, local/worktree/cloud edits, archiving threads, etc.

[1] https://github.com/Dimillian/CodexMonitor

◧◩◪
98. turble+qj[view] [source] [discussion] 2026-02-02 19:38:45
>>laborc+Re
Can you explain how to use it? I’ve tried asking it to do “create 3 files using multiple sub agents” and other similar wording. It never works.

Is it in the main Codex build? There doesn’t seem to be an experiment for it.

https://github.com/openai/codex/issues/2604

◧◩◪
108. mritch+Ak[view] [source] [discussion] 2026-02-02 19:44:23
>>IMTDb+c8
yeah, I wanted a better terminal for operating many TUI agent's at once and none of these worked because they all want to own the agent.

I ended up building a terminal[0] with Tauri and xterm that works exactly how I want.

0 - screenshot: https://x.com/thisritchie/status/2016861571897606504?s=20

◧◩◪
128. strong+cr[view] [source] [discussion] 2026-02-02 20:10:10
>>raw_an+yi
The trick to reach the usage limit is to run many agents in parallel. Not that it’s an explicit goal of mine but I keep thinking of this blog post [0] and then try to get Codex to do as much for me as possible in parallel

[0]: http://theoryofconstraints.blogspot.com/2007/06/toc-stories-...

◧◩◪
156. abshkb+Pw[view] [source] [discussion] 2026-02-02 20:33:33
>>OkGoDo+Bd
We did train Codex models natively on Windows - https://openai.com/index/introducing-gpt-5-2-codex/ (and even 5.1-codex-max)
◧◩◪
176. halfli+hA[view] [source] [discussion] 2026-02-02 20:48:04
>>iamnot+hz
Deploying from Antigravity is as easy as say connecting the Firebase MCP [1] and asking it "deploy my app to firebase".

[1] https://firebase.google.com/docs/ai-assistance/mcp-server

◧◩◪
191. kzahel+0E[view] [source] [discussion] 2026-02-02 21:06:15
>>tomash+Y4
I'm also so baffled by this. I had to write my own app to be able to do seamless handoff between my laptop/desktop/phone and it works for me (https://github.com/kzahel/yepanywhere - nice web interface for claude using their SDK, MIT, E2E relay included, no tailscale required) but I'm so baffled why this isn't first priority. Why all these desktop apps?
204. hmokig+MJ[view] [source] 2026-02-02 21:32:30
>>meetpa+(OP)
I'm still waiting for the big pivotal moment in this space, I think there is a lot of potential with rethinking an IDE to be Agent first, and lots of what is out there is still lacking. (It's like we all don't know what we don't know, so we are just recycling UX around trying to solve it)

I keep coming back to my basic terminal with tmux running multiple sessions. I recently though forked this https://github.com/tiann/hapi and been loving using tailscale to expose my setup on my mobile device for convenience (plus the voice input there)

208. hamash+lK[view] [source] 2026-02-02 21:35:02
>>meetpa+(OP)
More simple and similar app: vibe-kanban

https://www.vibekanban.com/

◧◩◪
226. harikb+FO[view] [source] [discussion] 2026-02-02 21:52:33
>>rafram+VL
I have been building Desktop apps with Go + Wails[1]. I happen to know Go, but if you are ai-coding even that is not necessary.

A full fledged app, that does everything I want, is ~ 10MB. I know Tauri+Rust can get it to probably 1 MB. But it is a far cry from these Electron based apps shipping 140MB+ . My app at 10MB does a lot more, has tons of screens.

Yes, it can be vibe coded and it is especially not an excuse these days.

[1] https://wails.io/

Microsoft Teams, Outlook, Slack, Spotify? Cursor? VsCode? I have like 10 copies of Chrome in my machine!

◧◩◪◨⬒
231. measur+JP[view] [source] [discussion] 2026-02-02 21:56:55
>>bopbop+cP
There is a guy on twitter documenting his progress with moltbot/openclaw: https://x.com/Austen/status/2018371289468072219. Apparently he has already registered his bot for an LLC so he can make money w/ it.
◧◩◪
239. Areibm+oR[view] [source] [discussion] 2026-02-02 22:02:41
>>raw_an+yi
Your goal should be to run agents all the time, all in parallel. If you’re not hitting limits, you’re massively underutilizing the VC intelligence subsidy

https://hyperengineering.bottlenecklabs.com/p/the-infinite-m...

◧◩◪◨⬒⬓⬔
243. measur+GS[view] [source] [discussion] 2026-02-02 22:06:46
>>bopbop+6R
What does the bootcamp have to do w/ anything? He is using AI slop to make money, that's all that matters in a socio-economic system wherein everyone & everything must make profits to persist. Edit: found another example from coinbase: https://x.com/0xEricBrown/status/2018082458143699035.

Edit: I'm not going to keep addressing your comment if you keep editing it. You asked for an example & I found two very easily. I am certain there are many others so at this point the onus is on you to figure out what exactly it is you are actually arguing.

◧◩◪
247. Olympi+pT[view] [source] [discussion] 2026-02-02 22:09:49
>>namelo+FQ
This is another common excuse.

You don't need to use microsoft's or apple's or google's shit UI frameworks. E.g. see https://filepilot.tech/

You can just write all the rendering yourself using metal/gl/dx. if you didn't want to write the rendering yourself there are plenty of libraries like skia, flutter's renderer, nanovg, etc

◧◩◪◨
257. incr_m+0Z[view] [source] [discussion] 2026-02-02 22:29:21
>>Olympi+pT
How is File Pilot for accessibility and for all of the little niceties like native scrolling, clipboard interaction, drag and drop, and so on? My impression is that the creator is has expertly focused on most/all of these details, but I don't have Windows to test.

I insist on good UI as well, and, as a web developer, have spent many hours hand rolling web components that use <canvas>. The most complicated one is a spreadsheet/data grid component that can handle millions of rows, basically a reproduction of Google Sheets tailored to my app's needs. I insist on not bloating the front-end package with a whole graph of dependencies. I enjoy my NIH syndrome. So I know quality when I see it (File Pilot). But I also know how tedious reinventing the wheel is, and there are certain corners that I regularly cut. For example there's no way a blind user could use my spreadsheet-based web app (https://github.com/glideapps/glide-data-grid is better than me in this aspect, but there's no way I'm bringing in a million dependencies just to use someone else's attempt to reinvent the wheel and get stuck with all of their compromises).

The answer to your original question about why these billion dollar companies don't create artisanal software is pretty straightforward and bleak, I imagine. But there are a few actually good reasons not to take the artisanal path.

◧◩◪◨⬒
268. hmokig+w31[view] [source] [discussion] 2026-02-02 22:45:26
>>hollow+S01
Claude, at least, already supports LSP servers though. https://code.claude.com/docs/en/plugins-reference#lsp-server...
◧◩
280. fabian+Ja1[view] [source] [discussion] 2026-02-02 23:12:30
>>rubslo+Vn
We have been working on this, letting any coding agent define infrastructure so we can define it effortlessly: https://specific.dev. We aren't just targeting non-developers though, we think this is useful to anyone building primarily through coding agents.
◧◩◪◨
285. embedd+dd1[view] [source] [discussion] 2026-02-02 23:24:19
>>Olympi+pT
> You don't need to use microsoft's or apple's or google's shit UI frameworks. E.g. see https://filepilot.tech/

That's only for Windows though, it seems? Maybe the whole "just write all the rendering yourself using metal/gl/dx" is slightly harder than you think.

286. inerci+Wd1[view] [source] 2026-02-02 23:27:22
>>meetpa+(OP)
Something similar but for any ACP server: https://github.com/inercia/mitto
◧◩◪◨
293. resfir+7h1[view] [source] [discussion] 2026-02-02 23:42:27
>>lvl155+tL
Surely it varies a lot and everyone is in an industry bubble to some extent, but from my experience in some non-tech industries (healthcare, manufacturing), Linux workstations were nonexistent and working with the Linux servers was more a sign of an ops role. People who wrote code for a living didn't touch them directly. Last StackOverflow survey [1] puts it at something like 50% use Windows at work, 30% Mac, 20-40% Linux (breakdown of WSL and Ubuntu as categories seems confusing, maybe the raw data is better).

[1] https://survey.stackoverflow.co/2025/technology/#1-computer-...

◧◩◪◨
303. saadn9+vm1[view] [source] [discussion] 2026-02-03 00:08:07
>>mritch+Ak
looks like we both did haha: https://github.com/saadnvd1/aTerm
◧◩
339. indigo+oE1[view] [source] [discussion] 2026-02-03 02:02:48
>>mdrzn+L6
Not going to solve your exact problem but I started this project with this approach in mind

https://github.com/jgbrwn/vibebin

◧◩◪◨⬒
341. indigo+ME1[view] [source] [discussion] 2026-02-03 02:05:25
>>theLim+0k
Not going to solve your exact problem but I started this project with this approach in mind

https://github.com/jgbrwn/vibebin

◧◩
342. indigo+2F1[view] [source] [discussion] 2026-02-03 02:07:03
>>daxfoh+KO
Not going to solve your exact problem but I started this project with this approach in mind. It is exposed to the Internet though on a VPS or server but using caddy basic auth in front of the coding url

https://github.com/jgbrwn/vibebin

◧◩◪◨
350. indigo+zG1[view] [source] [discussion] 2026-02-03 02:17:56
>>kzahel+0E
This looks awesome! And incredibly polished. Exactly the approach I take to vibebin-- I may have to integrate yep anywhere into it (if that's ok) as an additional webui!

https://github.com/jgbrwn/vibebin

Although I would need it to listen on 0.0.0.0 instead of localhost because I use LXC containers so caddy on the host proxies to the container 10.x address. Hopefully yep has a startup flag for that. I saw that you can specify the port but didn't see listening address mentioned.

◧◩◪◨
357. niobe+6I1[view] [source] [discussion] 2026-02-03 02:28:21
>>ryandr+iZ
isn't that what Simon Willison's `llm` does?

edit: [link](https://github.com/simonw/llm)

◧◩◪
380. dworks+GS1[view] [source] [discussion] 2026-02-03 03:56:51
>>energy+Zx1
The stages are modelled after a kanban board. So you can have whichever stages you think are important for your LLM development workflow. These are mine:

00: Iterate on requirements with ChatGPT outside of the IDE. Save as a markdown requirements doc in the repo

01: Inside the IDE; Analysis of current codebase based on the scope of the requirements

02: Based on 00 and 01, write the implementation plan. Implement the plan

03: Verification of implementation coverage and testing

04: Implementation summary

05: Manual QA based on generated doc

06: Update global STATE.md and DECISIONS.md that documents the app, and the what and why of every requirement

Every stage has a single .md as output and after the stage is finished the doc is locked. Every stage takes the previous stages' docs as input.

I have a half-finished draft with more details and a benchmark (need to re-run it since a missing dependency interrupted the runs)

https://dilemmaworks.com/implementing-recursive-language-mod...

◧◩◪◨⬒⬓
384. deaux+UT1[view] [source] [discussion] 2026-02-03 04:08:16
>>IhateA+xT1
Open weights means that the current prices for inference of Chinese models are indicative of their cost to run because.

https://openrouter.ai/moonshotai/kimi-k2.5

It's a fantasy to believe that every single one of these 8 providers is serving at incredibly subsidized dumping prices 50% below cost and once that runs out suddenly you'll pay double for 1M of tokens for this model. It's incredibly competitive with Sonnet 4.5 for coding at 20% of the token price.

I encourage you to become more familiar with the market and stop overextrapolating purely based on rumored OpenAI numbers.

◧◩◪◨
418. rubyma+A62[view] [source] [discussion] 2026-02-03 06:08:45
>>tomber+H22
I built my Block Editor (Notion-style) in Qt C++ and QML[1].

[1] https://get-notes.com

◧◩◪◨⬒
422. jackso+K72[view] [source] [discussion] 2026-02-03 06:18:52
>>heyste+4H
Hey, Conductor founder here. Conductor is built on Anthropic's Agents SDK, which exposes most (but not all) of Claude Code's features.

https://platform.claude.com/docs/en/agent-sdk/overview

◧◩
458. prodig+Dq2[view] [source] [discussion] 2026-02-03 08:55:23
>>sunaoo+Jp2
Claude responds differently to "think", "think hard", and "think very hard". Just because it's hidden to you, doesn't mean a user doesn't have a choice.

Saying gpt-3.5-turbo is better than gpt-5.2 makes me think something you got some of them hidden motives.

https://code.claude.com/docs/en/common-workflows#use-extende...

◧◩◪◨⬒
462. ogoffa+5r2[view] [source] [discussion] 2026-02-03 08:57:43
>>joseph+YG1
We're actually working on a native open source cross-platform UI toolkit called Slint that’s trying to do exactly that. https://slint.dev
◧◩◪◨
465. ogoffa+Mr2[view] [source] [discussion] 2026-02-03 09:04:03
>>tomber+H22
Qt is still used, but I think part of the reason it is less used is that C++ isn't always the right language anymore for building GUI application.

That’s actually why we're working on Slint (https://slint.dev): It's a cross-platform native UI toolkit where the UI layer is decoupled from the application language, so you can use Rust, JavaScript, Python, etc. for the logic depending on what fits the project better.

◧◩◪◨⬒
467. Comput+3s2[view] [source] [discussion] 2026-02-03 09:06:21
>>prodig+Nr2
I was getting this in my Claude code app, it seems clear to me that they didn’t want users to do that anymore and it was deprecated. https://i.redd.it/jvemmk1wdndg1.jpeg
◧◩◪◨⬒⬓
470. prodig+fs2[view] [source] [discussion] 2026-02-03 09:08:05
>>Comput+3s2
Thx for the correction. Changed a couple weeks ago. https://decodeclaude.com/ultrathink-deprecated/
◧◩◪
475. thekna+7t2[view] [source] [discussion] 2026-02-03 09:14:48
>>gloosx+I92
React Native is able to build abstractions on top of both Android and iOS that uses native UI. Microsoft even have a package for doing a "React Native" for Windows: https://github.com/microsoft/react-native-windows

It's weird that we don't have a unified "React Native Desktop" that would build upon the react-native-windows package and add similar backends for MacOS and Linux. That way we could be building native apps while keeping the stuff developers like from React.

◧◩◪
476. Bishon+Bt2[view] [source] [discussion] 2026-02-03 09:18:08
>>ProofH+yn2
My bad then. I meant that it's "Crazy Good" as in that the free tier gave me a tremendous amount of tokens.

What I didn't realize though, is that the limit doesn't reset each 5 hours as is the case for claude. I hit the limit of the free tier about 2 hours in, and while I was expecting to be able to continue later today, it alerts me that I can continue in a week.

So my hype for the amount of tokens one gets compared to claude was a bit too eager. Hitting the limit and having to wait a week probably means that we get a comparable token amount vs the $20 claude plan. I wonder how much more I'd get when buying the $20 plus package. The pricing page doesn't make that clear (since there was no free plan before yesterday I guess): https://developers.openai.com/codex/pricing/

◧◩◪◨
485. reveri+Ny2[view] [source] [discussion] 2026-02-03 09:54:51
>>thekna+7t2
There are such implementations for React Native: https://reactnative.dev/docs/out-of-tree-platforms
◧◩◪
500. nashas+3H2[view] [source] [discussion] 2026-02-03 11:07:09
>>tptace+2c
Check out >>46829029
◧◩◪◨⬒⬓⬔
506. sunaoo+tN2[view] [source] [discussion] 2026-02-03 11:50:53
>>prodig+fs2
Nice blog, this post is interesting: https://decodeclaude.com/compaction-deep-dive/ Didn't know about Microcompaction!
◧◩◪◨⬒⬓⬔⧯
527. prodig+Z03[view] [source] [discussion] 2026-02-03 13:21:31
>>sunaoo+tN2
If you're a big context/compaction fan and want another fun fact, did you know that instead of doing regular compaction (prompting the agent to summarize the conversation in a particular way and starting the new conversation with that), Codex passes around a compressed, encrypted object that supposedly preserves the latent space of the previous conversation in the new conversation.

https://openai.com/index/unrolling-the-codex-agent-loop/ https://platform.openai.com/docs/guides/conversation-state#c...

Context management is the new frontier for these labs.

◧◩◪◨⬒
528. rubyma+B13[view] [source] [discussion] 2026-02-03 13:26:17
>>rjh29+Kq2
That's false, see QML hot reload[1].

[1] https://www.qt.io/blog/speed-up-qt-development-with-qml-hot-...

◧◩◪◨⬒⬓
530. rubyma+123[view] [source] [discussion] 2026-02-03 13:28:45
>>FooBar+lv2
This is not because of Qt - it is due to some (most) Qt developers not caring enough. I created my Qt app feel native both on macOS and Windows[1]. It did require a lot of tuning - but those are things I'll reuse across other apps.

[1] https://get-notes.com/

◧◩◪◨⬒⬓⬔⧯▣
533. rubyma+z23[view] [source] [discussion] 2026-02-03 13:31:45
>>pitche+CX2
In modern Qt you don't write UI in C++ anymore - you do that in QML. It is far simpler to create amazing pixel perfect UIs with drooling-inducing animations in QML. I wrote a blog post that talks a bit about this[1].

[1] https://rubymamistvalove.com/block-editor

◧◩◪◨⬒
565. super2+ZC3[view] [source] [discussion] 2026-02-03 16:21:47
>>eloisa+J83
Also, modern native UIs became looking garbage on desktops / laptops, where you usually want a high information density.

Just look at this TreeView in WinUI2 (w/ fluent design) vs a TreeView in the good old event viewer. It just wastes SO MUCH space!

https://f003.backblazeb2.com/file/sharexxx/ShareX/2026/02/mm...

And imo it's just so much easier to write a webapp, than fiddle with WinUI. Of course you can still build on MFC or Win32, but meh.

◧◩◪◨⬒
577. sideef+j24[view] [source] [discussion] 2026-02-03 18:01:24
>>realus+vR2
https://reactnative.dev/docs/out-of-tree-platforms says otherwise

React Native Skia allegedly runs on Linux too

◧◩◪◨⬒⬓
580. sideef+e54[view] [source] [discussion] 2026-02-03 18:12:01
>>sideef+j24
React Native Skia seems abandoned. But maybe this will make React Native on Linux viable

https://github.com/gtkx-org/gtkx

◧◩
603. krzyza+Wo5[view] [source] [discussion] 2026-02-04 00:51:49
>>samuel+T4
until then, the https://CommanderAI.app tries to fill the gap on Mac
[go to top]