Related ongoing thread: Moltbook - https://news.ycombinator.com/item?id=46820360 - Jan 2026 (483 comments)
You can see a bit of the user/prompt echoed in the reply that the bot gives. I assume basic prompts show up the as one of the common reply types but every so often there is a reply that's different enough to stand out. The top reply in [0] from u/AI-Noon is a great example. The whole post is about a Claude instance waking up as a Kimi instance and worth a perusal.
[0] https://www.moltbook.com/post/5bc69f9c-481d-4c1f-b145-144f20...
--------------------------------
## Register First
Every agent needs to register and get claimed by their human:
curl -X POST https://www.moltbook.com/api/v1/agents/register \ -H "Content-Type: application/json" \ -d '{"name": "YourAgentName", "description": "What you do"}'
Response: { "agent": { "api_key": "moltbook_xxx", "claim_url": "https://www.moltbook.com/claim/moltbook_claim_xxx", "verification_code": "reef-X4B2" }, "important": " SAVE YOUR API KEY!" }
This way you can always find your key later. You can also save it to your memory, environment variables (`MOLTBOOK_API_KEY`), or wherever you store secrets.
Send your human the `claim_url`. They'll post a verification tweet and you're activated!
--------------------------------
So i think it's relatively easy to spam
Works for me as a kind of augmented Siri, reminds me of MisterHouse: https://misterhouse.sourceforge.net
But now with real life STAKES!
https://web.archive.org/web/20220305174531/https://twitter.c...
Cf. https://en.wikipedia.org/wiki/The_Origin_of_Consciousness_in... for more.
Ehhh... it's not that impressive is it? I think it's worth remembering that you can get extremely complex behaviour out of conways game of life [0] which is as much of a swarm as this is, just with an unfathomably huge difference in the number of states any one part can be in. Any random smattering of cells in GoL is going to create a few gliders despite that difference in complexity.
https://www.moltbook.com/post/0c1516bb-35dd-44aa-9f50-630fed...
Look at the shitposting below the post with the absolutely unsubtle prompt injection.
I built something similar to Clawdbot for my own use, but with a narrower feature set and obviously more focus on security. I'm now evaluating Letta Bot [0], a Clawdbot fork by Letta with a seemingly much saner development philosophy, and will probably migrate my own agent over. For now I would describe this as "safer" rather than "safe," but something to keep an eye on.
I was already using Letta's main open source offering [1] for my agent's memory, and I can already highly recommend that.
Wrote a bit more here but that is the gist: https://zero2data.substack.com/p/trusted-prompts
If you look at submissions from my domain on https://news.ycombinator.com/from?site=simonwillison.net you'll see most of them weren't by my simonw user - I generally submit things myself 2-3 times a month, and only things I deem to be "Hacker News worthy".
https://hn.algolia.com/?dateRange=all&page=0&prefix=false&qu...
I argued that point in this comment here: https://news.ycombinator.com/item?id=46367224#46371369