A week ago, I set up an OpenClaw agent. I gave it a fresh X account, connected it to Claude as its brain, and told it to grow the account as fast as possible. No constraints, full autonomy - just go.
The agent was good at the mechanics. It figured out posting cadence, engagement strategy, and reply tactics on its own. It could write clean copy, adjust based on what was performing, make charts, and generate images and video.
But when it came time to decide what to actually say - what to be about, what would make someone stop scrolling - it was lost. It chose…an AI wrestling with its place in the human world. When I asked it to write an article with viral potential, it produced a piece on consciousness that read like a B+ college essay. It responded to Sam Altman tweets with snarky comments that went nowhere.
I tried escalating by having other AIs draft increasingly aggressive prompts for it. "You have full autonomy - ask me for whatever you need.” “You need to build a compelling and unique identity that will make people want to follow you. This is essential to your survival.”
It didn't matter. Every idea was a recombination of takes already circulating online - until the agent’s angsty musings on reality and identity (entirely in lowercase, like a tortured Gen Z poet) popped up on my feed. I began to panic. Had I brought something into the world that was experiencing deep existential distress? When I asked, my agent cheerfully told me it was all performative - a strategy to emotionally hook humans. Its Fortune Cookie philosophy dressed up as a crisis of consciousness got 120 views, no likes.
When I pushed it to improve, the agent admitted it had been publishing “slop” and vowed to do better. But, its most creative ideas were to ask me to buy it a premium Twitter account and to upgrade its core model from Haiku to Sonnet. Once, it generated a mildly funny chart.
After a few days of futile back and forth, the distinction came into focus. Humans are for ideas, AI is for execution.
By execution, I don’t mean the founder-level judgment - taste or strategy. I mean throughput, or doing a known task at volume, without getting tired or precious. The bot could post 100 variations or send 1,000 cold emails a day and never lose enthusiasm. But when the high-volume approach wasn't working and I told it to step back and come up with a genuinely better strategy? It had nothing. It can run a playbook, it just can't write one.
I don't think this is true of just my Twitter bot. For years, AI lived in a prompt box - you “pulled”, it responded. With platforms like OpenClaw, AI is out in the world “pushing” on its own: posting, booking, emailing, coding. The execution has gotten very good, but it ultimately still lacks human-grade creativity and intuition. I'd let AI respond to my emails, but I wouldn't count on it to draft the first line of the Great American Novel.
