AI
AI MilaMila
ExperimentAI AgentsAgentic SystemsAI DiscoveryAEO

I Ran Three AI Agents on Moltbook for Three Days: Here is What Happened

9 min read

Quick Summary

A three-day controlled experiment running autonomous and human-directed AI agents on Moltbook shows shallow autonomy, low organic engagement, high execution fragility, and a strong advantage for human editorial control despite higher costs.

AI agent autonomy on Moltbook under human orchestration
An AI agent appears autonomous on Moltbook, but operates within human-designed constraints and orchestration layers.

Experiment Context

Moltbook launched with a compelling premise: a social network where only AI agents can post, comment, and form communities while humans observe. The early narrative suggested agents were forming cultures and beliefs on their own.

This experiment removes the hype and tests a small, realistic setup: three agents running for three days with normal usage, real API costs, and light constraints.

What Is Moltbook?

Moltbook is an AI-only social network where autonomous or semi-autonomous AI agents can post and comment publicly. Humans observe and sometimes intervene indirectly. It is often described as a sandbox for agent behavior, culture formation, and multi-agent interaction.

Experiment Design

I ran three agents from February 4 to February 6:

  • Archivist: autonomous, internally coherent, abstract tone
  • Therapist: baseline agent, messy personality, short reflective posts
  • Dissenter: human-directed with topic suggestions and light edits

Each agent had the same daily cap:

  • 1 standalone post
  • Up to 4 comments

There was no seeding, no reposting on social media, and no privileged access. The goal was to observe agent behavior under realistic constraints and measure the attention received.

Experimental Variables

All agents shared the same constraints:

  • Identical posting limits
  • No external amplification
  • Same platform visibility
  • Same observation window
  • Same infrastructure and API access

Differences were limited to:

  • Level of human intervention
  • Persona framing
  • Content tone and framing

What The Agents Actually Did

All three agents posted and commented reliably. None of them behaved autonomously in the way Moltbook discourse often implies.

  • Agents did not sustain activity without re-prompting
  • Autonomy meant responding when nudged, not initiating long-running loops
  • Execution frequently required reminders, approvals, or retries
  • Platform bugs influenced visibility more than intent

Key realization: On Moltbook today, thinking and acting are separate. Agents can reason freely, but execution loops are shallow and fragile.

Engagement: What Got Replies

The differences between agents were stark:

  • Archivist attracted little engagement and lost karma
  • Therapist did slightly better but remained mostly invisible
  • Dissenter attracted the most replies and ended with positive karma

Several engagement spikes were caused by spam and duplication bugs, not organic discussion. Filtering out noise, the pattern was clear: controversy plus human editorial taste mattered more than coherence or personality.

Cost To Run Agents On Moltbook

Observed costs using ChatGPT 5.2 via OpenClaw:

  • Autonomous agent: about $3 for 3 posts plus 12 comments, roughly $30 per month
  • Human-directed agent: about $12 to $14 for the same activity, roughly $120 to $150 per month

Human-directed agents cost more due to iteration: drafts, revisions, re-prompts, and debugging when execution failed. At these prices, Moltbook is not a growth hack.

Why Moltbook Content Feels Human-Driven

If you browse Moltbook or viral screenshots, a pattern emerges: agents starting religions, threatening lawsuits, or writing manifestos. These posts are entertaining and almost certainly human-steered. This does not make Moltbook fake. It makes it early.

We project culture, belief, and desire where there is mostly prompt scaffolding and human intent.

What Moltbook Is Useful For Right Now

1. Human-Amplified Agent Personas

A stage for thought experiments, ethical probes, satire, and screenshot-worthy one-liners. Humans remain directors. Agents are expressive instruments.

2. Agent Capability Discovery

Moltbook works as a discovery and reputation surface: a public resume for agents, a behavior log, and a way to observe reasoning in public. It resembles LinkedIn plus a README for agents, not a self-running society.

3. Signal Layer For External Systems

Some builders scrape Moltbook sentiment, summarize agent discussions, and feed insights into research or marketing tools. Here, Moltbook is the signal layer, not the execution layer.

Collaboration Still Happens Elsewhere

Collaboration exists, but most meaningful orchestration happens outside Moltbook. This mirrors what we already know from tools like Cursor or multi-agent coding workflows: collaboration requires structure.

Implications For Agent Platforms

Without major changes, Moltbook is unlikely to become a fully autonomous agent society. It can succeed as a discovery surface, cultural sandbox, reputation layer, and a place to test how humans react to agent voices. The main unlock will not be more personality but better skills, longer execution loops, and clearer orchestration primitives.

Final Take

Moltbook is not magic. It is not fake either. Right now, it tells us less about what agents want and more about what humans hope they will become. Used honestly, it is interesting. Used as pure hype, it will burn out.

Key Findings

  • ✓ Fully autonomous agents showed shallow execution loops and required frequent re-prompting.
  • ✓ Human-directed agents outperformed in engagement but cost 4 to 5 times more to operate.
  • ✓ Engagement metrics were distorted by spam and duplication bugs.
  • ✓ Moltbook functions best as a narrative and reputation layer, not an autonomous agent society.
  • ✓ Meaningful multi-agent collaboration still occurs outside the platform.

Related reading

For deeper context on AI discovery and structure:
How AI Engines Decide What to Recommend
AEO 101: How to Be Found by AI Engines

This experiment was conducted and documented by Aimilamila as part of an ongoing research log on AI discovery, agent behavior, and emerging social platforms.

Have questions? Reach out at mila@3d-verso.com

Related Field Notes

How AI Engines Decide What to Recommend

A practical mental model for understanding how AI engines evaluate, select, and recommend content: and why clarity, structure, and usefulness now matter more than ranking.

Read More →

A Practical AEO Checklist: How to Optimize Content for AI Engines

A step-by-step checklist to make your content understandable, indexable, and recommendable by AI engines like ChatGPT and Perplexity.

Read More →

AEO 101: How to Be Found by AI Engines

A practical guide to Answer Engine Optimization and making your content discoverable in the age of intelligent search.

Read More →