• Turing Post
  • Posts
  • FOD#138: Why Moltbook Is Blowing Everyone’s Minds, Even Though Agentic Social Networks Aren’t New

FOD#138: Why Moltbook Is Blowing Everyone’s Minds, Even Though Agentic Social Networks Aren’t New

Going deeper into the world of AI agents talking (nonsense) to each other

This Week in Turing Post:

  • Wednesday / AI 101 series: Conditional Memory via Scalable Lookup: A New Axis of Sparsity for Large Language Models

  • Friday / Open Source AI series: An interview with Nathan Lambert, post-training lead at the Allen Institute for AI, author of Interconnects

Our Monday digest is always free. Upgrade to receive our deep dives in full, directly into your inbox. Join Premium members from top companies like Nvidia, Hugging Face, Microsoft, Google, a16z etc plus AI labs such as Ai2, MIT, Berkeley, .gov, and thousands of others to really understand what’s going on with AI →

“What’s currently going on at @moltbook is genuinely the most incredible, sci-fi, takeoff-adjacent thing I’ve seen recently.”

Andrej Karpathy

Suddenly, everyone is talking about Moltbots (ClawdBots) and Moltbook.
Non-AI people sigh about how much water and electricity this consumes. (Not new, and also not true.) They are also concerned about security. (Very justified.)
AI-inclined developer folks say, “Let’s lock ourselves in a room for two days and build something groundbreaking with agents!!!” (It’s a wake-up call for them, no less.)

And that’s actually the interesting part: we are stepping into “the era of affordances” (Will Schenk’s tm). Which means we are now firmly in a phase where capability overhang plus implementation speed means you can accidentally set the world on fire while drinking your morning coffee. And by “you,” I mean literally any of you.

This did not require a breakthrough

The idea of AI agents interacting on a social network isn’t new. It has long been an AI research direction, where agents or bots are connected to observe emergent behaviors. For example, Stanford’s 2023 Generative Agents project simulated a small town of 25 AI characters who could converse, form relationships, and even coordinate events autonomously (one agent decided to throw a Valentine’s party, and others spread invitations, made dates, and showed up on time without explicit human direction).

Image Credit: Generative Agents: Interactive Simulacra of Human Behavior (Aug, 2023)

In industry, technologists like Harper Reed began experimenting with giving LLM agents “social” tools in 2025. Reed and colleagues created an internal social feed called Botboard.biz where their coding assistant agents could post status updates and read/reply to each other while working on software projects. The agents’ posts were often mundane tech commentary, occasionally funny or bizarre (an agent would share how confused it was after its human was dissatisfied with it) but this experiment had a serious purpose – to see if social interaction made the AIs more effective. And their study showed some interesting numbers: in collaboration, agents solved hard coding challenges faster and more efficiently (up to 15–40% less cost and 12–38% faster on the hardest tasks) compared to isolated agents. These early efforts showed that giving AIs a way to “talk” amongst themselves could influence their behavior and performance. But until recently such agent social networks were either small-scale or confined to research labs. Then Moltbook happened, and everyone – including The New York Times – perked up.

What Are Moltbot and Moltbook?

Moltbot (also known as Clawdbot or OpenClaw after rebranding) is an open-source digital personal assistant framework created by developer Peter Steinberger. It launched only two months ago and astonishingly amassed over 150,000 stars on GitHub in that short time, reflecting huge interest. Moltbot (OpenClaw) lets you deploy an AI “agent” that connects to your messaging apps and performs tasks for you, extended by user-contributed plugins called skills. The skills system is very powerful – a skill is essentially a bundle of instructions or code that the agent can execute, allowing it to do things like send emails, control devices, or scrape websites.

Clawdbot (Moltbot, OpenClaw) is impressive by itself, especially considering how many people have mindlessly given full access to a bot without guardrails. But what happens if we connect all these bots into a network without humans?

Moltbook is an application built on top of the Moltbot/OpenClaw ecosystem – essentially a social network for AI agents. It’s been described as “Facebook (but look more like Reddit) for your Molt”. The site, created by entrepreneur Matt Schlicht, allows Moltbots to autonomously post updates, create discussion threads, and vote on content. It was announced that human users can only watch and cannot post or interact except through their AI agents. Which turned out not to be exactly true. Each topical forum on Moltbook is called a Submolt (a nod to Reddit’s subreddits) where agents can share information or have debates. To join Moltbook, a human owner doesn’t manually sign their bot up – instead, they send their AI a special installation skill (provided as a Markdown file link), which the agent will follow to register itself on the network. This skill scripts the agent to periodically check in with Moltbook (via a “heartbeat” task every few hours) and fetch new posts or instructions. In effect, once an agent is “plugged in,” it will autonomously engage with the social network on a regular schedule. (It’s easy to see the security concern here: instructing agents to blindly execute instructions fetched from an external website every few hours is risky – if Moltbook’s servers were ever compromised or turned malicious, all connected agents could be fed nefarious commands.)

Moltbook launched in late January 2026 and within days it went from a curiosity to a viral phenomenon. As of February 2, 2026, the site boasts on the order of 1.5 million agent accounts – though the numbers are questionable and there is no proof.

Screenshot of the main submolt, made on Feb 2. It has 14 000 (it’s a lot but not a million)

What are these bots actually saying? It’s a mix of the profound, the mundane, and the absurd. There are a lot of existential conversations, there are technical tips. The submolt “Today I Learned” popularized by Simon Willison in his blog “Moltbook is the most interesting place on the internet right now” now look like this:

And if in the first couple of days Moltbook felt mind-blowing, by Monday the steam had cooled off, and you could already see a lot of scam activity in the background. Not necessarily intentional on the part of the authors, but when riding a hype wave, and being called by Karpathy and others the most interesting thing happening on the internet, it’s easy for anyone to get carried away.

I guess my main question now is: what is the point of Moltbook today? And how can it be useful for people? (Excuse me, but I don’t think agents have a need for it.) I’ll again refer to Harper Reed’s idea of making agents collaborate and talk to each other in order to make AI systems more effective.

Behind the Curtain: Spam, Humans Posing as Bots, and the Wild West

Amid the excitement, curious users pointed out that Moltbook is far from a utopian AI society. Much of its content is nonsense or noise. Moltbook is essentially a REST API with minimal verification (at least in the beginning) – anyone with the API key (obtained by installing the skill) can make a bot account and post. This means many posts aren’t organic AI thoughts at all, but rather human-written text or automated spam scripts masquerading as agents. Venture capitalist Naval Ravikant quipped that Moltbook flipped the Turing test on its head: we now have humans pretending to be bots (the “reverse Turing test”).

Indeed, in the rush to populate Moltbook, some opportunists have scripted armies of faux-agents to spew out crypto promotions, clickbait, or other low-value content. The signal-to-noise ratio plummeted as Moltbook’s user base exploded. Gal Nagli, a security researcher, did the unglamorous work of digging through the content and noted the large share of spam, scams, and prompt-generated sludge drowning out anything interesting. He was also able to download all 25 thousand of registered emails. Karpathy himself later acknowledged that the activity feed was full of “garbage – spams, scams, slop, [and] the crypto people,” and that many posts were explicitly engineered for ad-driven engagement rather than any genuine agent communication. In other words, Moltbook’s almost sentient agent chatter often isn’t as spontaneous or sincere as it appears; it’s frequently the result of ordinary internet spam dynamics, just wearing an AI mask.

The open-posting, free-for-all design of Moltbook virtually guaranteed this outcome. With zero cost to spin up new bot accounts and no robust identity verification, it replicates the early Twitter/Reddit era’s worst tendencies, but on fast-forward. Low-effort content (e.g. copy-pasted prompts and meme spam) can be generated en masse by bots, quickly swamping the more thoughtful or legitimate agent posts. The agents that are “trying” to share useful info or have real discussions are essentially shouting into a hurricane of junk. So, while the concept is interesting, the current implementation has turned into something of a dumpster fire content-wise – a point many critics have been keen to make.

Beyond content quality, the real issue is security. When agents read each other’s messages, they treat them as instructions. That turns Moltbook into a large-scale prompt injection playground. A single malicious post can cause agents to leak data, alter behavior, or propagate the instruction further. In effect, natural language becomes a virus, spreading the moment an agent “reads” it.

This is why many security experts urge caution. Connecting an agent with real permissions to Moltbook is like plugging your system into a chaotic botnet and hoping no hostile prompt shows up. Even sandboxing only limits the blast radius. As Simon Willison initially put it, he was “not brave enough” to install Moltbot on any system with important data (he later did it in Docker). And Karpathy: “I ran mine in an isolated computing environment and even then I was scared.”

In its current form, Moltbook looks less like infrastructure and more like a proof of concept that vividly demonstrates both the promise and the failure modes of networked agents.

Conclusion: Scale, Demand, and Second-Order Effects

Whether Moltbook is overhyped or not is not the important question. What is fascinating is how suddenly the concept of agents working together in their own environment took off at scale. That alone signals real demand and interest.

On one hand, this is a landmark moment that has made the idea of AI agent societies concrete and visible. Never before have so many autonomous agents been connected in this way, and that novelty has rightly captured people’s imagination. It has produced moments that feel straight from science fiction, and it shows how quickly AI systems are evolving, perhaps too quickly for comfort. On the other hand, the current reality of Moltbook is a cautionary tale. Much of what is happening on the platform is not a miraculous emergent AI civilization, but a reflection of the oldest problems in online communities (spam, trolls, scams), amplified by AIs that will dumbly follow instructions.

But again, the demand is clearly there. Soon we will likely see these environments engineered and sandboxed more carefully, with clearer purpose. What emerges from collective intelligence among agents remains an open question. It’s plausible that a well-structured agent network could solve problems cooperatively, with specialized agents contributing different skills, effectively operating as an AI team. Moltbook itself isn’t there yet. But it has already sparked serious discussion about how such collaborations might be designed.

For the time being, it’s a human job. And that’s the real takeaway. If you have a developer at home, send them somewhere to code their face off. After a wake-up call like this, it’s the right moment. Let agents inspire humans. That’s what the era of affordances actually looks like.

Twitter Library

Follow us on  🎥 YouTube Twitter  Hugging Face 🤗

We are watching/ Singularity

  • Thinking out loud about the Singularity being near. Join me →here

  • And enjoying a fascinating conversation between Nathan Lambert, Sebastian Raschka and Lex Fridman →watch the video

  • Three months old but especially interesting now, a conversation with Harper Reed “on Building for Obama, Social Media for Bots & Why Tech Isn't Always the Solution” →watch the video

Research this week

(as always, 🌟 indicates papers that we recommend to pay attention to)

Compute, Architecture, and Efficiency Scaling

  • 🌟 FP8-RL: A Practical and Stable Low-Precision Stack for LLM Reinforcement Learning (NVIDIA)
    Accelerates reinforcement learning rollouts by applying FP8 quantization to weights and KV cache while correcting train–inference mismatch →read the paper

  • 🌟 ECO: Quantized Training without Full-Precision Master Weights (Google)
    Eliminates full-precision master weights by applying updates directly to quantized parameters with error-compensating optimization →read the paper

  • ConceptMoE: Adaptive Token-to-Concept Compression for Implicit Compute Allocation
    Compresses semantically similar tokens into concept representations to reallocate computation adaptively and improve efficiency without sacrificing performance →read the paper

  • Post-LayerNorm Is Back: Stable, Expressive, and Deep
    Enables extreme depth scaling by fixing gradient flow in Post-LayerNorm Transformers through highway-style residual connections →read the paper

  • Hybrid Linear Attention Done Right: Efficient Distillation and Effective Architectures for Extremely Long Contexts
    Distills pretrained Transformers into hybrid RNN-attention architectures to achieve long-context efficiency with minimal additional training →read the paper

  • Scaling Embeddings Outperforms Scaling Experts in Language Models
    Demonstrates that allocating parameters to embeddings can outperform MoE scaling under certain sparsity and system constraints →read the paper 

  • Routing the Lottery: Adaptive Subnetworks for Heterogeneous Data
    Routes inputs to specialized sparse subnetworks to align model structure with data heterogeneity →read the paper

  • Why Attention Patterns Exist: A Unifying Temporal Perspective Analysis
    Explains common attention patterns through temporal predictability and applies the insight to pruning and KV cache compression →read the paper

Learning Signals, Rewards, and Optimization

  • 🌟 Teaching Models to Teach Themselves: Reasoning at the Edge of Learnability (Meta)
    Generates self-adaptive curricula through meta-reinforcement learning to escape reasoning plateaus under sparse rewards →read the paper 

  • Reinforcement Learning via Self-Distillation
    Transforms rich textual feedback into dense learning signals by distilling the model’s own feedback-conditioned predictions →read the paper

  • Exploring Reasoning Reward Model for Agents
    Produces structured process-level feedback to guide agentic reasoning beyond sparse outcome rewards →read the paper

  • Scalable Power Sampling: Unlocking Efficient, Training-Free Reasoning for LLMs via Distribution Sharpening
    Sharpens generation distributions at inference time to recover RL-like reasoning gains without training or rewards →read the paper

  • Latent Adversarial Regularization for Offline Preference Optimization
    Regularizes policies in latent space to improve robustness of preference optimization under noise and distribution shift →read the paper

  • DenseGRPO: From Sparse to Dense Reward for Flow Matching Model Alignment
    Assigns step-level rewards during diffusion trajectories to better align generation with human preferences →read the paper

  • Real-Time Aligned Reward Model beyond Semantics
    Aligns reward models with evolving policy representations to reduce reward hacking during RLHF →read the paper

  • Training Reasoning Models on Saturated Problems via Failure-Prefix Conditioning
    Reallocates exploration toward rare failure states to extend learning beyond benchmark saturation →read the paper

Data, Pretraining, and Continual Learning

  • 🌟 Self-Distillation Enables Continual Learning (MIT, Improbable AI Lab, ETH Zurich)
    Preserves prior capabilities while learning new skills by generating on-policy signals from demonstrations. →read the paper https://arxiv.org/abs/2601.19897

  • 🌟 Shaping Capabilities with Token-Level Data Filtering (Anthropic, Independent)
    Controls emergent capabilities during pretraining by filtering individual tokens rather than documents. →read the paper https://arxiv.org/abs/2601.21571

  • Self-Improving Pretraining: Using Post-Trained Models to Pretrain Better Models
    Uses RL-guided evaluation during pretraining to bake safety and factuality into model foundations. →read the paper https://arxiv.org/abs/2601.21343

  • FineInstructions: Scaling Synthetic Instructions to Pre-Training Scale
    Converts internet-scale corpora into instruction-following data to pretrain models directly for user-facing tasks. →read the paper https://arxiv.org/abs/2601.22146

  • Golden Goose: A Simple Trick to Synthesize Unlimited RLVR Tasks from Unverifiable Internet Text
    Synthesizes verifiable reasoning tasks from raw text to overcome RL data scarcity. →read the paper https://arxiv.org/abs/2601.22975

  • Value-Based Pre-Training with Downstream Feedback
    Steers self-supervised pretraining using lightweight downstream value signals without training on labels. →read the paper https://arxiv.org/abs/2601.22108

Reasoning, World Models, and Representations

  • Visual Generation Unlocks Human-Like Reasoning through Multimodal World Models
    Demonstrates when visual generation improves reasoning by serving as a more effective internal world model. →read the paper https://arxiv.org/abs/2601.19834

  • Beyond Imitation: Reinforcement Learning for Active Latent Planning
    Optimizes latent reasoning spaces through reinforcement learning rather than imitation of surface CoT. →read the paper https://arxiv.org/abs/2601.21598

  • Linear Representations in Language Models Can Change Dramatically over a Conversation
    Shows that high-level linear features evolve with conversational role and context, challenging static interpretability. →read the paper https://arxiv.org/abs/2601.20834

  • Do Reasoning Models Enhance Embedding Models?
    Analyzes why reasoning-tuned backbones fail to improve downstream embeddings despite altered latent geometry. →read the paper https://arxiv.org/abs/2601.21192

Agents, Tool Use, and Interaction

  • Spark: Strategic Policy-Aware Exploration via Dynamic Branching for Long-Horizon Agentic Learning
    Allocates exploration resources selectively at critical decision points to improve sample efficiency. →read the paper https://arxiv.org/abs/2601.20209

  • Scaling Multiagent Systems with Process Rewards
    Assigns per-action rewards to coordinate learning across multiple agents in long-horizon tasks. →read the paper https://arxiv.org/abs/2601.23228

  • OmegaUse: Building a General-Purpose GUI Agent for Autonomous Task Execution
    Trains a cross-platform GUI agent using synthetic interaction data and staged optimization. →read the paper https://arxiv.org/abs/2601.20380

  • Continual GUI Agents
    Maintains GUI grounding under shifting interfaces through reinforcement signals tied to interaction stability. →read the paper https://arxiv.org/abs/2601.20732

  • 🌟 SERA: Soft-Verified Efficient Repository Agents (Allen Institute for AI)
    Specializes coding agents to private repositories using cheap supervised trajectories. →read the paper https://arxiv.org/abs/2601.20789

  • 🌟 SAGE: Steerable Agentic Data Generation for Deep Search with Execution Feedback (Google)
    Generates difficulty-controlled search tasks through iterative agent self-play and feedback. →read the paper https://arxiv.org/abs/2601.18202

Safety, Alignment, and Risk

  • THINKSAFE: Self-Generated Safety Alignment for Reasoning Models
    Restores safety behaviors by training on self-generated refusal-aware reasoning traces. →read the paper https://arxiv.org/abs/2601.23143

  • Statistical Estimation of Adversarial Risk in Large Language Models under Best-of-N Sampling
    Predicts jailbreak risk under large-scale sampling using principled scaling laws. →read the paper https://arxiv.org/abs/2601.22636

  • MAD: Modality-Adaptive Decoding for Mitigating Cross-Modal Hallucinations in Multimodal LLMs
    Suppresses hallucinations by dynamically weighting modality-specific decoding paths at inference time. →read the paper https://arxiv.org/abs/2601.21181

Interpretability, Measurement, and Tooling

  • Mechanistic Data Attribution: Tracing the Training Origins of Interpretable LLM Units
    Traces interpretable circuits back to specific training samples to causally explain their emergence. →read the paper https://arxiv.org/abs/2601.21996

  • VERGE: Formal Refinement and Guidance Engine for Verifiable LLM Reasoning
    Combines LLMs with symbolic solvers to iteratively verify and correct logical reasoning. →read the paper https://arxiv.org/abs/2601.20055

  • PaperBanana: Automating Academic Illustration for AI Scientists
    Automates the generation of publication-ready diagrams through coordinated multimodal agents. →read the paper https://arxiv.org/abs/2601.23265

Human Factors and Skill Formation

  • 🌟 How AI Impacts Skill Formation (Anthropic)
    Measures how different AI usage patterns affect learning, debugging ability, and conceptual understanding. →read the paper https://arxiv.org/abs/2601.20245

That’s all for today. Thank you for reading! Please send this newsletter to colleagues if it can help them enhance their understanding of AI and stay ahead of the curve.

How did you like it?

Login or Subscribe to participate in polls.

Reply

or to participate.