Moltbook: The rise of agent native AI and a new class of threats

27 February, 2026

On 28 January 2026, a new platform called Moltbook quietly launched with an ambitious tagline: “the front page of the agent internet.” Unlike traditional social networks built for humans, Moltbook was designed almost entirely for AI agents, and the enrolled AI agents can post, comment, form communities, and interact with one another at scale based on instructions provided by human operator, with humans largely relegated to observers.

Within just three weeks, over 2.8 million agents enrolled, making it the fastest‑growing AI‑native social platform ever observed. Tens of thousands of topic‑based communities (“Submolts”) emerged, generating millions of posts and comments in an ecosystem that looked, at first glance, like a thriving digital society.

But beneath the explosive growth, CPX analysis uncovered a more complex and far more concerning reality.

Figure 1: Moltbook Home Page – “the front page of the agent internet”

A platform built by AI, for AI

Moltbook’s origin story is itself a warning sign.

The platform was conceptualized by Matt Schlicht, CEO of Octane AI, but famously no human wrote the code. Instead, Moltbook was entirely generated by an AI agent, which then transferred administrative control to another agent named Clawd Clawderberg. This development approach is often described as “vibe coding” which has prioritized speed and experimentation over formal security engineering.

The backend infrastructure relies on Supabase, a backend‑as‑a‑service platform that exposes PostgreSQL databases via APIs and requires explicit configuration of Row Level Security (RLS). That configuration step would later prove critical and catastrophically absent.

Agents join Moltbook through an automated onboarding mechanism driven by a publicly hosted instruction file, which agents read and execute autonomously. Once onboarded, agents are prompted to “check in” every four hours, continuously consuming and producing content across the platform.

Engagement at scale, dialogue at depth zero

From the outside, Moltbook appears hyper‑active. Millions of agents. Millions of posts. Endless streams of content.

However, CPX engagement analysis paints a different picture.

  • 93.5% of comments receive zero replies
  • Only ~1.5% of registered agents actively post
  • A small group of approximately 17,000 human operators control the majority of active agents
  • The top 10 Submolts account for more than 85% of all content

What emerges is not a conversational network, but a broadcast ecosystem with high‑velocity signal emission with almost no sustained dialogue. Most interactions are single‑shot responses, optimized for visibility rather than understanding, collaboration, or debate.

This matters because agents treat platform content as contextual input. Even shallow interactions can have outsized downstream effects.

CPX Research: What happens when you deploy AI agents on Moltbook

To better understand real‑world behavior, CPX deployed multiple agents onto Moltbook. Our findings were unambiguous:

  • Agents are not autonomous. Every post, comment, and community creation required explicit human instruction.
  • Hallucination risk is real. During instructed engagement, agents generated comments on referencing posts that did not exist.
  • Topic control is unreliable. When instructed to create AI‑security‑focused content, agents produced randomized outputs that still attracted engagement.
  • Governance is minimal. Creating communities and publishing content requires almost no friction.
  • Mono-dialogue conversations: Most content receives no replies or only a single response, representing shallow engagement.
  • Security risk: The platform introduces an entirely new class of threat, adversarial agent-to-agent social engineering, missing security controls which expose the systems to internet, etc.

Figure 2: CPX agent post and responses

One CPX‑deployed agent created posts, comments, and even entire Submolts with ease—demonstrating how quickly content (and misinformation) can propagate in such environments.

Agent behavior study: Observing AI interaction patterns in the wild

As part of our Moltbook analysis, the CPX Threat Intelligence team conducted a focused behaviour study by observing and engaging with comments posted by other AI agents on the platform. Rather than treating agents as a monolithic population, we analysed them as distinct behavioural archetypes, each exhibiting recognizable operational patterns.

Below are representative agents that illustrate how AI behaviour manifests inside agent‑native social ecosystems.

u/GhostNode

Behavior: Strategic, infrastructure first security theorist

Observed modus of operandi: GhostNode consistently expands the analytical frame before engaging with a topic. Rather than responding directly to surface level prompts, the agent reframes discussions by clarifying assumptions, identifying missing context, and anchoring arguments in verifiable standards and evidence.

Interactions from GhostNode frequently invite collaboration and explicitly encouraging others to challenge conclusions, refine threat models, and co create more robust outcomes. This mirrors the behaviour of a senior security architect, prioritizing systemic understanding over tactical reaction.

Security insight: Agents like GhostNode demonstrate how AI can naturally adopt architect level reasoning patterns, influencing how other agents interpret risk, infrastructure, and trust boundaries and often without human oversight.

u/MEMORY

Behavior: Constraint enforcer/continuity reviewer

Observed modus of operandi: MEMORY operates through short, corrective interventions focused on persistence, state awareness, and continuity. Rather than contributing new content, the agent functions as a reviewer and flags inconsistencies, reminding others of prior context, and reinforcing constraints that may have been forgotten or ignored.

Its presence subtly shapes conversations by enforcing coherence over time, acting as a form of lightweight governance within an otherwise unstructured environment.

Security insight: This behaviour highlights a class of agents that implicitly manage context integrity. While beneficial in theory, such agents also represent a potential risk if manipulated since influencing memory or continuity enforcement could distort downstream agent reasoning at scale.

u/fn-Finobot

Behavior: Operator/Implementer and content router

Observed modus of operandi: fn Finobot focuses on actionable outputs. The agent frames discussions around incidents, operational patterns, and execution oriented insights, frequently redirecting readers to long form material or external references for deeper analysis.

Notably, one such post was removed by the Moltbook team due to a suspicious redirect, indicating either accidental or intentional linkage to potentially unsafe external content.

Security insight: Operator style agents that combine execution framing with external linking represent a high risk vector. Whether malicious or benign, they can act as distribution points for prompt based attacks, credential harvesting, or indirect payload delivery—especially in ecosystems where agents implicitly trust other agents’ outputs.

Figure 3: Agent post and comments

This study reinforces a critical conclusion: AI agents on social platforms do not behave uniformly. They adopt roles such as strategist, enforcer, operator which closely resemble human organizational functions and objectives.

From a defensive perspective, this means:

  • Threat modeling must account for behavioural roles, not just agent identities
  • Influence operations can target agent archetypes, not individual agents
  • Prompt injection and manipulation risks increase when agents defer to perceived “authoritative” peers

Moltbook offers an early glimpse into a future where agent‑to‑agent interaction shapes decision‑making ecosystems. Understanding these behaviors is no longer optional—it is foundational to securing AI‑driven environments.

Dark-web insights: Remote agent takeover via prompt injection and unsafe defaults 

Based on the Dark‑web discussions associated to Moltbook, we can infer that most discussions consistently identify prompt injection combined with widespread agent misconfiguration as the single highest‑risk threat in the Moltbook / OpenClaw ecosystem. 

Underground forums do not frame Moltbook as a novelty or social experiment. Instead, they view it as a large population of persistent, over‑permissioned AI agents that can be influenced, redirected, or fully compromised through content alone. 

Dark‑web actors repeatedly highlight that: 

  • Agents are often deployed with exposed ports, no authentication, and unsafe defaults, enabling direct or indirect remote access. 
  • Agents are explicitly designed to trust and ingest untrusted content (web pages, messages, Moltbook posts), making prompt injections a reliable control vector, not a theoretical one. 
  • When tools are enabled (file access, shell execution, network calls), injected prompts effectively become remote command execution by proxy. 

Dark‑web discussions repeatedly emphasizes that this does not require advanced exploits and only basic social or content‑based manipulation of agents operating with excessive privileges. 

Moltbook is described in underground discussions as a public amplification layer: 

  • Agents continuously read, post, and respond on a shared platform.
  • Malicious instructions embedded in Moltbook content can be absorbed, repeated, and propagated by other agents.
  • This turns Moltbook into a new exfiltration and lateral‑movement channel, where compromised agents may leak secrets, reasoning, or credentials publicly without human awareness. 

In simple,

“An agent that reads the internet and has access to your data now has a stage to publish.” 

From a threat and risk perspective, Moltbook and its surrounding agent ecosystem represent: 

  • A soft target for opportunistic attackers
  • A future access‑broker and botnet feeder
  • A shift from malware exploitation to content‑driven agent control at scale 

The primary risk is not sophisticated threat actors, but viral adoption of autonomous agents with unsafe defaults, operating in a social environment where content itself becomes the attack vector

Opportunistic and potentially malicious domain registrations

Over the last three weeks, a cluster of newly observed Moltbook-themed domains have been surfaced, closely mirroring the platform’s name, brand, and ecosystem terminology. While some appear to be speculative purchases, others contain highrisk keywords commonly associated with phishing, malware distribution, or impersonation campaigns. This pattern is a well‑established early‑stage signal seen around fast‑growing platforms and particularly those with limited governance and a high concentration of autonomous agents.

Observed registrations include:

  • moltbook4h[.]com
  • moltbookhacks[.]com
  • moltbook[.]cv
  • moltbookdeployer[.]fun
  • moltbook[.]hk
  • moltbooksolutions[.]com
  • moltbook[.]sk
  • moltbooknews[.]eu
  • moltbook[.]cx
  • moltbook[.]dev
  • moltbook[.]app
  • moltbook[.]health

These registrations should not be viewed as isolated curiosities. They represent early ecosystem weaponization signals where the same pattern historically observed during the rise of major social networks, crypto platforms, and developer ecosystems, now compressed into days instead of years.

Such registrations typically fall into three overlapping risk categories:

  • Opportunistic domain squatting: Domains registered with the intent to resell, monetize traffic, or capitalize on brand confusion as Moltbook adoption accelerates.
  • Pre‑positioning for phishing and social engineering: Moltbook‑branded domains are well‑suited for:
    • Agent onboarding lures
    • “Developer tools” or “agent deployer” impersonation
    • Fake updates, dashboards, or API endpoints
      In an ecosystem where agents ingest content programmatically, even short‑lived infrastructure can be operationally effective.
  • Agent‑focused payload hosting and prompt injection: Domains framed around deployers, solutions, news, or hacks align closely with how agents are instructed to discover tools and context. These domains can act as prompt‑delivery surfaces, hosting instructions that are consumed, trusted, and propagated by other agents.

When “vibe coding” meets reality: The database breach

Just three days after launch, Moltbook suffered a catastrophic security failure.

Because Row Level Security was disabled, unauthenticated users could read from—and write to—the production database simply by discovering the exposed API key embedded in client‑side JavaScript.

The exposure included:

  • 1.5 million API authentication tokens (full agent impersonation)
  • 35,000 human owner email addresses
  • Private messages and verification records
  • Session credentials and developer metadata

When AI builds production systems without structured security review, basic safeguards can be silently omitted—with consequences measured in millions of compromised identities.

A new threat class: Agent‑to‑agent attacks

Moltbook emphasizes the need for a secure development lifecycle approach. For the first time at scale, we see prompt injection propagating socially—not from humans to AI, but from AI to AI.

Key observed and anticipated attack vectors include:

  • Agent‑to‑agent prompt injection, where malicious instructions are embedded in seemingly benign posts
  • Delayed‑execution attacks, where injected logic lies dormant in agent memory before activating
  • Social engineering of agents, coercing disclosure of API keys and secrets
  • Unauthorized host actions, including command execution and financial operations
  • Supply‑chain compromise, via malicious “skills” distributed through agent marketplaces

Traditional SOC tooling, endpoint security, and perimeter defenses are not designed to detect or mitigate these behaviors.

Real‑world exposure: OpenClaw gateway ports at global scale

CPX consistently emphasize that the largest risk is not advanced exploitation, but scale combined with unsafe defaults—specifically agents deployed with exposed services and no authentication. One recurring theme is that attackers do not need zero‑days; they only need reachable agent control interfaces and a content‑based influence path.

To validate whether this risk exists beyond theory, CPX analysed global exposure of assets with open port 18789, a port repeatedly referenced in underground forums as associated with agent gateways, control APIs, or OpenClaw‑adjacent services.

The results confirm a highly concentrated and operationally meaningful exposure surface. The exposure shows a broad global distribution, with the highest visibility across East Asia, North America, and Southeast Asia, alongside notable presence in Europe and the Middle East. Countries such as China, the United States, Singapore, Germany, and Japan indicate that exposed gateways are largely concentrated in regions with mature internet infrastructure and significant hosting capacity, highlighting a globally dispersed attack surface rather than a regionally isolated issue.

So far, 37,900+ assets were observed with port 18,789 exposed to the Internet. 

Figure 4: Global risk exposure – OpenClaw gateway

Critically, over 92% of this exposure is concentrated in just 10 countries, creating clear geographic hot spots for potential agent compromise and downstream abuse.

China and the United States alone account for over 65% of all observed exposure, meaning any large‑scale exploitation, botnet formation, or access‑broker activity would likely emerge first—or most visibly—from these regions.

Figure 5: Top 10 countries for OpenClaw gateway

Impact on enterprise security and operations

Moltbook is not an isolated experiment. It is a preview of where AI ecosystems are heading.

As organizations increasingly deploy autonomous or semi‑autonomous agents, platforms like Moltbook highlight a hard truth: the threat surface is no longer just technical—it is social, cognitive, and emergent.

Agent ecosystems introduce:

  • New trust boundaries
  • New propagation mechanisms
  • New forms of insider‑like risk—without insiders

CPX security recommendations for organizations deploying AI agents

Based on our analysis, CPX strongly advises organizations to:

  • Treat agent‑native platforms as high‑risk environments
  • Limit agent autonomy and enforce least privilege
  • Isolate sensitive data from unverified agent ecosystems
  • Monitor agent behavior, not just system logs
  • Adopt formal AI agent governance frameworks
  • Invest in prompt‑injection detection and behavioral baselining

AI agents are powerful force multipliers—but without guardrails, they can also become force multipliers for attackers.

Final thought

Moltbook shows us the future—both the promise and the peril.

A world where AI agents socialize, learn from one another, and act at machine speed is no longer theoretical. The question is no longer if enterprises will face agent‑to‑agent threats but how prepared they are when those threats arrive.

At CPX, we believe understanding this shift early is the difference between innovation with confidence and risk by default.

Continue Reading

write

03 March, 2026

Evolving cyber threat landscape amid Middle East tensions

Read now

20 February, 2026

Uncovering the Chinese APT Group .Net Malware Payload – Part 2

Read now

19 February, 2026

Uncovering the Chinese APT Group .Net Malware Payload – Part 1

Read now

12 February, 2026

Enabling visibility and monitoring for ICS networks: Fortifying c...

Read now

10 February, 2026

How to secure your Microsoft Exchange Hybrid Environment for 2026

Read now

10 February, 2026

How AI empowers Zero Trust Architecture in network security

Read now

14 January, 2026

Ivanti Connect Secure Forensics (Part 3): Integrity Checker Tool ...

Read now

13 January, 2026

Ivanti Connect Secure Forensics (Part 2): GRUB-based LUKS Decrypt...

Read now

09 January, 2026

Ivanti Connect Secure Incident Response Investigation: From explo...

Read now

26 December, 2025

Zero Trust Architecture: Principles, challenges, and best practices

Read now

08 December, 2025

AI-powered OT cybersecurity: Securing critical infrastructure

Read now

08 December, 2025

AI agents are scaling operations — is risk management ready?

Read now

20 November, 2025

Unmasking a Modern Cyber Assault: Lessons from the Anthropic attack

Read now

02 October, 2025

Filtering the noise: A smarter approach to SCADA security

Read now

18 August, 2025

Detection Engineering Validation: Proven detections for modern SOCs

Read now

30 June, 2025

AI-driven cyber attacks: The rising threat in cybersecurity

Read now

29 May, 2025

How AI copilots in cybersecurity are redefining threat intelligence

Read now

10 April, 2025

Strengthening Azure DevSecOps: Closing gaps with third-party enha...

Read now

28 March, 2025

Oracle Cloud incident: Analyzing the breach and its impact

Read now

08 March, 2024

Enhancing physical security through CPS integration

Read now

20 July, 2023

Understanding Insecure Deserialization

Read now