A former Meta executive who helped navigate one of Facebook’s most turbulent trust-and-safety periods is now building what he believes is missing from modern content moderation: a way to make policies executable, not just written.
Brett Levenson, previously part of Facebook’s business integrity team during and after the Cambridge Analytica fallout, has launched Moonbounce, a startup focused on what it calls “policy as code.” The company has now raised $12 million in fresh funding to scale its technology, which aims to translate moderation rules directly into machine-readable systems that AI can enforce in real time.
At its core, Moonbounce is betting that the biggest weakness in content moderation today is not detection capability, but the delay between writing rules and actually enforcing them across AI systems.
Levenson’s central argument is straightforward.
Content moderation policies are written by humans, but enforced by machines. And the translation between the two is slow, inconsistent, and often incomplete.
Moonbounce is designed to eliminate that gap.
Instead of relying on fragmented rule lists and manual updates, the platform builds a control layer where policies can be written, tested, versioned, and deployed like software. In practical terms, that means a rule change about hate speech, political ads, or AI-generated content could be pushed directly into moderation systems without weeks or months of engineering work.
The system acts as an orchestration layer between policy teams and AI models. It ensures that classifiers, filters, and large language model guardrails are aligned with the latest rules, rather than operating on outdated interpretations.
This approach positions Moonbounce not as another detection tool, but as infrastructure for enforcement consistency.
The timing of Moonbounce’s launch is not accidental.
Meta, along with other major platforms, is accelerating its shift toward AI-driven moderation. The company has already signaled plans to rely more heavily on automated systems to detect harmful content across Facebook and Instagram, while reducing dependence on large third-party human moderation teams.
AI systems are now tasked with identifying everything from explicit content and scams to terrorism-related material and misinformation. Internal testing at Meta suggests these systems can achieve high accuracy in clear-cut cases, allowing platforms to scale moderation far beyond what human teams could handle.
At the same time, Meta has begun loosening certain moderation policies. Moves such as reducing third-party fact-checking and shifting toward community-driven moderation models reflect a broader recalibration of how content is governed.
This creates a structural tension.
Platforms are automating more decisions while simultaneously adjusting the rules those systems are supposed to enforce. Without a reliable way to update AI behavior in sync with policy changes, inconsistencies become inevitable.
Moonbounce is positioning itself as the solution to that exact problem.
To understand what Moonbounce is trying to fix, it helps to look at how moderation works today.
Large platforms like Facebook operate hybrid systems where automation handles the majority of decisions, and humans intervene only in complex or borderline cases.
The pipeline typically includes several layers:
Most clear violations are removed automatically. Human moderators are reserved for appeals, edge cases, and policy interpretation.
But the system has a known weakness.
When policies change, updating all these layers takes time. Engineers must reinterpret rules, adjust models, and validate outcomes. During that window, enforcement can drift away from stated policy.
Moonbounce is designed to compress that delay.

Moonbounce’s approach reframes moderation as a software problem.
Policies are treated as structured inputs that can be versioned, tested, and deployed. Instead of informal guidelines, they become explicit constraints that AI systems must follow.
This has several implications.
First, it makes enforcement more predictable. Platforms can test how a rule change affects moderation outcomes before deploying it globally.
Second, it improves auditability. Regulators and internal teams can verify whether AI systems are enforcing rules as intended, rather than relying on opaque model behavior.
Third, it reduces operational lag. Updates can be pushed quickly across systems without manual reconfiguration.
In theory, this creates a tighter feedback loop between policy decisions and real-world enforcement.
The rise of generative AI has fundamentally changed the scale of the moderation challenge.
Platforms are no longer dealing only with user-generated text and images. They are now handling AI-generated content that can be produced in massive volumes, often blurring the line between authentic and synthetic media.
At the same time, moderation systems are not just removing content. They are shaping what users see through ranking and recommendation algorithms.
This expands the role of moderation from enforcement to influence.
As a result, the stakes are higher.
Errors in moderation are no longer isolated incidents. They can affect public discourse, political narratives, and user trust at scale.
Tools like Moonbounce are emerging in response to this shift, offering a way to align AI behavior with rapidly evolving rules.
Moonbounce is entering a space where demand is increasing but solutions remain fragmented.
Most existing tools focus on detection, improving the ability to identify harmful content. Fewer address the governance layer, where policies are defined and translated into system behavior.
This gap is becoming more visible as platforms face regulatory scrutiny across multiple regions, each with its own legal standards for content moderation.
A system that can encode policies as testable, auditable rules could appeal not only to social media companies but also to enterprises deploying AI in customer-facing environments.
The question is whether platforms will adopt an external control layer or continue building internal solutions.
Moonbounce’s $12 million funding round signals early confidence, but the real test lies ahead.
The company will need to prove that its system can integrate with complex moderation pipelines and deliver measurable improvements in consistency and speed.
Observers will also be watching how regulators respond. As governments push for greater transparency in AI systems, tools that make enforcement auditable could gain traction.
At the same time, platforms like Meta are continuing to evolve their moderation strategies, balancing automation, user-driven systems, and policy changes.
Whether Moonbounce becomes a core part of that ecosystem or remains a niche infrastructure layer will depend on how effectively it can bridge the gap it has identified.
Moonbounce is not trying to build better AI moderators.
It is trying to make sure those moderators follow the rules.
In an era where content moderation is increasingly automated, the ability to turn policy into code may become as important as the models themselves.
Be the first to post comment!
Cognichip, a semiconductor AI startup, has secured $60 milli...
by Vivek Gupta | 4 days ago
Salesforce has announced one of the most significant updates...
by Vivek Gupta | 5 days ago
Bluesky has introduced a new standalone AI application calle...
by Vivek Gupta | 1 week ago
ByteDance has begun integrating its latest AI video generati...
by Vivek Gupta | 1 week ago
A startup called Littlebird has raised $11 million to build...
by Vivek Gupta | 1 week ago
Spotify is making a clear strategic shift: the future of use...
by Vivek Gupta | 2 weeks ago