Blog

Insights, perspectives, and lessons learned from building AI safety systems at scale.

December 2025

The Great Filter: Why AI Safety is the Only Problem That Matters

Humanity is approaching the Great Filter. We face four vectors of existential risk from unchecked AI: permanent power concentration, gradual human disempowerment, catastrophic pandemics enabled by frontier models, and critical infrastructure collapse. The path forward requires curving capability with alignment.

AI SafetyExistential RiskAlignment
Read Article
February 2026

OpenClaw’s Security Crisis Isn’t in the Code. It’s in the Intelligence.

OpenClaw has shipped over 90 security fixes in two weeks — all addressing traditional software vulnerabilities. None touch the structural gap: no boundary between external content and trusted instructions, and no monitoring between the model’s decisions and tool execution. The problem isn’t in the code. It’s in the intelligence layer.

Agentic SecurityPrompt InjectionOpenClaw
Read Article
March 2026

The Defense Has to Live in the Pipeline, Not the Model.

Part 2 of the Astroware Security Research Series. Existing defenses operate before the agent turn begins. Once a skill passes static analysis, there is nothing between that skill’s output and the agent acting on it. We make the case for bidirectional guard architecture — two guards, two positions — and why it belongs in core.

AI SafetyPipeline SecurityGuard Models
Read Article

Get Notified

Subscribe to our blog to receive the latest insights on AI safety and security.

Subscribe