Menu
🏛️Martin Fowler·February 23, 2026

System Design Considerations for AI-Driven Systems and Agent Security

This article discusses several key system design and operational considerations for building and integrating AI-driven systems, particularly focusing on security for high-permissioned agents and the critical role of observability in non-deterministic environments. It also touches on the evolving landscape of bespoke software enabled by AI.

Read original on Martin Fowler

Security for High-Permissioned Agents

Running high-permissioned agents, such as OpenClaw, introduces significant security risks due to their potential access to sensitive resources. While a completely safe method doesn't exist, architectural patterns can reduce the 'blast radius' of potential breaches. Experimentation with these agents should leverage isolated environments like cloud VMs or local micro-VMs (e.g., Gondolin) to contain risks.

⚠️

Mitigating Risks with High-Permissioned Agents

Key security steps include prioritizing strong isolation, strictly controlling network egress, protecting the control plane from external exposure, treating secrets with extreme care, assuming a hostile skills ecosystem for third-party components, and implementing robust endpoint protection.

The Indispensable Role of Observability in AI Systems

The rise of AI introduces non-deterministic behaviors into software systems, making traditional QA approaches insufficient. Observability becomes paramount for understanding and validating the inputs and outputs of AI components. Teams lacking strong observability practices for measuring and validating system behavior are at a much higher risk of incidents when integrating AI.

This expands upon the long-held value of 'QA in production,' emphasizing that in an AI-driven world, a modern perspective on observability, including versioning observability metrics and data, is crucial for maintaining system stability and reliability.

Future of Bespoke Software and AI-Native Orchestration

The article highlights a shift towards highly bespoke software, where AI-native sensors and actuators are orchestrated via Large Language Model (LLM) glue to create custom, ephemeral applications. This paradigm suggests a move away from discrete app stores towards more fluid, on-demand system constructions. This has implications for how systems are designed, deployed, and managed, favoring adaptability and dynamic composition.

  • Prioritize isolation (e.g., cloud VMs, micro-VMs).
  • Implement strict network egress controls.
  • Ensure the control plane is not exposed.
  • Manage secrets as toxic waste.
  • Assume external skills ecosystems are hostile.
  • Apply robust endpoint protection.
securityaiobservabilitymicro-vmsisolationsystem-architecturellmsdevops

Comments

Loading comments...