George Jor

The Guide to AI-Friendly Systems

As the web evolves beyond traditional SEO, AI agents are increasingly reading and interpreting online content. Ensuring that both the content and the underlying systems are sufficiently structured has become a critical consideration in modern software development.

  1. Structured & Semantic Data
    To make web content more AI-friendly, embed structured data using JSON-LD with widely adopted schemas from Schema.org such as FAQ, Product, Service, and Article. Combine this with proper semantic HTML markup (including appropriate use of H1–H6 headings and elements like <section>, <article>, and <nav>), to enhance machine readability and reduce the likelihood of misinterpretation by AI agents. Adopting predictable, standardized formats for data exchange, such as JSON, CSV, or XML, further enables AI systems to parse and process information reliably and efficiently.

  2. Context & Relationships
    The Model Context Protocol (MCP) is an open-source standard that provides a universal interface for connecting AI applications to external systems, functioning much like a USB-C port for AI context sharing. Implementing an llms.txt file can further streamline discovery and interaction for large language models and AI crawlers by clearly signaling intended usage patterns, supported endpoints, and schemas. This approach helps prevent information overload, improves overall readability, and minimizes the risk of hallucinations.

  3. Permissions & Ethical Boundaries
    Access control can be explicitly managed through robots.txt and emerging AI-specific directives such as llms.txt. These mechanisms allow developers to clearly define which content may be used for training, which is restricted to inference only, and which areas are entirely prohibited, thereby upholding ethical standards and ensuring compliance with legal requirements.

  4. Observability & Error Resilience
    Maintaining transparency in AI processes requires capturing raw prompt payloads, monitoring token usage for optimization, implementing span-level tracing with OpenTelemetry, and logging short-term memory states to facilitate debugging. Incorporating automatic fallback mechanisms, such as switching to alternative models during downtime, further enhances system reliability and error resilience.

  5. Scalability & Compatibility
    Conventional rate limiting designed for human users often proves inadequate for AI agents that operate at much higher volumes. Infrastructure should therefore be designed to support high concurrency, batch processing, and low latency. Compatibility with established AI toolchains such as LangChain and prevalent API designs like those from OpenAI will encourage seamless integration by third-party agents.