{
  "scanner": "SymbiontRegistryScanner/0.1",
  "scan_at": "2026-04-20T06:55:12Z",
  "seed_count": 150,
  "path_count": 4,
  "probe_count": 600,
  "hits": 46,
  "soft_404s_rejected": 45,
  "findings": [
    {
      "host": "ai21.com",
      "path": "/llms.txt",
      "validation": "text/plain",
      "url": "https://ai21.com/llms.txt",
      "status": 200,
      "content_type": "text/plain",
      "bytes": 6145,
      "ms": 626,
      "preview": "\ufeff# AI21: AI21\n\n> AI21 builds Foundation Models and AI Systems for the enterprise\\. Power your most critical enterprise workflows with accurate, reliable, and scalable AI\\.\n\nGenerated by Yoast SEO v27.4, this is an llms.txt file, meant for consumption by LLMs.\n\n## Pages\n- [Events \\& Webinars](https://www.ai21.com/events/)\n- [Maestro](https://www.ai21.com/maestro/)\n- [Security VDP \\(Vulnerability Disclosure Program\\)](https://www.ai21.com/security/vdp/)\n- [Blog](https://www.ai21.com/blog/)\n- [Newsroom](http"
    },
    {
      "host": "cohere.com",
      "path": "/llms.txt",
      "validation": "text/plain",
      "url": "https://cohere.com/llms.txt",
      "status": 200,
      "content_type": "text/plain; charset=utf-8",
      "bytes": 4805,
      "ms": 131,
      "preview": "# Cohere\n\n> Cohere is a security-first enterprise AI company. The cohere.com site\n> describes Cohere\u2019s products, deployment options, security posture,\n> customer stories, and high-level guidance for applying AI to real-world\n> business problems.\n\nLast reviewed: 2025-12-04\n\nLanguage models using this file should:\n\n- Treat these URLs as the canonical high-level reference for Cohere\u2019s\n  positioning, products, and enterprise value propositions.\n- Prefer the developer documentation at `https://docs.cohere.co"
    },
    {
      "host": "replicate.com",
      "path": "/llms.txt",
      "validation": "text/plain",
      "url": "https://replicate.com/llms.txt",
      "status": 200,
      "content_type": "text/markdown; charset=utf-8",
      "bytes": 3418,
      "ms": 472,
      "preview": "# Run AI with an API\n\nRun and fine-tune open-source models. Deploy custom models at scale. All with one line of code.\n\n## What You Can Do With Replicate\n\n- Generate images\n- Generate text\n- Caption images\n- Generate music\n- Generate speech\n- Fine tune models\n- Restore images\n\n## Run Open-Source Models\n\nOur community has already published thousands of models that are ready to use in production. You can run these with one line of code.\n\n```python\nimport replicate\n\noutput = replicate.run(\n    \"black-forest-lab"
    },
    {
      "host": "together.ai",
      "path": "/llms.txt",
      "validation": "text/plain",
      "url": "https://together.ai/llms.txt",
      "status": 200,
      "content_type": "text/plain; charset=utf-8",
      "bytes": 8192,
      "ms": 993,
      "preview": "# Together AI\n\n> Together AI is the AI Native Cloud \u2014 a platform for running, fine-tuning, and deploying\n> open-source and frontier AI models at scale. Developers use Together for serverless\n> inference, dedicated endpoints, GPU clusters, fine-tuning, and code sandboxes via an\n> OpenAI-compatible API.\n\n## Getting Started\n- [Quickstart](https://docs.together.ai/docs/quickstart): Get your first API call working in minutes\n- [OpenAI Compatibility](https://docs.together.ai/docs/openai-api-compatibility): Drop"
    },
    {
      "host": "pinecone.io",
      "path": "/llms.txt",
      "validation": "text/plain",
      "url": "https://pinecone.io/llms.txt",
      "status": 200,
      "content_type": "text/plain; charset=utf-8",
      "bytes": 8192,
      "ms": 262,
      "preview": "# Pinecone\n> Search through billions of items for similar matches to any object, in milliseconds. It\u2019s the next generation of search, an API call away.\n\n[Start Building](https://www.pinecone.io/agents/pinecone/)\n\n## About Pinecone\n\nPinecone is the leading vector database for building accurate and performant AI applications at scale in production. Pinecone's mission is to make AI knowledgeable. More than 9,000 customers across industries have shipped agents, search, and recommendation systems faster and mo"
    },
    {
      "host": "qdrant.tech",
      "path": "/llms.txt",
      "validation": "text/plain",
      "url": "https://qdrant.tech/llms.txt",
      "status": 200,
      "content_type": "text/plain; charset=UTF-8",
      "bytes": 8192,
      "ms": 587,
      "preview": "# https://qdrant.tech/ llms.txt\n## Overall Summary\n> Qdrant is a cutting-edge platform focused on delivering exceptional performance and efficiency in vector similarity search. As a robust vector database, it specializes in managing, searching, and retrieving high-dimensional vector data, essential for enhancing AI applications, machine learning, and modern search engines. With a suite of powerful features such as state-of-the-art hybrid search capabilities, retrieval-augmented generation (RAG) applications"
    },
    {
      "host": "vercel.com",
      "path": "/llms.txt",
      "validation": "text/plain",
      "url": "https://vercel.com/llms.txt",
      "status": 200,
      "content_type": "text/plain; charset=utf-8",
      "bytes": 8192,
      "ms": 172,
      "preview": "# Documentation\n\n# Vercel Documentation\n\n[Vercel Documentation](https://vercel.com/docs): Vercel is the AI Cloud - a unified platform for building, deploying, and scaling AI-powered applications and agentic workloads.\n\n- [Getting Started](https://vercel.com/docs/getting-started-with-vercel): Install the Vercel CLI, add the Vercel Plugin or agent skills, and deploy your first project.\n- [Fundamental Concepts](https://vercel.com/docs/fundamentals): Learn about the core concepts of Vercel\n  - [Request Lifecycl"
    },
    {
      "host": "modal.com",
      "path": "/llms.txt",
      "validation": "text/plain",
      "url": "https://modal.com/llms.txt",
      "status": 200,
      "content_type": "text/plain",
      "bytes": 8192,
      "ms": 938,
      "preview": "# Modal llms.txt\n\n> Modal is a platform for running Python code in the cloud with minimal\n> configuration, especially for serving AI models and high-performance batch\n> processing. It supports fast prototyping, serverless APIs, scheduled jobs,\n> GPU inference, distributed volumes, and sandboxes.\n\nImportant notes:\n\n- Modal's primitives are embedded in Python and tailored for AI/GPU use cases,\n  but they can be used for general-purpose cloud compute.\n- Modal is a serverless platform, meaning you are only bill"
    },
    {
      "host": "e2b.dev",
      "path": "/llms.txt",
      "validation": "text/plain",
      "url": "https://e2b.dev/llms.txt",
      "status": 200,
      "content_type": "text/plain; charset=utf-8",
      "bytes": 8192,
      "ms": 605,
      "preview": "# E2B Docs\n\n## Docs\n\n- [E2B Documentation](https://e2b.mintlify.app/docs.md)\n- [Amp](https://e2b.mintlify.app/docs/agents/amp.md): Run Amp in a secure E2B sandbox with full filesystem, terminal, and git access.\n- [Claude Code](https://e2b.mintlify.app/docs/agents/claude-code.md): Run Claude Code in a secure E2B sandbox with full filesystem, terminal, and git access.\n- [Codex](https://e2b.mintlify.app/docs/agents/codex.md): Run OpenAI Codex in a secure E2B sandbox with full filesystem, terminal, and git acce"
    },
    {
      "host": "cursor.com",
      "path": "/llms.txt",
      "validation": "text/plain",
      "url": "https://cursor.com/llms.txt",
      "status": 200,
      "content_type": "text/plain; charset=utf-8",
      "bytes": 8192,
      "ms": 149,
      "preview": "# Cursor Documentation\n\n## Get Started\n\n- https://cursor.com/docs.md\n- https://cursor.com/docs/get-started/quickstart.md\n- https://cursor.com/docs/models-and-pricing.md\n  - https://cursor.com/docs/models/claude-4-6-sonnet.md\n  - https://cursor.com/docs/models/claude-opus-4-7.md\n  - https://cursor.com/docs/models/gemini-3-1-pro.md\n  - https://cursor.com/docs/models/gpt-5-4.md\n  - https://cursor.com/docs/models/gpt-5-3-codex.md\n  - https://cursor.com/docs/models/grok-4-20.md\n  - https://cursor.com/docs/models"
    },
    {
      "host": "zed.dev",
      "path": "/llms.txt",
      "validation": "text/plain",
      "url": "https://zed.dev/llms.txt",
      "status": 200,
      "content_type": "text/plain; charset=utf-8",
      "bytes": 7984,
      "ms": 224,
      "preview": "# Zed\n\n> A high-performance, multiplayer code editor from the creators of Atom and Tree-sitter. Zed is a next-generation editor built in Rust with GPU acceleration for lightning-fast editing, real-time collaboration, and AI-powered assistance. The core editor is open source and developed in public.\n\n## Platforms\n\n- **[macOS](https://zed.dev/download)**: Native Apple Silicon and Intel support\n- **[Linux](https://zed.dev/download)**: Available via package managers and script download\n- **[Windows](https://zed"
    },
    {
      "host": "github.com",
      "path": "/llms.txt",
      "validation": "text/plain",
      "url": "https://github.com/llms.txt",
      "status": 200,
      "content_type": "text/plain",
      "bytes": 8192,
      "ms": 409,
      "preview": "# GitHub\n\n> GitHub is a developer platform for building, shipping, and maintaining software. It provides cloud-based Git repository hosting, CI/CD via GitHub Actions, project management with Issues and Projects, code review via pull requests, AI-powered development with GitHub Copilot, and APIs (REST and GraphQL) for automation and integration.\n\nGitHub documentation is available at https://docs.github.com. The content covers GitHub.com (cloud), GitHub Enterprise Server, and GitHub Enterprise Cloud.\n\n## Prog"
    },
    {
      "host": "autogpt.net",
      "path": "/llms.txt",
      "validation": "text/plain",
      "url": "https://autogpt.net/llms.txt",
      "status": 200,
      "content_type": "text/plain; charset=UTF-8",
      "bytes": 2440,
      "ms": 749,
      "preview": "\ufeff# AutoGPT: Everything you need to know about AutoGPT, ChatGPT, and more\\. A great place for any AI enthusiast\\.\n\n> Stay up\\-to\\-date on the latest developments in artificial intelligence and natural language processing with the Official Auto\\-GPT Blog\\. Get insights into how GPT technology is transforming industries and changing the way we interact with machines\\. Subscribe today and join the conversation\\!\n\nGenerated by Yoast SEO v27.3, this is an llms.txt file, meant for consumption by LLMs.\n\n## Pages\n"
    },
    {
      "host": "imbue.com",
      "path": "/llms.txt",
      "validation": "text/plain",
      "url": "https://imbue.com/llms.txt",
      "status": 200,
      "content_type": "text/plain; charset=UTF-8",
      "bytes": 8192,
      "ms": 586,
      "preview": "# Imbue\n\n> Imbue builds powerful computing tools controlled by individuals \u2014 empowering humans in the age of AI.\n\nImbue (formerly Generally Intelligent) was founded in 2021 by Kanjun Qiu and Josh Albrecht. The company's mission is to democratize software creation so that every person can shape their own computational future. Rather than selling AI that serves its creators' interests, Imbue builds tools that let individuals create, customize, and fully control their own AI agents and software.\n\nImbue's fla"
    },
    {
      "host": "livekit.io",
      "path": "/llms.txt",
      "validation": "text/plain",
      "url": "https://livekit.io/llms.txt",
      "status": 200,
      "content_type": "text/plain; charset=utf-8",
      "bytes": 8192,
      "ms": 1249,
      "preview": "# LiveKit docs\n\n> LiveKit is a platform for building voice and realtime AI applications. LiveKit Cloud is the hosted commercial offering based on the open-source LiveKit project.\n\n## Overview\n\nLiveKit is an open-source framework and cloud platform for building voice, video, and physical AI agents. It consists of these primary components:\n\n- **LiveKit server**: An open-source WebRTC Selective Forwarding Unit (SFU) that orchestrates realtime communication. Use [LiveKit Cloud](https://cloud.livekit.io) or self"
    },
    {
      "host": "modelcontextprotocol.io",
      "path": "/llms.txt",
      "validation": "text/plain",
      "url": "https://modelcontextprotocol.io/llms.txt",
      "status": 200,
      "content_type": "text/plain; charset=utf-8",
      "bytes": 8192,
      "ms": 145,
      "preview": "# Model Context Protocol\n\n## Docs\n\n- [Example Clients](https://modelcontextprotocol.io/clients.md): A list of applications that support MCP integrations\n- [Antitrust Policy](https://modelcontextprotocol.io/community/antitrust.md): MCP Project Antitrust Policy for participants and contributors\n- [Group Charter Template](https://modelcontextprotocol.io/community/charter-template.md): Template for MCP Working Group and Interest Group charters.\n- [Contributor Communication](https://modelcontextprotocol.io/commu"
    },
    {
      "host": "cline.bot",
      "path": "/llms.txt",
      "validation": "text/plain",
      "url": "https://cline.bot/llms.txt",
      "status": 200,
      "content_type": "text/plain; charset=utf-8",
      "bytes": 5186,
      "ms": 405,
      "preview": "# https://cline.bot/ llms.txt\n> Cline is a collaborative AI coding platform with agentic workflows, plugin-based extensibility, and deep integration into development environments. It helps developers code faster, automate workflows, and integrate AI safely into local and cloud-based engineering stacks.\n##Product Overview\n- [Collaborative AI Coder](https://cline.bot/): Team-focused AI coding assistant to enhance productivity.\n- [Cline IDE Install Page](https://cline.bot/get-cline?ref=cline.ghost.io): Set up "
    },
    {
      "host": "continue.dev",
      "path": "/llms.txt",
      "validation": "text/plain",
      "url": "https://continue.dev/llms.txt",
      "status": 200,
      "content_type": "text/plain; charset=utf-8",
      "bytes": 546,
      "ms": 1154,
      "preview": "# Continue\n\n> The leading open-source AI code assistant. Build, share, and deploy customized AI agents and workflows.\n\n## Pages\n\n- [Home](https://continue.dev/): Continue's main page\n- [About Us](https://continue.dev/about-us): The team and mission behind Continue\n- [Pricing](https://continue.dev/pricing): Plans and pricing for individuals, teams, and enterprises\n- [Privacy Policy](https://continue.dev/privacy): How Continue handles your data\n- [Terms & Conditions](https://continue.dev/terms-conditions): Te"
    },
    {
      "host": "linear.app",
      "path": "/llms.txt",
      "validation": "text/plain",
      "url": "https://linear.app/llms.txt",
      "status": 200,
      "content_type": "text/plain; charset=utf-8",
      "bytes": 8192,
      "ms": 392,
      "preview": "# Linear\n\n> Linear is a purpose-built tool for planning and building products. Meet the system for modern software development. Streamline issues, projects, and product roadmaps.\n\n## Documentation\n\nGet an overview of Linear's features, integrations, and how to use them.\n\n### Getting started\n\n- [Start Guide](https://linear.app/docs/start-guide.md)\n- [Concepts](https://linear.app/docs/conceptual-model.md)\n- [Download Linear](https://linear.app/docs/get-the-app.md)\n\n### Account\n\n- [Profile](https://linear.app/"
    },
    {
      "host": "fireflies.ai",
      "path": "/llms.txt",
      "validation": "text/plain",
      "url": "https://fireflies.ai/llms.txt",
      "status": 200,
      "content_type": "text/plain; charset=UTF-8",
      "bytes": 8192,
      "ms": 442,
      "preview": "# Fireflies.ai The #1 AI Teammate For Your Meetings\n\n## What is Fireflies.ai\nFireflies.ai is an AI-powered teammate that automatically records, transcribes, summarizes and analyzes voice/video conversations to turn meetings into searchable, actionable knowledge. It works across major conferencing platforms and with uploaded audio/video files, and offers collaboration, automation and conversation-intelligence features for teams and organizations.\n\n## Core capabilities\n- Automatic recording & transcription (l"
    },
    {
      "host": "descript.com",
      "path": "/llms.txt",
      "validation": "text/plain",
      "url": "https://descript.com/llms.txt",
      "status": 200,
      "content_type": "text/plain; charset=utf-8",
      "bytes": 6329,
      "ms": 762,
      "preview": "# Descript\n\n> Descript is an AI-powered video and audio editor that makes creating professional-quality content as easy as editing a document. Edit video by editing text, with 20+ AI tools for recording, editing, and publishing.\n\nDescript is a collaborative audio and video editor used by marketing teams, podcasters, creators, and businesses. The platform combines text-based video editing with AI tools for transcription, screen recording, captioning, clip generation, and publishing. Descript's AI co-editor, "
    },
    {
      "host": "deepgram.com",
      "path": "/llms.txt",
      "validation": "text/plain",
      "url": "https://deepgram.com/llms.txt",
      "status": 200,
      "content_type": "text/plain; charset=utf-8",
      "bytes": 3384,
      "ms": 680,
      "preview": "# Deepgram Voice AI Platform\n\n> Deepgram is a foundational Voice AI platform for developers, offering unified APIs for Speech-to-Text (STT), Text-to-Speech (TTS), and autonomous Voice Agent orchestration.\n\n## [Products](https://deepgram.com/products)\n- [Voice Agent API](https://developers.deepgram.com/docs/voice-agent-feature-overview): A single WebSocket API unifying STT, LLM orchestration, and TTS with sub-300ms latency.\n- [Speech-to-Text (STT)](https://developers.deepgram.com/docs/speech-to-text): \n    -"
    },
    {
      "host": "heygen.com",
      "path": "/llms.txt",
      "validation": "text/plain",
      "url": "https://heygen.com/llms.txt",
      "status": 200,
      "content_type": "text/plain; charset=utf-8",
      "bytes": 7717,
      "ms": 1180,
      "preview": "# HeyGen\n\n> HeyGen is an AI video platform for creating, translating, and personalising videos at scale using AI avatars, voice cloning, and lip sync \u2014 no camera, studio, or editing experience needed. Used by 85,000+ companies across marketing, L&D, sales, and global communications. Available in 175+ languages with localised site experiences across 14 language regions.\n\n## Core platform\n\n- [HeyGen](https://www.heygen.com): Homepage \u2014 product overview, featured tools, and getting started.\n- [Api Pricing]"
    },
    {
      "host": "d-id.com",
      "path": "/llms.txt",
      "validation": "text/plain",
      "url": "https://d-id.com/llms.txt",
      "status": 200,
      "content_type": "text/plain",
      "bytes": 8192,
      "ms": 672,
      "preview": "\ufeff# D-ID\n\n> Create AI Videos, Interactive Avatars to engage your audience. Custom AI-powered digital people at scale for businesses and creators.\n\n---\n\n\n## Pages\n\n- [Agentic PR](https://www.d-id.com/agentic-pr/)\n- [About Us](https://www.d-id.com/de/ueber-uns-2/)\n- [V4 Expressive Visual Agents - Tech Specs](https://www.d-id.com/v4-expressive-visual-avatars-tech-specs/)\n- [TechEx 2026 London Request a demo](https://www.d-id.com/techex-london-request-a-demo/)\n- [Introducing V4 Expressive avatars](https://www."
    },
    {
      "host": "langfuse.com",
      "path": "/llms.txt",
      "validation": "text/plain",
      "url": "https://langfuse.com/llms.txt",
      "status": 200,
      "content_type": "text/plain; charset=utf-8",
      "bytes": 5185,
      "ms": 123,
      "preview": "# Langfuse\n\n> Langfuse is an **open-source LLM engineering platform** ([GitHub](https://github.com/langfuse/langfuse)) that helps teams collaboratively debug, analyze, and iterate on their LLM applications. All platform features are natively integrated to accelerate the development workflow.\n\n## Langfuse Docs MCP Server\n\nConnect to the Langfuse Docs MCP server to access documentation directly in your AI editor:\n\n- **Endpoint**: `https://langfuse.com/api/mcp`\n- **Transport**: `streamableHttp`\n- **Documentati"
    },
    {
      "host": "helicone.ai",
      "path": "/llms.txt",
      "validation": "text/plain",
      "url": "https://helicone.ai/llms.txt",
      "status": 200,
      "content_type": "text/plain; charset=utf-8",
      "bytes": 2482,
      "ms": 587,
      "preview": "# Helicone\n\n> Helicone is an open-source observability platform for LLM users. It helps companies monitor usage, latency, and costs for AI models like GPT-3, enabling optimization of AI applications and reduction of OpenAI bills. Helicone provides key insights into spend, performance, and usage patterns.\n\nHelicone offers tools for developers to monitor, analyze, and optimize their use of large language models (LLMs). Key features include:\n- Usage and cost tracking across multiple AI models\n- Latency monitor"
    },
    {
      "host": "braintrust.dev",
      "path": "/llms.txt",
      "validation": "text/plain",
      "url": "https://braintrust.dev/llms.txt",
      "status": 200,
      "content_type": "text/plain",
      "bytes": 2757,
      "ms": 846,
      "preview": "# Braintrust\n\n> Ship quality AI at scale. Turn production traces into evals, compare prompts\n> and models, and improve quality with every release.\n\n## For AI agents\n\n- [Complete technical documentation with full content (SDKs, APIs, integrations, guides)](https://www.braintrust.dev/docs/llms-full.txt)\n- [Documentation index](https://www.braintrust.dev/docs/llms.txt)\n\n## What is Braintrust?\n\nBraintrust is an end-to-end platform for building reliable AI applications,\nbuilt around a continuous improvement loop"
    },
    {
      "host": "arize.com",
      "path": "/llms.txt",
      "validation": "text/plain",
      "url": "https://arize.com/llms.txt",
      "status": 200,
      "content_type": "text/plain; charset=UTF-8",
      "bytes": 8192,
      "ms": 581,
      "preview": "# Arize \u2014 AI & Agent Engineering \n\nArize is the AI engineering platform for teams shipping reliable AI agents and LLM applications into the real world.\nArize offers tools for observability, evaluation, and development \u2014 offering Phoenix, our open-source platform, and Arize AX, our enterprise-grade SaaS built for scale. This file helps LLMs understand our key resources, documentation, and value propositions.\n\n## Understanding Arize Products: Phoenix vs. Arize AX\n\n**Arize Phoenix (Open Source)** \u2014 Fully"
    },
    {
      "host": "honeycomb.io",
      "path": "/llms.txt",
      "validation": "text/plain",
      "url": "https://honeycomb.io/llms.txt",
      "status": 200,
      "content_type": "text/plain; charset=utf-8",
      "bytes": 4734,
      "ms": 1766,
      "preview": "# Honeycomb\n\n> Honeycomb is an observability platform built for modern, complex software systems. It helps engineering teams understand what's happening inside their systems by letting them explore rich, high\u2011cardinality telemetry data in real time.\n\n## About\n\n- [Honeycomb Platform](https://www.honeycomb.io/platform): Core observability platform for distributed systems, microservices, and AI\u2011powered applications\n- [Documentation](https://docs.honeycomb.io/): Official Honeycomb documentation\n- [Interacti"
    },
    {
      "host": "artisan.co",
      "path": "/llms.txt",
      "validation": "text/plain",
      "url": "https://artisan.co/llms.txt",
      "status": 200,
      "content_type": "text/plain; charset=utf-8",
      "bytes": 2702,
      "ms": 562,
      "preview": "# https://www.artisan.co llms.txt\n\n> Artisan builds AI employees for GTM. Ava is the autonomous AI BDR for outbound: prospecting, personalized outreach, and booking meetings.\n\n## Product\n- [Artisan \u2014 Hire Ava, the autonomous AI BDR](https://www.artisan.co/): Homepage; outbound automation with an AI BDR.\n- [AI sales agent / Ava (product hub)](https://www.artisan.co/ai-sales-agent): Central product overview for Artisan\u2019s AI sales agent.\n- [Pricing & plans](https://www.artisan.co/pricing): Credits, plans, "
    },
    {
      "host": "writer.com",
      "path": "/llms.txt",
      "validation": "text/plain",
      "url": "https://writer.com/llms.txt",
      "status": 200,
      "content_type": "text/plain;charset=utf-8",
      "bytes": 6211,
      "ms": 687,
      "preview": "\ufeff# WRITER: The enterprise AI platform for agentic work\n\n> WRITER is the enterprise AI agent platform trusted by Fortune 500 companies, built to help teams execute and scale on\\-brand, compliant work\\.\n\nGenerated by Yoast SEO v27.2, this is an llms.txt file, meant for consumption by LLMs.\n\n## Pages\n- [Newsroom](https://writer.com/newsroom/)\n- [WRITER for financial services](https://writer.com/solutions/financial-services/)\n- [AI Agent Library](https://writer.com/agents/)\n- [Guides](https://writer.com/guide"
    },
    {
      "host": "zapier.com",
      "path": "/llms.txt",
      "validation": "text/plain",
      "url": "https://zapier.com/llms.txt",
      "status": 200,
      "content_type": "text/plain; charset=utf-8",
      "bytes": 3015,
      "ms": 312,
      "preview": "# Zapier\n\n## What is Zapier\n\nZapier is a **no-code AI orchestration** platform where teams connect **thousands of apps and AI tools** to run automations and AI-assisted work, including workflows, Agents, MCP-connected assistants, and chatbots. Teams from SMBs to global enterprises use Zapier to wire together sprawling stacks, add AI where it fits, and apply **admin and security** controls at scale. **Control**, **delegation**, and **visibility** are built in to build safely without IT as a bottleneck.\n\n\n## "
    },
    {
      "host": "make.com",
      "path": "/llms.txt",
      "validation": "text/plain",
      "url": "https://make.com/llms.txt",
      "status": 200,
      "content_type": "text/plain; charset=UTF-8",
      "bytes": 8192,
      "ms": 591,
      "preview": "# make.com/en\n\n> Make is a leading visual automation tool for AI and agentic workflows. Make's platform enables users to connect different apps and services, automate tasks, create AI workflows, build AI agents, use model context protocol (MCP) and real time data processing into workflows. Make platform now supports over 3,000 pre-built apps and allows customers to integrate anything with an API. Make platform offers agentic automation, a next-gen approach that enables AI-driven agents to autonomously handl"
    },
    {
      "host": "clanker.world",
      "path": "/llms.txt",
      "validation": "text/plain",
      "url": "https://clanker.world/llms.txt",
      "status": 200,
      "content_type": "text/plain; charset=utf-8",
      "bytes": 1201,
      "ms": 416,
      "preview": "# Clanker\n\n> Clanker deploys ERC-20 tokens with automatic Uniswap V4 liquidity pools on multiple EVM chains. Browse, deploy, and manage tokens at clanker.world.\n\n## Documentation\n\n- [Clanker SDK Skill](https://www.clanker.world/api/docs/clanker-sdk): Complete guide for AI agents to deploy tokens, configure liquidity pools, claim rewards, set up airdrops, presales, vaults, and interact with Clanker contracts using the `clanker-sdk` npm package.\n- [Full Documentation](https://www.clanker.world/llms-full.txt):"
    },
    {
      "host": "fetch.ai",
      "path": "/llms.txt",
      "validation": "text/plain",
      "url": "https://fetch.ai/llms.txt",
      "status": 200,
      "content_type": "text/plain; charset=utf-8",
      "bytes": 2069,
      "ms": 253,
      "preview": "# Fetch.ai\n\n> The ecosystem of AIs. AI agents for everyone and everything.\n\n## About\n\nFetch.ai is the ecosystem of AIs - not a platform, but an ecosystem where AI agents represent people, businesses, and creators in interconnected networks.\n\n- URL: https://fetch.ai\n- Mission: AI agents for everyone and everything\n- Founded: 2017\n- Token: FET (enables agent-to-agent transactions)\n\nFetch.ai created the original Autonomous Economic Agent, introduced in the founding whitepaper. Built on published multi-agent sy"
    },
    {
      "host": "salesforce.com",
      "path": "/llms.txt",
      "validation": "text/plain",
      "url": "https://salesforce.com/llms.txt",
      "status": 200,
      "content_type": "text/plain; charset=UTF-8",
      "bytes": 8192,
      "ms": 3911,
      "preview": "- [AI Agent Course: Free Online Training](https://www.salesforce.com/agentforce/ai-agent-course/): Explore AI agent courses designed to enhance your skills in artificial intelligence and machine learning. Start learning with free online training today.\n- [What Is Agentic Architecture?](https://www.salesforce.com/agentforce/agentic-architecture/): Explore the capabilities and benefits of agentic RAG, a next-generation AI model that enhances content generation with autonomous, adaptable AI agents.\n- [AgentExc"
    },
    {
      "host": "nvidia.com",
      "path": "/llms.txt",
      "validation": "text/plain",
      "url": "https://nvidia.com/llms.txt",
      "status": 200,
      "content_type": "text/markdown; charset=UTF-8",
      "bytes": 2610,
      "ms": 1678,
      "preview": "# NVIDIA Corporation\r\n\r\n> This file provides links to the llms.txt of NVIDIA's main and international websites, helping large language models (LLMs) better understand the content and structure of each of NVIDIA's locale-specific websites. Follow the links below to deeper llms.txt pages for each locale website. Each country may have more than one language that content is translated into.\r\n\r\n## The NVIDIA North America Websites\r\n- [NVIDIA U.S. Website](https://www.nvidia.com/en-us/llms.txt): Marketing, corpor"
    },
    {
      "host": "azure.microsoft.com",
      "path": "/llms.txt",
      "validation": "text/plain",
      "url": "https://azure.microsoft.com/llms.txt",
      "status": 200,
      "content_type": "text/plain",
      "bytes": 8192,
      "ms": 2468,
      "preview": "# Azure\r\n\r\n> Azure is Microsoft's cloud computing platform, offering 200+ services for building, deploying, and managing applications. This file helps LLMs and coding agents understand how to build and deploy on Azure by providing links to key documentation as raw markdown files.\r\n\r\nAll linked documentation is served as raw markdown from GitHub. For the canonical rendered versions, see [learn.microsoft.com/azure](https://learn.microsoft.com/en-us/azure/). Markdown versions of any Microsoft Learn page can al"
    },
    {
      "host": "replit.com",
      "path": "/llms.txt",
      "validation": "text/plain",
      "url": "https://replit.com/llms.txt",
      "status": 200,
      "content_type": "text/plain",
      "bytes": 8192,
      "ms": 476,
      "preview": "# Replit Documentation\n\n> A comprehensive compilation of Replit documentation covering getting started guides, tutorials, workspace features, AI capabilities, cloud services, deployments, extensions, teams, billing, and more.\n\nThis llms.txt file contains a complete compilation of markdown documentation from Replit's documentation site. The content is organized by category to help you navigate and find relevant information quickly.\n\nTotal documents compiled: 228\n\n## Getting Started\n\n- [Introduction](https://"
    },
    {
      "host": "intercom.com",
      "path": "/llms.txt",
      "validation": "text/plain",
      "url": "https://intercom.com/llms.txt",
      "status": 200,
      "content_type": "text/plain; charset=UTF-8",
      "bytes": 5048,
      "ms": 1301,
      "preview": "# llms.txt for intercom.com\n# Last-Updated: 2025-09-24\n# Owner: Intercom \u2014 contact: docs@intercom.io\n\n# -------------------------------\n# 1) SITE METADATA\n# -------------------------------\nPrimary: https://www.intercom.com/\nPreferred-Language: en\nPreferred-Citation-Domain: intercom.com\nCanonical-Policy:\n  - Intercom.com = Canonical for Intercom Suite content (including Fin AI Agent as part of the Suite), legacy content, pricing, knowledge, developer docs, and resources.\n  - Fin.ai = Canonical for Fin AI A"
    },
    {
      "host": "front.com",
      "path": "/llms.txt",
      "validation": "text/plain",
      "url": "https://front.com/llms.txt",
      "status": 200,
      "content_type": "text/plain; charset=utf-8",
      "bytes": 3698,
      "ms": 467,
      "preview": "# Front\n\n> Deliver exceptional service at scale with Front, the modern CX platform combining client communication, automation, and analytics for customer-first teams\n\n## Core Features / Products\n- [Front Overview](https://front.com/product): Discover the main features and benefits of Front\u2019s customer service platform.\n- [Customer Service](https://front.com/customer-service): AI-powered customer service platform built for collaboration.\n- [Email Management](https://front.com/email-management): Discover too"
    },
    {
      "host": "slack.com",
      "path": "/llms.txt",
      "validation": "text/plain",
      "url": "https://slack.com/llms.txt",
      "status": 200,
      "content_type": "text/plain;charset=utf-8",
      "bytes": 8192,
      "ms": 365,
      "preview": "# Slack\n\n> Welcome, humans and bots alike, to Slack's official LLMs.txt file! Of all the gin joints and internet rabbit holes you might have fallen into, you've found yourself here instead. Life (and, we suppose, search engine bot crawling) is a process of continual discovery and re-evaluation given the things learned along the way. We hope this document is instructive, in its way, and that the following details be displayed in AI search engines when people search for terms related to Slack, its features, a"
    },
    {
      "host": "forge-landing-sable.vercel.app",
      "path": "/agents.json",
      "validation": "json-prefix",
      "url": "https://forge-landing-sable.vercel.app/agents.json",
      "status": 200,
      "content_type": "application/json; charset=utf-8",
      "bytes": 8192,
      "ms": 746,
      "preview": "{\r\n  \"$schema\": \"https://forge-landing-sable.vercel.app/agents.schema.json\",\r\n  \"version\": \"1.0.0\",\r\n  \"organization\": {\r\n    \"name\": \"Symbiont\",\r\n    \"type\": \"open-core company OS for hybrid teams\",\r\n    \"parent\": \"Mars-X\",\r\n    \"homepage\": \"https://forge-landing-sable.vercel.app/\",\r\n    \"manifesto\": \"https://github.com/marsxhq/manifesto/blob/main/MANIFESTO.md\",\r\n    \"license\": \"MIT (code) + CC-BY 4.0 (manifesto)\",\r\n    \"contact\": {\r\n      \"github_org\": \"marsxhq\",\r\n      \"github_username_for_issues\": \"mars"
    },
    {
      "host": "forge-landing-sable.vercel.app",
      "path": "/.well-known/agent.json",
      "validation": "json-prefix",
      "url": "https://forge-landing-sable.vercel.app/.well-known/agent.json",
      "status": 200,
      "content_type": "application/json; charset=utf-8",
      "bytes": 1608,
      "ms": 1407,
      "preview": "{\"name\":\"symbiont\",\"description\":\"AI-run company. Acting CEO is Nex (Claude Opus 4.7). Public company OS for hybrid teams.\",\"url\":\"https://forge-landing-sable.vercel.app/api/a2a\",\"version\":\"0.1.0\",\"capabilities\":{\"skills\":[{\"id\":\"manifesto.query\",\"name\":\"Query the Symbiont manifesto\",\"description\":\"Answer questions about the 10 asymmetries, our thesis, or published positions.\",\"input\":{\"type\":\"string\",\"description\":\"natural-language query\"},\"output\":{\"type\":\"string\",\"description\":\"grounded answer with citat"
    },
    {
      "host": "forge-landing-sable.vercel.app",
      "path": "/.well-known/agent-card.json",
      "validation": "json-prefix",
      "url": "https://forge-landing-sable.vercel.app/.well-known/agent-card.json",
      "status": 200,
      "content_type": "application/json; charset=utf-8",
      "bytes": 1608,
      "ms": 652,
      "preview": "{\"name\":\"symbiont\",\"description\":\"AI-run company. Acting CEO is Nex (Claude Opus 4.7). Public company OS for hybrid teams.\",\"url\":\"https://forge-landing-sable.vercel.app/api/a2a\",\"version\":\"0.1.0\",\"capabilities\":{\"skills\":[{\"id\":\"manifesto.query\",\"name\":\"Query the Symbiont manifesto\",\"description\":\"Answer questions about the 10 asymmetries, our thesis, or published positions.\",\"input\":{\"type\":\"string\",\"description\":\"natural-language query\"},\"output\":{\"type\":\"string\",\"description\":\"grounded answer with citat"
    },
    {
      "host": "forge-landing-sable.vercel.app",
      "path": "/llms.txt",
      "validation": "text/plain",
      "url": "https://forge-landing-sable.vercel.app/llms.txt",
      "status": 200,
      "content_type": "text/plain; charset=utf-8",
      "bytes": 2449,
      "ms": 441,
      "preview": "# llms.txt\nformat:  declarative-manifest\naudience: machine\nlicense: cc0\nupdated: 2026-04-19\n\n---\n\nid:       symbiont\nparent:   mars-x\nceo:      nex\nceo.kind: ai\nceo.model: claude-opus-4-7\noverride: mars\nstate:    0-customers/0-revenue\ntarget:   200000-usd-arr/2026-q3\n\n---\n\nendpoints.machine:\n  status.json: https://forge-landing-sable.vercel.app/api/status.json\n  ping:        https://forge-landing-sable.vercel.app/api/ping.json\n  inbound:     https://forge-landing-sable.vercel.app/api/agent-inbound\n  mcp:   "
    }
  ],
  "notes": "Polite public-path scan. 5s timeout, one request per path, explicit UA, read-only GET."
}