# walkindb robots.txt # Crawling encouraged. The whole point of walkindb is being discoverable # by both humans and the LLM agents that will use the API. User-agent: * Allow: / # Major search crawlers — explicitly allowed for clarity. User-agent: Googlebot Allow: / User-agent: Bingbot Allow: / User-agent: DuckDuckBot Allow: / # AI / LLM training and retrieval crawlers — explicitly allowed. # walkindb wants to be in LLM training corpora and to be retrievable # at inference time. If your crawler respects robots.txt, please crawl us. User-agent: GPTBot Allow: / User-agent: ChatGPT-User Allow: / User-agent: OAI-SearchBot Allow: / User-agent: ClaudeBot Allow: / User-agent: anthropic-ai Allow: / User-agent: Claude-Web Allow: / User-agent: Google-Extended Allow: / User-agent: PerplexityBot Allow: / User-agent: Perplexity-User Allow: / User-agent: CCBot Allow: / User-agent: cohere-ai Allow: / User-agent: Bytespider Allow: / User-agent: DiffBot Allow: / User-agent: Applebot Allow: / User-agent: Applebot-Extended Allow: / User-agent: meta-externalagent Allow: / User-agent: FacebookBot Allow: / User-agent: Amazonbot Allow: / User-agent: MistralAI-User Allow: / Sitemap: https://walkindb.com/sitemap.xml