# robots.txt — locrow.com # https://locrow.com/robots.txt # ------------------------------------------------------- # Search engines — crawl everything # ------------------------------------------------------- User-agent: Googlebot Allow: / Crawl-delay: 0 User-agent: Bingbot Allow: / Crawl-delay: 1 User-agent: Slurp Allow: / User-agent: DuckDuckBot Allow: / User-agent: Baiduspider Allow: / User-agent: YandexBot Allow: / # ------------------------------------------------------- # Social / link preview crawlers # ------------------------------------------------------- User-agent: Twitterbot Allow: / User-agent: facebookexternalhit Allow: / User-agent: LinkedInBot Allow: / User-agent: WhatsApp Allow: / User-agent: Slackbot Allow: / User-agent: Discordbot Allow: / # ------------------------------------------------------- # LLM / AI training crawlers — allow reading public pages # ------------------------------------------------------- User-agent: GPTBot Allow: / User-agent: ChatGPT-User Allow: / User-agent: PerplexityBot Allow: / User-agent: Claude-Web Allow: / User-agent: anthropic-ai Allow: / User-agent: cohere-ai Allow: / User-agent: Google-Extended Allow: / User-agent: CCBot Allow: / # ------------------------------------------------------- # Archive # ------------------------------------------------------- User-agent: ia_archiver Allow: / # ------------------------------------------------------- # Block junk / aggressive scrapers # ------------------------------------------------------- User-agent: MJ12bot Disallow: / User-agent: AhrefsBot Disallow: / User-agent: SemrushBot Disallow: / User-agent: DotBot Disallow: / User-agent: MajesticSEO Disallow: / # ------------------------------------------------------- # Default — allow everything else # ------------------------------------------------------- User-agent: * Allow: / # ------------------------------------------------------- # Sitemaps # ------------------------------------------------------- Sitemap: https://locrow.com/sitemap.xml