From f4df0e72ba90ef3a58dbabbe42c7d73f0be21370 Mon Sep 17 00:00:00 2001 From: Ido Shamun <1993245+idoshamun@users.noreply.github.com> Date: Sun, 12 Apr 2026 14:30:23 +0300 Subject: [PATCH] fix: unblock sitemap fetching in robots.txt Replace broad Disallow: /api/ with specific Disallow: /api/r/ for the redirector. The broad rule was preventing Google from fetching sitemaps at /api/sitemaps/ despite the Allow override. Also add the API sitemap index as a second Sitemap directive. --- packages/webapp/public/robots.txt | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/packages/webapp/public/robots.txt b/packages/webapp/public/robots.txt index 4b82fabcc76..f276aef9446 100644 --- a/packages/webapp/public/robots.txt +++ b/packages/webapp/public/robots.txt @@ -58,17 +58,17 @@ Allow: / # All other crawlers User-agent: * Allow: / -Allow: /api/sitemaps/ Disallow: /join Disallow: /error Disallow: /callback Disallow: /helloworld Disallow: /helloworld/ Disallow: /welcome -Disallow: /api/ +Disallow: /api/r/ -# Sitemap index +# Sitemap indexes Sitemap: https://app.daily.dev/sitemap.xml +Sitemap: https://app.daily.dev/api/sitemaps/index.xml # LLM-friendly documentation # https://app.daily.dev/llms.txt