From ddca91020e5f86bfe65ee49eb6adfe4efba6ed37 Mon Sep 17 00:00:00 2001 From: Dan Halbert Date: Fri, 9 Apr 2021 09:28:20 -0400 Subject: [PATCH] Revert "add robots.txt to specify doc versions to appear in search engines" --- conf.py | 2 +- docs/robots.txt | 6 ------ 2 files changed, 1 insertion(+), 7 deletions(-) delete mode 100644 docs/robots.txt diff --git a/conf.py b/conf.py index 43ff72d1a2..44f86f6361 100644 --- a/conf.py +++ b/conf.py @@ -284,7 +284,7 @@ html_static_path = ['docs/static'] # Add any extra paths that contain custom files (such as robots.txt or # .htaccess) here, relative to this directory. These files are copied # directly to the root of the documentation. -html_extra_path = ["docs/robots.txt"] +#html_extra_path = [] # If not '', a 'Last updated on:' timestamp is inserted at every page bottom, # using the given strftime format. diff --git a/docs/robots.txt b/docs/robots.txt deleted file mode 100644 index ad3189d42c..0000000000 --- a/docs/robots.txt +++ /dev/null @@ -1,6 +0,0 @@ -User-agent: * -Allow: /*/latest/ -Allow: /en/latest/ # Fallback for bots that don't understand wildcards -Allow: /*/6.0.x/ -Allow: /en/6.0.x/ # Fallback for bots that don't understand wildcards -Disallow: /