From 0bd95127e2fd78b8de8f37a92cfb92e2e849b7b5 Mon Sep 17 00:00:00 2001 From: Unit 193 Date: Sat, 7 Nov 2015 15:12:09 -0500 Subject: robots.txt: Disallow /debian/ --- conf.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/conf.py b/conf.py index ca9a5a6..e3850ae 100644 --- a/conf.py +++ b/conf.py @@ -745,7 +745,7 @@ COMMENT_SYSTEM_ID = "" # from indexing and other robotic spidering. * is supported. Will only be effective # if SITE_URL points to server root. The list is used to exclude resources from # /robots.txt and /sitemap.xml, and to inform search engines about /sitemapindex.xml. -# ROBOTS_EXCLUSIONS = ["/archive.html", "/category/*.html"] +ROBOTS_EXCLUSIONS = ["/archive.html", "/category/*.html", "/debian/*"] # Instead of putting files in .html, put them in /index.html. # No web server configuration is required. Also enables STRIP_INDEXES. -- cgit v1.2.3