diff options
Diffstat (limited to 'nikola/plugins/task/robots.py')
| -rw-r--r-- | nikola/plugins/task/robots.py | 83 |
1 files changed, 83 insertions, 0 deletions
diff --git a/nikola/plugins/task/robots.py b/nikola/plugins/task/robots.py new file mode 100644 index 0000000..9944c0d --- /dev/null +++ b/nikola/plugins/task/robots.py @@ -0,0 +1,83 @@ +# -*- coding: utf-8 -*- + +# Copyright © 2012-2014 Roberto Alsina and others. + +# Permission is hereby granted, free of charge, to any +# person obtaining a copy of this software and associated +# documentation files (the "Software"), to deal in the +# Software without restriction, including without limitation +# the rights to use, copy, modify, merge, publish, +# distribute, sublicense, and/or sell copies of the +# Software, and to permit persons to whom the Software is +# furnished to do so, subject to the following conditions: +# +# The above copyright notice and this permission notice +# shall be included in all copies or substantial portions of +# the Software. +# +# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY +# KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE +# WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR +# PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS +# OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR +# OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR +# OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE +# SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. + +from __future__ import print_function, absolute_import, unicode_literals +import codecs +import os +try: + from urlparse import urljoin, urlparse +except ImportError: + from urllib.parse import urljoin, urlparse # NOQA + +from nikola.plugin_categories import LateTask +from nikola import utils + + +class RobotsFile(LateTask): + """Generate a robots.txt.""" + + name = "robots_file" + + def gen_tasks(self): + """Generate a robots.txt.""" + kw = { + "base_url": self.site.config["BASE_URL"], + "site_url": self.site.config["SITE_URL"], + "output_folder": self.site.config["OUTPUT_FOLDER"], + "files_folders": self.site.config['FILES_FOLDERS'], + "robots_exclusions": self.site.config["ROBOTS_EXCLUSIONS"] + } + + if kw["site_url"] != urljoin(kw["site_url"], "/"): + utils.LOGGER.warn('robots.txt not ending up in server root, will be useless') + + sitemapindex_url = urljoin(kw["base_url"], "sitemapindex.xml") + robots_path = os.path.join(kw['output_folder'], "robots.txt") + + def write_robots(): + with codecs.open(robots_path, 'wb+', 'utf8') as outf: + outf.write("Sitemap: {0}\n\n".format(sitemapindex_url)) + if kw["robots_exclusions"]: + outf.write("User-Agent: *\n") + for loc in kw["robots_exclusions"]: + outf.write("Disallow: {0}\n".format(loc)) + + yield self.group_task() + + if not utils.get_asset_path("robots.txt", [], files_folders=kw["files_folders"]): + yield { + "basename": self.name, + "name": robots_path, + "targets": [robots_path], + "actions": [(write_robots)], + "uptodate": [utils.config_changed(kw)], + "clean": True, + "task_dep": ["sitemap"] + } + elif kw["robots_exclusions"]: + utils.LOGGER.warn('Did not generate robots.txt as one already exists in FILES_FOLDERS. ROBOTS_EXCLUSIONS will not have any affect on the copied fie.') + else: + utils.LOGGER.debug('Did not generate robots.txt as one already exists in FILES_FOLDERS.') |
