--- title: Robots.txt File linktitle: Robots.txt description: Hugo can generate a customized robots.txt in the same way as any other template. date: 2017-02-01 publishdate: 2017-02-01 lastmod: 2017-02-01 categories: [templates] keywords: [robots,search engines] menu: docs: parent: "templates" weight: 165 weight: 165 sections_weight: 165 draft: false aliases: [/extras/robots-txt/] toc: false --- To generate a robots.txt file from a template, change the [site configuration][config]: {{< code-toggle file="config">}} enableRobotsTXT = true {{< /code-toggle >}} By default, Hugo generates robots.txt using an [internal template][internal]. ```text User-agent: * ``` Search engines that honor the Robots Exclusion Protocol will interpret this as permission to crawl everything on the site. ## Robots.txt Template Lookup Order You may overwrite the internal template with a custom template. Hugo selects the template using this lookup order: 1. `/layouts/robots.txt` 2. `/themes//layouts/robots.txt` ## Robots.txt Template Example {{< code file="layouts/robots.txt" download="robots.txt" >}} User-agent: * {{ range .Pages }} Disallow: {{ .RelPermalink }} {{ end }} {{< /code >}} This template creates a robots.txt file with a `Disallow` directive for each page on the site. Search engines that honor the Robots Exclusion Protocol will not crawl any page on the site. {{% note %}} To create a robots.txt file without using a template: 1. Set `enableRobotsTXT` to `false` in the [site configuration][config]. 2. Create a robots.txt file in the `static` directory. Remember that Hugo copies everything in the [static directory][static] to the root of `publishDir` (typically `public`) when you build your site. [config]: /getting-started/configuration/ [static]: /getting-started/directory-structure/ {{% /note %}} [config]: /getting-started/configuration/ [internal]: https://github.com/gohugoio/hugo/blob/master/tpl/tplimpl/embedded/templates/_default/robots.txt