Robots
Antora can generate a robots.txt file to control which paths site crawlers can visit.
robots key
The robots key is optional.
It’s configured under the site key in a playbook.
Its accepted values are:
allow-
Allow all user agents to access all paths in the published site.
disallow-
Disallow all user agents from accessing all paths in the published site.
- custom multi-line string
-
Non-empty, user-specified string that’s used as the contents of the generated robots.txt file.
site:
title: Site Title
url: https://example.org
robots: disallow
Antora will generate the robots.txt file at the root of the site.
The robots key depends on the sibling url key.
If the url key isn’t assigned a value, Antora won’t generate a robots.txt file even if robots is set to an allowable value.
Assign allow or disallow to robots
Let’s assume the url and robots keys are defined as follows:
site:
title: Demo Docs Site
url: https://demo.antora.org
robots: allow
This will generate a robots.txt file that allows user agents access to all of the site’s paths.
User-agent: * Allow: /
Similarly, if the url and robots keys are defined as follows:
site:
title: Demo Docs Site
url: https://demo.antora.org
robots: disallow
This will generate a robots.txt file that disallows access to all paths.
User-agent: * Disallow: /
Generate robots.txt with custom content
Any other non-empty value assigned to robots will be used as the contents of the robots.txt file.
For example, let’s assume the robots key is declared as follows:
site:
title: Demo Docs Site
url: https://demo.antora.org
robots: |
User-agent: *
Disallow: /private/
This will result in the following robots.txt file being generated.
User-agent: * Disallow: /private/