robots.txt
file is essential for controlling search engine crawlers’ access to your site. It provides information on which pages and sections can be visited by web crawlers.
If you need to customize the contents of this file when your application is being deployed, you can add a command to the Build Commands section of your environment’s deployment settings.
For example, if you want to disallow crawling any part of your website:
/cdn-cgi
path on all hosted domains. This path appears on your domain as https://www.example.com/cdn-cgi/l/email-protection
and similar endpoints.
These /cdn-cgi
endpoints are:
robots.txt
file to exclude these paths from search engine crawling: