Adding robots.txt to your website
Robots.txt is a text file with instructions for search engine crawlers. It defines which areas of a website crawlers are allowed to search.
You can configure whether to allow / disallow the entire website or configure the rules for each folder individually.
The robots.txt configuration is available via the Tools menu.
Specifies whether to generate robots.txt file when the website is published. The file will be published in the root of the website.
For example: https://www.yourwebsite.com/robots.txt
This sets the rule for the entire website. This can be overridden for each page and folder.
Allow all, allow robots to index all pages and folders on the web site .
Disallow, disallow robots from indexing all pages and folders on the web site.
Pages and Folders
Under 'Pages and Folders' you can override rules for individual pages and folders.
First select the page in the site tree and then select one of the rules from the drop down list:
The page will use the rule as configured in the 'General' section.
Tells a search engine to index a page.
Tells a search engine not to index a page.
Even if the page isnít indexed, the crawler should follow all the links on a page and pass equity to the linked pages.
Tells a crawler not to follow any links on a page or pass along any link equity.
Note that you can also set Robots index/follow options in the Page Properties. These are the same settings.
In Quick 'n Easy Web Builder, assets (like images, pdf, videos, css, etc.) can be published to a different folder.
These folders can also be included / excluded for indexing. This can be useful if you do not want your documents (PDF) or images to be included in search results.
Asset folders can be configured in Preferences -> Publish