How to Add Bots to Your Robots.txt File (Ahrefs Example)

Link: https://support.brilliantdirectories.com/support/solutions/articles/12000099791-how-to-add-bots-to-your-robots-txt-file-ahrefs-example-

A robots.txt file is a text file placed on a website's server that instructs web crawlers and other automated processes which pages or sections of the site should or should not be accessed. It's a protocol used to communicate with web robots and is commonly used for managing website indexing and crawling by search engines like Google, Bing, SEO tools and others.


Adding a Web Robot to the site

Having a web robot crawl a site involves granting permission for it to access and index the site's content.


To allow third-party tools to crawl a website, they must be included in the Allowed list within the website's Robots.txt file.


The ROBOTS.TXT file can be accessed through:


  • Navigate to the Developer Hub sidebar menu link.
  • Click on the Robots File option.

To incorporate https://ahrefs.com/ into the list, please add the following lines to the beginning of the code:


User-agent: AhrefsSiteAudit
Allow: /


User-agent: AhrefsBot
Allow: /



Lastly, click on "Save Changes". Now, Ahrefs has permission to crawl the website without encountering any blockages.