Is your Site Audit not running properly?
There are a number of reasons why pages could be blocked from the Site Audit crawler based on your website’s configuration and structure, including:
- Robots.txt blocking crawler
- Crawl scope excluding certain areas of the site
- Website is not directly online due to shared hosting
- Pages are behind a gateway / user base area of site
- Crawler blocked by noindex tag
- Domain could not be resolved by DNS - the domain entered in setup is offline -
Follow these troubleshooting steps to see if you can make any adjustments on your own before reaching out to our support team for help.
A Robots.txt file gives instructions to bots about how to crawl (or not crawl) the pages of a website. You can allow and forbid bots such as Googlebot or Semrushbot from crawling all of your site or specific areas of your site using commands such as Allow, Disallow, and Crawl Delay.
If your robots.txt is disallowing our bot from crawling your site, our Site Audit tool will not be able to check your site.
You can inspect your Robots.txt for any disallow commands that would prevent crawlers like ours from accessing your website.
To allow the Semrush Site Audit bot (SiteAuditBot) to crawl your site, add the following into your robots.txt file:
(leave a blank space after “Disallow:”)
Here’s an example of how a robots.txt file may look:
Note the various commands based on the user agent (crawler) that the file is addressing.
These files are public and in order to be found must be hosted on the top level of a site. To find a website’s robots.txt file, enter the root domain of a site followed by /robots.txt into your browser. For example, the robots.txt file on Semrush.com is found at https://semrush.com/robots.txt.
Some terms you may see on a robots.txt file include:
- User-Agent = the web crawler you are giving instructions to.
- Ex: SiteAuditBot, Googlebot
- Allow = a command (only for Googlebot) that tells the bot it can crawl a specific page or area of a site even if the parent page or folder is disallowed.
- Disallow = a command that tells the bot not to crawl a specific URL or subfolder of a site.
- Ex: Disallow: /admin/
- Crawl Delay = a command that tells bots how many seconds to wait before loading and crawling another page.
- Sitemap = indicating where the sitemap.xml file for a certain URL is.
- / = use the “/” symbol after a disallow command to tell the bot not to crawl the entirety of your site
- * = a wildcard symbol that represents any string of possible characters in a URL, used to indicate an area of a site or all user agents.
- Ex: Disallow: /blog/* would indicate all URLs in a site’s blog subfolder
- Ex: User agent: * would indicate instructions for all bots
If you see the following code on the main page of a website, it tells us that we’re not allowed to index/follow links on it and our access is blocked.
<meta name="robots" content="noindex, nofollow" >
Or, a page containing at least one of the following: "noindex", "nofollow", "none", will lead to the error of crawling.
To allow our bot to crawl such a page, remove these “noindex” tags from your page’s code. For more information on the noindex tag, please refer to this Google Support article.
To whitelist the bot, contact your webmaster or hosting provider and ask them to whitelist SiteAuditBot.
The bot's IP addresses are: 220.127.116.11/25 (a subnet used by Site Audit only)
The bot is using standard 80 HTTP and 443 HTTPS ports to connect.
If you use any plugins (Wordpress, for example) or CDNs (content delivery networks) to manage your site, you will have to whitelist the bot IP within those as well.
For whitelisting on Wordpress, contact Wordpress support.
Common CDNs that block our crawler include:
- Cloudflare - read how to whitelist here
- Imperva - read how to whitelist here
- ModSecurity - read how to whitelist here
- Sucuri - read how to whitelist here
Please note: If you have shared hosting, it is possible that your hosting provider may not allow you to whitelist any bots or edit the Robots.txt file.
Below is a list of some of the most popular hosting providers on the web and how to whitelist a bot on each or reach their support team for assistance:
- Siteground - whitelisting instructions
- 1&1 IONOS - whitelisting instructions
- Bluehost* - whitelisting instructions
- Hostgator* - whitelisting instructions
- Hostinger - whitelisting instructions
- GoDaddy - whitelisting instructions
- GreenGeeks - whitelisting instructions
- Big Commerce - Must contact support
- Liquid Web - Must contact support
- iPage - Must contact support
- InMotion - Must contact support
- Glowhost - Must contact support
- A2 Hosting - Must contact support
- DreamHost - Must contact support
* Please note: these instructions work for HostGator and Bluehost if you have a website on VPS or Dedicated Hosting.
To see how much of your current crawl budget has been used, go to Profile - Subscription Info and look for “Pages to crawl” under “My plan.”
Depending on your subscription level, you are limited to a set number of pages that you can crawl in a month (monthly crawl budget). If you go over the amount of pages allowed within your subscription, you’ll have to purchase additional limits or wait until the next month when your limits will refresh.
If the domain could not be resolved by DNS, it likely means that the domain you entered during configuration is offline. Commonly, users have this issue when entering a root domain (example.com) without realizing that the root domain version of their site doesn’t exist and the WWW version of their site would need to be entered instead (www.example.com).
To prevent this issue, the website owner could add a redirect from the unsecured “example.com” to the secured “www.example.com” that exists on the server. This issue could also occur the other way around, if someone’s root domain is secured, but their WWW version is not. In such a case, you would just have to redirect the WWW version to the root domain.
We crawl JS and CSS and make performance checks (minify, compressed). We can not render JS because we cannot get content and links which show only after rendering.
In order to not miss the most important pages on your website with our crawl, you can change your crawl source from website to sitemap - this way we won’t miss any pages that are mentioned in the sitemap.
Your website may be blocking the SemrushBot in your robots.txt file. You can change the User Agent from SemrushBot to GoogleBot and your website is likely to allow Google’s User Agent to crawl. To make this change, find the settings gear in your Project and select User Agent.
If this option is used, blocked internal resources and pages blocked from crawl checks will not be triggered. Keep in mind that to use this, site ownership will have to be verified.
This is useful for sites that are currently under maintenance. It’s also helpful for when the site owner does not want to modify the robots.txt file.
To audit private areas of your website that are password protected enter your credentials in the “Crawling with your credentials” option under the settings gear. This slide allows the Site Audit bot to reach those pages and audit them for you.
This is highly recommended for sites still under development or are private and fully password protected.
Contact Semrush Support
If you still are having issues running your Site Audit, contact our Support team, we are happy to help!
Further reading: Check out our post about the Biggest SEO Mistakes Damaging Websites in 2021.