
By default, the tool shows 1,000 pages (which can be adjusted up to 10,000 pages) that are arranged by Click Depth. If any of the factors have an error or warning status, study the Details section to see problem pages and get how-to's on fixing them.ġ. Under Indexing and crawlability in the Site Audit module, go through Resources with 4xx status code, Resources with 5xx status code, and 404 page set up correctly. Indexing issues can be also caused by HTTP response codes errors. In the menu that pops up, you can either fetch your robots.txt from server to revise it, or create a robots.txt file from scratch and upload it to your website.Ĥ. Now, if you need to create a robots.txt file, or fix its instructions, simply switch to the Pages module, click and choose. Revise your robots.txt file (or create it from scratch)

The main aspects to pay attention to are proper indexing instructions in your robots.txt file and proper HTTTP response codes.ġ.Ĝheck if your robots.txt file is in place. If your site is hard to crawl and index, you're probably missing out on lots of ranking opportunities on the other hand, you may well be willing to hide certain parts of your site from search engines (say, pages with duplicate content). It's crucial that users and search engines can easily reach all the important pages and resources on your site, including JavaScript and CSS. Step 2: Check crawlability and indexing issues Finally, hit Next to proceed with the crawling. Select the bot you'd like to crawl your site as if you'd like to discard robots.txt during the crawl, simply uncheck the Follow robots.txt instructions box. At Step 2, click on the drop-down menu next to the Follow robots.txt instructions option. At Step 1, enter your site's URL and check the Enable expert options box.

To do this, create a WebSite Auditor project (or rebuild an existing one). – or discard robots instructions altogether and collect all pages of your site, even the ones disallowed in your robots.txt. You may want to tweak this setting to crawl the site as Google, Bing, Yahoo, etc. By default, WebSite Auditor crawls your site using a spider called SEO-PowerSuite-bot, which means it will obey robots instructions for all bots (user agent: *). Crawl your site as Googlebot (or any other bot).
