WebSite Auditor is equipped with a powerful SEO spider that crawls your site just like search engine bots do.
Not only it looks at your site's HTML, but also sees your CSS, JavaScript, Flash, images, videos, PDFs, and other resources, both internal and external, and shows you how all pages and resources interlink. This allows you to see the whole picture and make decisions based on the complete comprehensive data.
Crawler settings can be tailored to your needs and preferences: for instance, to collect or exclude certain sections of a site, crawl a site on behalf of any search engine bot, find pages that are unlinked from site, etc. To configure the settings, simply tick Enable expert options box once you are creating a project.
In Step 2 you’ll be able to specify crawler settings for the current project.
Robots.txt Instructions section features the following settings:
In the Filtering section you can specify various filtering conditions WebSite Auditor will use while collecting pages and resources to your project.
The Speed section allows limit the number of requests to the website, to decrease the load on the server. This prevents slower sites (or sites with high security restrictions) from blocking the crawler.
In the JavaScript section, you can control whether the app should execute the scripts while crawling your website pages - this will allow crawling websites built with Ajax, for instance, and/or parsing any script-generated content in full.
In the URL Parameters section, you can specify whether the program should collect the dynamic pages.
Advanced Options section contains additional crawler settings such as:
After configuring Crawler Settings, hit Finish for the program to start crawling your site. The settings can be accessed and changed any time later under ‘Preferences > Crawler Settings’ in each project.