Carrying out a successful and efficient Technical SEO Audit starts with gathering the correct audit data.
This article is designed as a companion to Sitebulb's Technical SEO Audit Checklist & Template. We will look at each template section and point you to the settings you'll need to enable in order to gather the data point you'll need to find useful insight and points of action.
Some of these settings are optional, and should only be enabled if you know you'll require the specific point of data they provide for your specific auditing goals.
Indexability & Crawlability
Choosing the correct Crawler
Choosing the right crawler is essential in ensuring that all your page content is rendered and crawled. Sitebulb offers two crawler options:
The HTML crawler executes traditional HTML extraction; it is the fastest option.
The Chrome Crawler renders JavaScript using the latest stable version of Chromium.
Choose the Chrome crawler if you know your website renders content using JavaScript.
If in doubt, use the Single Page Analysis tool to determine whether your website renders content with JavaScript. Adjust the crawler accordingly.
Save Disallowed URLs to review disallowed pages
By default, Sitebulb will respect all robots directives, unless otherwise specified in your Robots Directives Audit Settings. As a result, disallowed directives will be respected, and disallowed pages will not be part of your audit by default.
If you wish to analyse how robots' directives impact the crawlability of your pages, you can choose to save disallowed URLs to your Audit by ticking ‘Save disallowed URLs’ under the Politeness tab in the Robots Directives settings:
Review robots.txt politeness to crawl staging websites
Staging sites are likely to have directives in place to prevent the content from being crawled and indexed. Since Sitebulb respects robots directives by default, you will need to adjust your robots Politeness settings to ensure that Sitebulb can crawl successfully.
Navigate to the Robots Directives tab on your Audit Settings and find the Politeness section. Here, you can tick 'Is Staging Site' - this will remove the root path disallow from the robots.txt file, but instruct Sitebulb to respect the other robots directives.
Additionally, you can also choose to fully disable the 'Respect Robots Directives' setting, and even enable the 'Crawl disallowed URLs' and 'Crawl links with nofollow directives' settings, giving Sitebulb full access to the content you may want to crawl and save.
Add your XML Sitemap as a crawl source
By default, Sitebulb will only crawl the pages found through internal links. You'll need to add your XML Sitemaps as a crawl source if you wish to analyse its content, find orphan pages, and pick up hreflang implemented through XML Sitemaps.
To crawl and analyse the content of your XML Sitemaps with Sitebulb, ensure that XML Sitemaps are selected as a Crawl Source in your Audit Settings.
Internal Linking
By default, Sitebulb reports on all internal and external links found on your pages. When 'Crawler' is selected as the Crawl Source (default), Sitebulb discovers new internal pages by following internal links on every page it crawls.
External links are not crawled. By default, Sitebulb will only report on the HTTP status of all external links found on your pages.
To reduce the size of your audit, you can opt to exclude external links from your audit by disabling the External Link Analysis setting under Advanced Search Engine Optimisation Settings.
Find and review orphaned pages
We have a whole article dedicated to how to find orphaned pages on your website. Here is a summary of the key steps to help you identify orphan pages in your Audit.
You will need to set up your audit to allow you to find as many of your Internal pages as possible. To find orphan pages, you need to compare crawl data with other URL sources.
Connect Google Analytics and enable the ‘Extract and crawl any URLs found in Google Analytics’ feature.
Connect Google Search Console and enable the ‘Extract and crawl any URLs found in Google Search Console’
Add your XML Sitemaps as a crawl source
Optionally, add seed URLs for any isolated areas of the site.
Optionally, add a URL list of any pages that may not be found through other sources (if you know about these)
On Page Content
Review Alt Text
To analyse alt text, you will first need to ensure that Page Resources (specifically, images) are part of your Audit.
Before starting your audit, navigate to Audit Settings > Audit Data and enable the Page Resources report - this will ensure that Sitebulb collects all page resources, including images, and can create reports based on this data.
Performance & Mobile Friendly
Establish the scope of your performance Audit
Performance audits can be time and resource-intensive, as the crawler needs to collect and analyse every single page resource in order to report accurately on performance.
If you deal with bigger sites, you may want to consider setting up a separate Sitebulb Project specifically set to audit performance.
Since performance issues tend to be common across pages with the same templates/underlying HTML. As such, analysing a selection of key pages may be enough to find these bigger trends and opportunities for performance optimisation.
This may look like:
Setting up a Sample Audit to gather performance data from a varied sample of pages from different directories and at different crawl depths, without crawling the whole website.
Using a URL List of your key pages as the sole crawl source for your Performance audit.
Limiting the scope of your audit by using URL Inclusion and Exclusion rules to limit the crawl to specific areas of the website, like revenue-generating areas.
Remember, there are no limits to how many Projects or Audits you can create with Sitebulb.
Enable the Performance report
Once you have a clear idea of the scope of your performance Audit, go ahead and set up a new Project. Set your project name and start URL, then move on to the Audit Settings page, where you will need to:
Choose the Chrome Crawler under Crawler Settings. The Performance & Mobile Friendly report is only available when crawling with Chrome.
Enable the Performance & Mobile Friendly report under Audit Data settings
You also have the option to adjust the options found under Advanced Settings. For more on what each of these options means, and how to adjust them, see our Performance Report documentation.
Structured Data
To instruct Sitebulb to gather and report on your structured data, you will need to enable the Structured Data report under Audit Settings > Audit Data:
If your website loads structured data via JavaScript, make sure to select the Chrome Crawler so Sitebulb can render and audit the markup correctly.
International Implementation
To gather International data, you’ll need to make sure that your International Report is enabled. To build your International Report, Sitebulb will crawl your entire site, including alternate URLs found in HTML, HTTP headers, or XML sitemaps, to find and audit international hreflang and HTML annotation.
Enable your International report under Audit Settings > Audit Data:
If your hreflang markup is set up in sitemaps, you should also enable XML Sitemaps as a Crawl Source during setup, to ensure that these tags can be found and analysed by Sitebulb.
Security
To review your website's security settings with Sitebulb, enable the Security report under Audit Settings > Audit Data:
Next Steps
Once you have the basics of your audit settings in place, you can go ahead and start your crawl. Alternatively, think about any other goals or requirements you may have for your audit, and follow the resources below to dig deeper into the audit settings:
Audit Settings reference
Audit Data reference
Recommendations for crawling large websites



