Screaming Frog SEO Spider Update – Version 21.0

Dan Sharp

Posted 12 November, 2024 by in Screaming Frog SEO Spider

Screaming Frog SEO Spider Update – Version 21.0

We’re delighted to announce Screaming Frog SEO Spider version 21.0, codenamed internally as ‘towbar’.

This update contains new features and improvements based upon user feedback and as ever, a little internal steer.

So, let’s take a look at what’s new.


1) Direct AI API Integration

In our version 20.0 release we introduced the ability to connect to LLMs and query against crawl data via custom JavaScript snippets.

In this update, you’re now able to directly connect to OpenAI, Gemini and Ollama APIs and set up custom prompts with crawl data.

You can configure up to 100 custom AI prompts via ‘Config > API Access > AI’.

Direct AI Integration with OpenAI

You’re able to select the category of model, the AI model used, content type and data to be used for the prompt such as body text, HTML, or a custom extraction, as well as write your custom prompt.

The SEO Spider will auto-control the throttling of each model and data will appear in the new AI tab (and Internal tab, against your usual crawl data).

AI Tab results

In a similar way as custom JS snippets, this can allow you to create alt text at scale, understand the language of a page, detect inappropriate content, extract embeddings and more.

The ‘Add from Library’ function includes half a dozen prompts for inspiration, but you can add and customise your own.

OpenAI Add From Library

The benefits of using the direct integration over custom JS snippets are –

  • You can input your API key once for each AI platform, which will be used for all prompts.
  • You don’t need to edit any JavaScript code! You can just select requirements from dropdowns and enter your prompt into the relevant field.
  • JavaScript rendering mode isn’t required, data can be returned through any crawl mode.
  • The APIs are automatically throttled as per their requirements.

This new AI integration should make it even more efficient to create custom prompts when crawling. We hope users will utilise these new AI capabilities responsibly for genuine ‘value-add’ use cases.


2) Accessibility

You can now perform an accessibility audit in the SEO Spider using the open-source AXE accessibility rule set for automated accessibility validation from Deque.

This is what powers the accessibility best practices seen in Lighthouse and PageSpeed Insights. It should allow users to improve their websites to make them more inclusive, user friendly and accessible for people with disabilities.

Accessibility can be enabled via ‘Config > Spider > Extraction’ (under ‘Page Details’) and requires JavaScript rendering to be enabled to populate the new Accessibility tab.

Accessibility Config

The Accessibility tab details the number of accessibility violations at different levels of compliance based on the Web Content Accessibility Guidelines (WCAG) set by the W3C.

Accessibility Tab

An accessibility score for each page can also be collected by connecting to Lighthouse via PageSpeed Insights (‘Config > API Access > PSI’).

WCAG compliance levels build upon each other and start from WCAG 2.0 A to 2.0 AA, then 2.0 AAA before moving onto 2.1 AA and 2.2 AA. To reach the highest level of compliance (2.2 AA), all violations in previous versions must also be achieved.

The Accessibility tab includes filters by WCAG with over 90 rules within them to meet that level of compliance at a minimum.

Accessibility Tab filters in Overview tab

The right-hand Issues tab groups them by accessibility violation and priority, which is based upon the WCAG ‘impact’ level from Deque’s AXE rules and includes an issue description and further reading link.

Accessibility issues in the right hand Issues tab

The lower Accessibility Details tab includes granular information on each violation, the guidelines, impact and location on each page.

Accessibility Details tab

You can right-click on any of the violations on the right-hand side, to ‘Show Issue in Browser’ or ‘Show Issue In Rendered HTML’.

All the data including the location on the page can be exported via ‘Bulk Export > Accessibility > All Violations’, or the various WCAG levels.

Accessibility Bulk Exports

There’s also an aggregated report under the ‘Reports’ menu.


3) Email Notifications

You can now connect to your email account and send an email on crawl completion to colleagues, clients or yourself to pretend you have lots of friends.

This can be set up via ‘File > Settings > Notifications’ and adding a supported email account.

Email Notifications

You can select to ‘Email on Crawl Complete’ for every crawl to specific email address(es).

Crawl complete emails

So many friends.

Alternatively, you can send emails for specific scheduled crawls upon completion via the new ‘Notifications’ tab in the scheduled crawl task as well.

Email Notifications from scheduled crawls

The email sent confirms crawl completion and provides some top-level data from the crawl.

Email Notification Delivered

We may expand this functionality in the future to include additional data points and data exports.

Please read about notifications in our user guide.


4) Custom Search Bulk Upload

There’s a new ‘Bulk Add’ option in custom search, which allows you to quickly upload lots of custom search filters, instead of inputting them individually.

Bulk Upload Custom Search

If you’re using this feature to find unlinked keywords for internal linking, for example, you can quickly add up to 100 keywords to find on pages using ‘Page Text No Anchors’.

Custom search bulk upload filters

Please see our ‘How to Use Custom Search‘ tutorial for more.


Other Updates

Version 21.0 also includes a number of smaller updates and bug fixes.

  • Additional crawl statistics are now available via the arrows in the bottom right-hand corner of the app. Alongside URLs completed and remaining, you can view elapsed and estimated time remaining, as well as crawl start time date and time. This data is available via ‘Reports > Crawl Overview’ as well.
  • Custom Extraction has been updated to support not just XPath 1.0, but 2.0, 3.0 and 3.1.
  • Scheduling now has ‘Export’ and ‘Import’ options to help make moving scheduled crawl tasks less painful.
  • The Canonicals tab has two new issues for ‘Contains Fragment URL’ and ‘Invalid Attribute In Annotation’.
  • The Archive Website functionality now supports WARC format for web archiving. The WAR file can be exported and viewed in popular viewers.
  • You can now open database crawls directly via the CLI using the –load-crawl argument with the database ID for the crawl. The database ID can be collected in the UI by right-clicking in the ‘File > Crawls’ table and pasting it out, or viewed in the CLI using the cli –list-crawls argument.
  • There’s a new right click ‘Show Link In Browser’ and ‘Show Link in HTML’ option in Inlinks and Outlinks tab to make it more efficient to find specific links.

That’s everything for version 21.0!

Thanks to everyone for their continued support, feature requests and feedback. Please let us know if you experience any issues with this latest update via our support.

Small Update – Version 21.1 Released 14th November 2024

We have just released a small update to version 21.1 of the SEO Spider. This release is mainly bug fixes and small improvements from the latest major release –

  • Fixed issue with custom database locations not being picked up.
  • Fixed bug in OpenAI Tester.
  • Fixed a couple of crashes, including for users that had ‘auto connect’ selected for the old GA API, which hasn’t been available for sometime (and is now removed!).

Dan Sharp is founder & Director of Screaming Frog. He has developed search strategies for a variety of clients from international brands to small and medium-sized businesses and designed and managed the build of the innovative SEO Spider software.

11 Comments

  • javi 1 week ago

    happy 21!

    Reply
  • Stefan 1 week ago

    Hello everyone!

    The accessibility part is brilliant!
    Especially for the Europeans next year.

    I tried to bring the data into the scheduled crawl (as gsheet in gdrive), but all files are empty?!
    What am I doing wrong? or is it currently not working?

    Greets

    Stefan

    Reply
    • screamingfrog 1 week ago

      Hi Stefan,

      That’s awesome to hear!

      For accessibility to run, you’d need to supply a saved config in scheduling that has both acccessibility and JavaScript rendering enabled (both mentioned above in the post).

      So that’s the first place I’d check!

      Please send any issues into us via support though (https://www.screamingfrog.co.uk/seo-spider/support/).

      Cheers,

      Dan
      Dan

      Reply
  • Ahmet Çadırcı 1 week ago

    The addition of AI to the programme has been incredible. Thank you for your work.

    Reply
  • Alex 1 week ago

    Accessibility audit in the SEO Spider is something I waited for a long time. Great news!

    Reply
  • Dominik Stein 1 week ago

    Amazing work, as always, @ScreamingFrogTeam :)
    The Direct AI API Integration will save us a huge amount of time and also the Email Notifications give us dozens of new possibilities to trigger certain events in the future.
    Love it! Great update.

    Reply
  • Appreciate it, guys!!

    Reply
  • Fabian Auler 6 days ago

    Wow, great work again. I have been waiting for this update. I like this small update too: “Email Notifications”. Because we are working in a bigger team and will use these notifications for our members.

    Best
    -Fabian

    Reply
  • Rachael Bews 3 days ago

    Hello! I’m having big problems using the import vs open function (which is no longer available to me in 21.1) – is this just me? Its taking significantly longer to import crawls than the time it used to take to simply open them?

    Reply
    • screamingfrog 3 days ago

      Hi Rachael,

      If you’re importing an old .seospider crawl file, it does take longer to convert it into the database format and open it.

      After it has been imported though, it will open instantly. Obviously any crawls run in database storage mode, will also open instantly.

      So it is only for crawls in the older format where you have previously had to ‘save’ crawl files. As mentioned on support, you could switch back to memory storage mode to open them if you prefer.

      Cheers

      Dan

      Reply
  • Yerai Lorenzo 3 days ago

    Accesibility audit is an amazing upgrade, but what I love the most is the custom search upload in bulk. This makes data extraction way easier than before :)

    Reply

Leave A Comment.

Back to top