微信客服
Telegram:guangsuan
电话联系:18928809533
发送邮件:[email protected]

How to Use Screaming Frog for SEO丨2025 Usage Guide

Author: Don jiang

Anyone doing Google SEO knows that tools are the leverage for efficiency. Taking Screaming Frog as an example, this crawler tool can complete 8 hours of manual work within 20 minutes: it can crawl every URL of your website and accurately locate 80-120 common SEO issues (such as 404 broken links, duplicate titles, and images missing Alt attributes).

This article takes you from installation and settings to data implementation, turning Screaming Frog into your “SEO Microscope”.

How to use Screaming Frog for SEO

Installation and Basic Settings

Installing Screaming Frog sounds like a simple “click next several times” operation. However, some users have reported that failing to pay attention to system compatibility during installation caused the Mac version to run sluggishly, with crawling speeds 40% slower than normal;

Others set the crawl depth haphazardly, resulting in small websites taking 2 hours without finishing the core pages.

Pre-installation Preparation

1. System Compatibility

Screaming Frog supports Windows 10/11 (64-bit) and macOS 10.15 and above. If your computer is Windows 7 or macOS 10.14, downloading the installer directly will prompt “Incompatible,” and forcing it to run may cause crashes (actual tests show a crash rate of about 35% for Win7 users).

2. Permission Issues

  • Windows: It is recommended to install using an administrator account (Right-click installer → “Run as administrator”), otherwise, it may fail to write crawl data due to insufficient permissions (Common error: “Unable to save log file”).
  • Mac: Do you need to disable “System Integrity Protection” (SIP)? No, but when running for the first time, you might need to click “Open Anyway” in “System Preferences → Security & Privacy,” otherwise it will be blocked (About 20% of Mac users get stuck at this step).

3. Network Environment

Turn off proxy software (such as VPNs or accelerators) before crawling. Local network latency exceeding 200ms will cause crawling speed to drop by 50% (Actual test: at 200ms latency, 10 URLs per second; at 50ms latency, 25 URLs per second).

Formal Installation

Windows System

  1. Visit the Screaming Frog official website (www.screamingfrog.co.uk), click “Download Free Version” (The free version is sufficient for small to medium websites);
  2. Select “Windows Installer” and double-click to run after the download is complete;
  3. Follow the prompts to select the installation path (default C drive is recommended to avoid losing configuration files), check “Create desktop shortcut,” and click “Install”;
  4. After installation, a green spider icon will appear on the desktop; double-click to open.

macOS System

  • Download from the same official website, selecting “macOS DMG”;
  • Double-click the downloaded .dmg file and drag the “Screaming Frog SEO Spider” icon into the “Applications” folder;
  • When opening for the first time, the system may prompt “Cannot be opened because it is from an unidentified developer.” Go to “System Preferences → Security & Privacy” and click “Open Anyway.”

4 Basic Settings

After installation, you need to configure “Spider” parameters the first time you open the software.

If set incorrectly, the subsequent crawl data may be completely useless.

User Agent

  • Function: Tells the website server “who I am.” The user agent for the Google crawler is “Mozilla/5.0 (compatible; Googlebot/2.1; +http://www.google.com/bot.html)”.
  • Setting Method: Click the top menu “Configuration → Spider,” and select “Googlebot” in the “User Agent” dropdown box (the default is “Screaming Frog”).
  • Why it’s important: If you use the default “Screaming Frog” user agent, some websites will block the crawler (e.g., setting “Disallow: /screamingfrog”), making it impossible to crawl content. Using “Googlebot” simulates the actual Google crawler to obtain data closer to reality (Actual test: After switching, the crawl success rate of an e-commerce site increased from 65% to 92%).

Crawl Depth

  • Definition: Starting from the homepage, the maximum number of clicks to follow links (e.g., Homepage → Category Page → Product Page is 3 levels).
  • Setting Recommendations:
    • Small and medium websites (URLs ≤ 1000): Set to 5 levels (covering over 90% of core pages);
    • Large websites (URLs > 1000): Set to 10 levels, but use with “Limit number of URLs” (see below) to avoid excessive crawl time (10 levels might extend crawl time from 10 minutes to 1 hour).

Limit Max URLs to Crawl

  • Function: Prevents the software from crawling indefinitely due to too many links (e.g., forums, infinite scroll pages).
  • Setting Method: In “Configuration → Spider,” check “Limit number of URLs to crawl” and enter a specific value (5,000-10,000 for small/medium sites, no more than 50,000 for large sites).
  • Consequences of not setting: A user once crawled an e-commerce site with “recommended products” dynamic links. Because no limit was set, the software crawled for 24 hours, eventually capturing 230,000 URLs (80% of which were duplicate product detail pages).

Exclude Parameters

  • Issue: Many website URLs contain redundant parameters (e.g., ?utm_source=weibo, ?page=2). These parameters do not change the content but are recognized as different URLs by Screaming Frog, leading to duplicate crawling (e.g., “Product Page” and “Product Page?page=2” are counted as 2 URLs).
  • Setting Method: Click “Configuration → Exclude,” and in “Query Parameters,” enter the parameters to filter (separated by commas), such as “utm_source,utm_medium,page”.
  • Effect: After an educational website filtered 12 tracking parameters, the number of crawled URLs decreased from 12,000 to 4,500, and the crawl time was shortened by 40%.

Run a “Mini Crawl” with the Homepage

After setting up, don’t rush to crawl the whole site—input the homepage URL and click “Start” for a small-scale test (limit the crawl to 100 URLs) to check 3 things:

  1. Whether key pages are missed: For example, whether “About Us” or “Contact Us” in the homepage navigation are captured (search keywords in the “Internal” report);
  2. Whether there are duplicate URLs: In the “URL” report, see if there are different parameter versions of the same page (e.g., “/product” and “/product?color=red”);
  3. Whether 404s are triggered: Check the 404 status codes in “Response Codes” to ensure no deleted pages are being crawled (e.g., old campaign pages).

If issues are found, return to “Configuration” to adjust parameters (e.g., increasing crawl depth or adding exclude parameters) and test again.

Quickly Start a Basic Crawl

Many think “clicking start” is all there is to crawling, but in reality, 30% of people get invalid data due to ignoring details.

For example: some start without checking the network, causing the crawl to get stuck due to high latency; some set no limits, causing the software to grab duplicates for 2 hours; others enter the wrong URL format and get “0 results.”

3 Checks Before Starting

1. Confirm Basic Settings are Complete

  • User Agent: Must be set to “Googlebot” (check in “Configuration → Spider”), otherwise you might be blocked (Actual test: without this, a corporate site success rate was only 45%; with it, it reached 90%).
  • Crawl Depth: Adjust based on site size (5 levels for small/medium, 10 for large) to avoid missing core pages or wasting time.
  • Exclude Parameters: Filter useless tracking parameters (e.g., ?utm_source) to reduce duplicate URLs (Without filtering, an e-commerce site’s URL count was 3 times the actual amount).

2. Test Network Stability

  • Latency Requirement: Local latency to the target site should ideally be ≤100ms (test with “ping target domain” command).
    • Latency ≤100ms: Can crawl 20-30 URLs per second;
    • Latency 100-200ms: Crawl 10-15 URLs per second;
    • Latency >200ms: Crawl <10 URLs per second; crawl time will double (e.g., for 1,000 URLs, low latency takes 10 mins, high latency might take 25 mins).
  • Avoid Interference: Close VPNs, accelerators, or download tools (Actual test: with Thunder running, crawl speed dropped by 60%).

3. Confirm Target Website is Accessible

  • Enter the target URL directly in the browser (e.g., https://example.com) to check if it opens normally (to avoid crawling “403 Forbidden” pages).
  • If the site has login restrictions (e.g., member systems), log out beforehand (Screaming Frog cannot handle login states and will crawl blank pages or 403 errors).

4 Steps to Operate, Get Results in 10 Minutes

1. Enter Target URL

  • Format Requirement: Must enter the full URL (including http:// or https://), otherwise the software will error “Invalid URL.”
    • Example: Correct input “https://www.example.com”, incorrect input “www.example.com” or “example.com”.
  • Multiple Domains: If you need to crawl related domains (e.g., www and m. sites), start them separately (Screaming Frog crawls one domain at a time).

2. Set Restrictions (Optional but Recommended)

  • Limit Crawl Number: In “Configuration → Spider,” check “Limit number of URLs to crawl” and enter a value (5,000-10,000 for small/medium, no more than 50,000 for large).
    • Role: Prevents infinite crawling due to dynamic links like “load more.”
  • Exclude Specific Pages: Add “Disallow” rules in “Configuration → Exclude” (e.g., “/admin/” backend pages) to avoid irrelevant content.

3. Click “Start” and Observe Real-time Status

  • Progress Bar: Top progress bar shows overall crawl progress (Green=Normal, Yellow=Slowing, Red=Stuck).
  • Status Bar: Bottom right shows “Crawled X, Remaining Y, Speed Z/sec.”
    • Normal: Speed stabilizes at 10-30 URLs/sec (at low latency);
    • Abnormal: Speed suddenly drops to 0 or 1/sec, possibly due to server limits (anti-crawl mechanism) or network issues.

4. Handling Problems Mid-crawl

  • Stuck:
    • Check network: Ping the target domain again to see if latency spiked;
    • Manual interruption: Click the “Stop” button, wait 10 seconds, and restart (some servers temporarily block IPs; restarting might help);
    • Bypass limits: If you hit “403 Forbidden,” try changing the User Agent to “Bingbot” in “Configuration → Spider” (some sites are more lenient with Bingbot).

Crawl Completion

When the crawl ends, a “Crawl Complete” prompt appears. At this point, do 3 things to confirm data quality:

1. Check if Total Crawl Count is Reasonable

  • Calculation: Small/medium sites (within 100 pages) usually crawl 50-200 URLs; larger sites (within 1,000 pages) crawl 500-3,000 (depending on link complexity).
  • Abnormalities:
    • Count=0: Likely URL format error, total network disconnection, or site blocking Googlebot;
    • Count much smaller than expected: Crawl depth too shallow (e.g., set to 2 but core page is at 3), or blocked by robots.txt (check “Robots.txt blocked” in the “Directives” report).

2. Verify if Key Pages are Crawled

  • Method: Click “Internal” in the left menu → search for core page keywords (e.g., “Product,” “About Us”) to confirm they appear in results.
  • Example: If optimizing a “New Phones” page and it’s missing, the link might be too deep or broken (404).

3. Look for High Volume of Error Status Codes

  • Key focus:
    • 404 (Broken Link): If more than 10 appear, record the URLs (export via “Response Codes” report);
    • 500 (Server Error): Single instances might be glitches; >3% requires technical troubleshooting;
    • 301/302 (Redirect): Check if the target is valid (not a 404 or irrelevant page).

SEO Report Interpretation (Focus on these 6)

SEO professionals often say “data doesn’t lie,” but among dozens of reports in Screaming Frog, the information affecting Google ranking is in these 6 reports.

We’ve found that fixing these 6 categories (excluding complex content creation) can increase indexation rates for small websites from 65% to 85%, with average organic traffic growth of 20%.

Response Codes Report

This report logs the HTTP status code of every page. If the status code is wrong, crawlers might skip your page entirely.

Key Data and Actions

  • 200 (OK): Should be >85% (for SMB sites). If lower than 80%, many pages may be blocked or content-errored.
  • 404 (Broken Link): Common when deleted pages aren’t cleaned up (E-commerce sites often see 8-12% 404s).
    • Action: Export 404 URL list → check link source (Nav/Internal/External) → delete invalid links or set 301 redirects to relevant pages.
  • 301/302 (Redirect): Be wary if >5% (could be un-updated legacy pages).
    • Action: Check if target is valid; prioritize 301 permanent redirects to pass equity.
  • 500 (Server Error): Single occurrences might be glitches; >3% needs tech investigation.

Example: A corporate site fixed 12 404 links and deleted internal links pointing to them; daily crawl volume rose from 800 to 1,200.

URL Length and Structure Report

Google’s patience for long URLs is limited. The longer the URL, the lower the probability of being fully crawled.

Key Data and Actions

  • Length Distribution: Statistics show about 20-30% of URLs exceed 100 characters (Ideal is <80).
    • Action: Filter “Length > 100” URLs → shorten paths (e.g., change “/product?id=123” to “/red-running-shoes-123”).
  • Dynamic Parameters: URLs with 3+ parameters (e.g., “?id=123&cat=456&sort=date”) >15% need optimization.
    • Action: Consolidate parameters or use static links.

Comparison: An e-commerce site changed “/product?category=shoes&brand=nike&id=123” (102 chars) to “/nike-shoes-123” (45 chars); the index status changed from “Not indexed” to “Indexed”.

Page Titles Report

Titles are core to Google’s judgment of page themes. Duplicate or invalid titles directly lower rankings.

Key Data and Actions

  • Duplicate Rate: About 30-40% of pages have duplicate titles (e.g., multiple e-commerce product pages all titled “Product Details”).
    • Action: Filter “Duplicate Titles” → add unique identifiers for each page (e.g., “[Product Name]-[Brand]”).
  • Length Distribution: Ideal length is 50-60 chars (Google truncates at 600px, roughly 60 chars).
    • Action: Filter “Length > 60” → shorten content (keep core keywords, remove fluff).

Case: An education site changed a title from “Course Intro” to “2024 Python Entry Course – XX Education (with Study Materials)”; the page CTR rose from 1.2% to 2.1%.

Meta Description Report

Meta descriptions don’t directly impact rank but determine whether users click your page.

Key Data and Actions

  • Missing Rate: About 15-20% of pages lack descriptions (crawlers generate them from content, but quality is inconsistent).
    • Action: Filter “No Meta Description” → write manually (target 150-160 chars).
  • Length Distribution: About 25% exceed 160 chars (truncated), 10% are too short (<120 chars).
    • Action: Filter “Length > 160” or “< 120" → add value-driven info (e.g., "30-day free trial," "Genuine guarantee").

Data: An e-commerce site optimized 200 product descriptions; organic clicks grew by an average of 15%.

H1 Tag Report

H1 is the main header; Google uses H1 to judge core content (Ideally one H1 per page).

Key Data and Actions

  • Quantity Anomalies: About 10-15% of pages have no H1; 5% have multiple H1s (confusing theme).
    • Action: Filter “No H1” or “Multiple H1s” → add main titles to H1-less pages; remove extra H1 tags.
  • Relevance: About 30% of H1s don’t match content (e.g., H1 says “Summer Sale” but page is for winter coats).
    • Action: Filter “Content Mismatch” → modify H1 to be consistent.

Effect: A clothing brand optimized 100 H1s; average dwell time increased from 45 to 70 seconds.

Image Alt Attribute Report

Alt attributes are text descriptions for images; missing or keyword-stuffed Alts waste image search traffic (About 30% of users find content via image search).

Key Data and Actions

  • Missing Rate: About 40-50% of images lack Alt attributes (especially product/detail images).
    • Action: Filter “No Alt Text” → add descriptions (e.g., “Red sneakers side breathable mesh close-up”).
  • Keyword Stuffing: About 10-15% of Alts contain repetitive keywords.
    • Action: Filter “Keyword Stuffing” → change to natural descriptions.

Case: A sports brand added specific Alts to 200 product images; image search traffic grew by 25%.

Batch Check Internal Link Issues

We’ve found that sites not performing batch internal link checks have 15-20% of pages failing effective indexation. After fixing these, crawl volume for those pages can rise by over 30%.

Batch checking isn’t “looking at links one by one,” but using the “Internal” report in Screaming Frog.

Broken Internal Links

Broken internal links point to deleted or inaccessible pages (404). Users click these and bounce; crawlers reduce crawling when encountering frequent 404s.

Data and Actions

  • Common Sources: Nav bars (30-40%), old article recommendations (25-30%), user comments (15-20%).
  • Detection Method:
    1. Click “Internal” in left menu → filter “404” in “Status Code” column;
    2. Export results (Right-click → Export) to track “Source URL” and “Target URL.”

Example: An education site had 12 “Popular Courses” links in the nav; 8 pointed to 404s. Removing these increased daily crawl count for the host page from 150 to 220.

Fix Actions

  • Delete broken links;
  • Replace with valid links;
  • Set 301 redirects if the target page must be preserved.

Orphan Pages

Orphan pages have content but no internal links pointing to them (“Incoming Links=0”). Crawlers only find these via external links or direct URL input; index probability is 60% lower than pages with internal links.

Data and Actions

  • Common Types:
    • Temporary campaign pages (not deleted after event);
    • Test pages (unreleased features);
    • Low-quality duplicate product parameter pages.
  • Detection Method:
    1. Filter “Linked From=0” in “Indexability” report;
    2. Or filter “Incoming Links=0” and “Word Count > 100” in “Internal” report.

Data: An e-commerce site found 200 orphan pages (mostly old products). After adding internal links, indexation rates for these rose from 15% to 70%.

Fix Actions

  • Add internal links to high-value orphan pages;
  • Delete or block low-value orphan pages via robots.txt;
  • Periodically check for new orphans after weekly crawls.

Equity Concentration

This occurs when homepages or core pages have too many links (e.g., footer with 50 links), causing crawlers to “scatter energy,” meaning other important pages (product pages, blog posts) get fewer crawl opportunities.

Data and Actions

  • Typical Sign: Homepage “Outgoing Links” exceeds 50 (ideal is 20-30);
  • Quantified Impact: A home decor site with 68 homepage links saw core product crawl depth go from 2 levels to 4, reducing daily crawl volume by 40%.

Detection Method

  1. Sort by “Outgoing Links” in the “Internal” report;
  2. Check link counts on home and category pages.

Fix Actions

  • Streamline non-core links (move “Contact Us” to footer, keep 5-8 core sections on home);
  • Move secondary links to “More” dropdown menus;
  • Increase internal links to core pages (top sellers, high-conversion articles).

3 Tips for Batch Processing

  1. Use Excel to filter high-frequency issues: Export data and use Excel to locate sources pointing repeatedly to 404s.
  2. Prioritize high-equity pages: Fix links on the home and category pages first as they have the widest impact.
  3. Regular re-checks: Crawl every two weeks to compare data and ensure the internal structure stays healthy.

Finally, tools are only aids; the core of Google ranking will always be “content users need.”

滚动至顶部