Recovering from Algorithm Updates
Diagnosing the actual cause, recovery playbooks, the patience the timeline demands, and the rewrite-vs-redirect-vs-delete decision.
Most algorithm-recovery work fails for one reason: the team starts changing things before they have proven what is wrong. This module is the diagnostic-first playbook that distinguishes the sites that recover within two core updates from the sites that never come back.
TL;DR
- Diagnose before you operate. A traffic drop is not a diagnosis. The cause is one of: content quality, link profile, technical decay, query-intent shift, site-reputation-abuse policy, or genuine SERP competition. Each has a different recovery path.
- Recovery is gated by the next core update. The classifier that demoted you typically refreshes only during core updates. Expect 60 to 270 days between fix and recovery, with August and November core updates being the most common partial-recovery windows for HCU-affected sites.
- Rewrite, redirect, or delete is not a stylistic preference. It is a decision driven by content equity, query value, and user intent overlap. Get the decision wrong and you either keep the demotion (rewrite that should have been deleted) or lose link equity (delete that should have been redirected).
The mental model
Algorithm recovery is like wound triage in an emergency room. You do not start suturing before you know whether the patient has internal bleeding. The drop in traffic is the symptom; the algorithmic cause is the bleed.
A tier-one trauma team works in a strict sequence: assess airway, breathing, circulation, then move to the secondary survey. SEO recovery has the same discipline. Phase 1 is correlation: match the drop window to the update. Phase 2 is segmentation: identify which page clusters, query clusters, and traffic surfaces (Web, Discover, Images, News) were affected. Phase 3 is hypothesis: name the specific signal that demoted you. Phase 4 is intervention: ship a fix targeted at that signal. Phase 5 is patience: wait for the next refresh of the relevant classifier.
The fatal mistake is jumping from “traffic dropped” straight to “intervention.” Every intervention without a hypothesis is a coin flip, and core-update refreshes only happen 4–8 times per year, so each wasted cycle costs you 45–90 days.
Deep dive: the 2026 reality
The recovery landscape changed materially in 2024–2025. Three shifts you need to internalize:
Shift 1: HCU is now per-cluster. The original 2022/2023 Helpful Content Update was a binary site-wide classifier — once flagged, every page was demoted. After the March 2024 core merge, the helpful-content signal evaluates page clusters (e.g., /blog/finance/, /reviews/laptops/, /glossary/) independently. This is why some HCU-affected sites saw partial recoveries in August 2024 and November 2024 on specific clusters while others stayed flat.
Shift 2: Recovery requires deletion at scale, not just rewriting. Google’s own communications around the March 2024 update specifically called out scaled content abuse. For sites with hundreds of thin or AI-spun pages, the only successful recovery pattern observed in 2024–2025 was aggressive pruning — frequently 40–70% of the URL set deleted or no-indexed. Sites that “rewrote” without pruning have stayed flat or worsened.
Shift 3: AI Overviews citation changes everything. Since the June 2025 core update, sites cited in AI Overviews are receiving residual authority signals. This means a recovery now has two visible markers: classic blue-link rank recovery and AIO citation recovery. They can move independently. Recovery is no longer just “keywords come back”; it is also “we are cited in the AI synthesis.”
The diagnostic decision tree
| Drop signature | Likely cause | First diagnostic |
|---|---|---|
| Sharp, single-day, 30%+ across most pages | Manual action or core update | Check GSC Manual Actions report; cross-reference Search status dashboard |
| Gradual decline over 2–4 weeks aligning with named update | Core or HCU classifier | Cluster-level performance segmentation |
Specific URL pattern (e.g., /coupons/) hit | Spam policy or vertical re-evaluation | Compare to Google’s spam policy updates |
| Discover collapse with stable Web traffic | Discover refresh | News/Discover-specific signals |
| Branded queries fine, non-brand collapsed | Quality classifier (HCU lineage) | Page-quality audit on top losers |
| Drop coincides with a competitor’s launch | Genuine SERP competition | Manual SERP comparison |
| Drop right after a site migration | Technical regression | Crawl diff before/after |
| Drop in one country only | hreflang or geo-policy issue | International segmentation |
Recovery timelines, observed
Based on cohort data from BrightEdge, Sistrix, and Glenn Gabe’s case studies through Q1 2026:
| Cause | Median recovery window | Notes |
|---|---|---|
| Core-update content demotion (single cluster) | 60–90 days | Often partial in next core, full in second |
| HCU-class site-wide demotion | 180–270 days | Many sites never fully recover; survivor sites pruned heavily |
| Penguin / link-spam | 90–180 days after disavow | Faster after SpamBrain real-time detection |
| Manual action (resolved correctly) | 7–30 days after reconsideration approval | Reconsideration covered in module 99 |
| Site-reputation abuse (parasite SEO) | Indefinite while violating | Recovery only after policy compliance |
| Technical regression | 7–30 days after fix | Fastest of all categories |
Rewrite vs. redirect vs. delete
This is the call that determines whether your recovery effort works.
| Option | When to choose | Risk |
|---|---|---|
| Rewrite | URL still has link equity, query has commercial value, intent overlap with your business >70% | Slow; rewrite must be substantially different to count |
| 301 redirect | URL has equity but topic is consolidated into another, equivalent page exists | Diluted equity if target is not topically aligned |
| Delete (410) | Thin, AI-spun, no equity, no business value, no intent fit | Lose any residual signal; simplest classifier signal |
| Noindex (kept live) | Useful for users but not for search; product pages, internal docs | Crawl budget cost remains; not a true delete |
| Consolidate (canonical) | Two pages compete for the same intent | Wrong canonical can cause both to lose ranking |
The 2026 default for HCU recovery: delete with 410 for any page that scores below a quality threshold and has no inbound links. Rewrite only when the page demonstrably retains equity. Hard but proven across hundreds of recovery cases.
Visualizing it
flowchart TD
A["Traffic drop detected"] --> B{"Match to update timeline?"}
B -->|"No"| C["Investigate technical, manual, or competitor"]
B -->|"Yes"| D["Segment by cluster and query"]
D --> E{"Site-wide or cluster-specific?"}
E -->|"Site-wide"| F["HCU-class hypothesis: prune aggressively"]
E -->|"Cluster"| G["Targeted rewrite or delete in cluster"]
F --> H["Audit every URL: keep, rewrite, redirect, delete"]
G --> H
H --> I["Ship changes in batches"]
I --> J["Wait for next core update refresh"]
J --> K{"Recovery signal?"}
K -->|"Partial"| L["Repeat audit on stragglers"]
K -->|"None"| M["Reassess hypothesis"]
K -->|"Full"| N["Document playbook"]
L --> J
M --> D
Bad vs. expert
The bad approach
Recovery plan after September 2023 HCU:
1. Refresh 30 top-traffic articles
2. Add an author bio box
3. Improve internal linking
4. Build 10 new backlinks
5. Wait for Google to re-evaluate
This fails for HCU-class demotions because it does not address the actual signal. The classifier looks at the entire site’s helpful-content profile. Polishing the top 30 pages while leaving 800 thin AI-generated pages live keeps the site flagged. An author bio without expertise depth does not move the E-E-A-T needle. The team waits 9 months and recovers nothing.
The expert approach
# recovery-audit.py — content equity scoring for prune decisions
import csv
from datetime import datetime, timedelta
def score_url(metrics):
score = 0
# Inbound links from referring domains
if metrics['referring_domains'] >= 3:
score += 30
# Recent organic clicks in last 90 days
if metrics['clicks_90d'] >= 50:
score += 25
# Word count gate
if 600 <= metrics['word_count'] <= 4000:
score += 10
# First-hand evidence markers
if metrics['has_original_image']:
score += 10
if metrics['has_author_byline']:
score += 5
if metrics['has_first_person_evidence']:
score += 15
# AI generation penalty
if metrics['detected_ai_unedited']:
score -= 20
# Last meaningful update
days_old = (datetime.now() - metrics['last_updated']).days
if days_old > 730:
score -= 10
return score
def decide(score, metrics):
if score >= 50:
return 'KEEP'
if score >= 25 and metrics['referring_domains'] >= 1:
return 'REWRITE'
if metrics['referring_domains'] >= 2 and metrics['has_close_topical_match']:
return 'REDIRECT_301'
return 'DELETE_410'
This works because the decision is driven by data — equity, traffic, evidence — not by sentiment. Sites that pruned 40–70% of low-scoring URLs and rebuilt the rest with first-hand evidence are the ones that came back in 2024–2025. The audit must run end to end before any changes ship.
Do this today
- In Google Analytics 4, go to Reports → Acquisition → Traffic acquisition. Add a date comparison spanning 30 days before and after the suspected update. Export organic-search sessions per page-path-prefix.
- In Google Search Console → Performance, segment by Query and by Page. Identify the top 50 query and page losers. Tag each with its cluster prefix.
- In Ahrefs Site Explorer, run “Top pages: Lost” for the affected window. Cross-reference with GSC losers.
- Run Screaming Frog on the full site with a custom extraction for word count, author byline, structured data, and last-modified. Export to CSV.
- Combine the three exports in a sheet. Score every URL using the rubric in the expert section. Sort by score ascending.
- Build a single decision sheet: column A URL, column B current score, column C decision (KEEP / REWRITE / REDIRECT / DELETE), column D rationale.
- Ship deletes first as HTTP 410 Gone (not 404). Use a sitemap of deleted URLs in GSC → Sitemaps to accelerate processing. Submit via Bing Webmaster Tools as well.
- Ship redirects in batches with verified topical match. Validate with Screaming Frog’s redirect chain checker that no redirect lands on a 404 or another redirect.
- Schedule rewrites with a content team brief that requires first-hand evidence (original photography, expert quote, original data, or testing notes). No rewrite ships without one.
- Set a recovery monitoring dashboard in Looker Studio pulling from GSC, with traffic-by-cluster, AIO citation tracking via Profound or Athena HQ, and a marker line for each Google update from the Search Status Dashboard.
Mark complete
Toggle to remember this module as mastered. Saved to your browser only.
More in this part