IP Infusion Documentation: Google Search Remediation Plan
Prepared by: Mighty Digital
For: IP Infusion Technical Publications
Date: April 2026
Status: Draft for Review
Executive Summary
Unintended legacy documentation URLs from docs.ipinfusion.com (the Digital Ocean/Legacy server) have been indexed by Google and are surfacing in search results alongside—and sometimes instead of—the preferred CURRENT documentation at documentation.ipinfusion.com. This document analyzes the problem, prioritizes remediation actions, and provides detailed technical recommendations, resource links, and a proposed division of labor between IP Infusion and Mighty Digital.
The two primary goals are:
- Remove or suppress specific legacy URLs from the Google Search index
- Curate and steer search results toward preferred, current documentation links
1. Understanding the Two Servers
| Legacy Server | Current Server | |
|---|---|---|
| Domain | docs.ipinfusion.com | documentation.ipinfusion.com |
| Hosting | Digital Ocean (Mighty Digital managed) | Hosted MadCap Flare service |
| URL Pattern | docs.ipinfusion.com/service-provider/index.html#page/OcNOS-SP/... | documentation.ipinfusion.com/ocnos-sp-layer-2-7.0/Content/Home.htm |
| Control Level | Full (file system, Nginx config, custom scripting) | Dependent on MadCap Flare platform |
| Primary Action | Deindex, redirect, restrict | Optimize, signal authority |
The Legacy server is fully under Mighty Digital's control. This is where the most powerful technical interventions can be made. The Current server is hosted on a third-party platform (MadCap Flare), which limits direct server-level changes, though page-level metadata and sitemap management may still be accessible.
2. Root Cause Analysis
Before remediating, it is important to understand how legacy URLs entered the Google index. Likely causes include:
- Google Programmable Search Engine (PSE): If a PSE was previously configured with
docs.ipinfusion.comas a source, Googlebot may have followed and indexed those pages as part of normal crawling. Investigate whether a PSE still references the legacy domain. - External inbound links: If third-party sites, forums, vendor documentation, or older IP Infusion marketing pages linked to
docs.ipinfusion.com, Googlebot will have followed those links and indexed the destination pages. - No robots.txt or permissive crawling rules: If the legacy server has no
robots.txt(or an overly permissive one), Googlebot has had unrestricted access to all content. - Sitemaps submitted or discoverable: If an XML sitemap referencing legacy URLs was ever submitted to Google Search Console, those URLs were explicitly invited into the index.
Action item: Audit the Google Search Console account for docs.ipinfusion.com to check submitted sitemaps, crawl history, and coverage reports.
3. Inventorying Legacy URLs in the Google Index
Before removing anything, you need to know what is indexed. This is the necessary first step.
3.1 Google Search Operators
Use Google's site: operator to enumerate indexed legacy URLs:
site:docs.ipinfusion.com
site:docs.ipinfusion.com/service-provider
site:docs.ipinfusion.com/sp-6.3.3-configuration-guide
Record results in a spreadsheet, noting: URL, page title, approximate version (e.g., 6.3, 6.3.3), and whether a corresponding current URL exists. This becomes your master remediation inventory.
3.2 Google Search Console Coverage Report
In Google Search Console (GSC), navigate to Index > Pages for the docs.ipinfusion.com property (if verified). The coverage report shows all URLs Google has crawled and indexed, including those it chose not to index. Export this list.
GSC Documentation: Google Search Console Help Center
Coverage Report Guide: Understand the Index Coverage report
3.3 Third-Party Crawl Tools
Supplement GSC data with a full crawl of docs.ipinfusion.com using one of these tools:
- Screaming Frog SEO Spider (free up to 500 URLs, paid for larger sites): Crawls the site and exports all discovered URLs, response codes, redirect chains, and meta tags. screaming frog.co.uk
- Sitebulb: Similar crawler with visual reporting. sitebulb.com
- Ahrefs or Semrush Site Audit: Can also reveal which legacy URLs have external backlinks pointing to them (important for redirect planning). ahrefs.com / semrush.com
3.4 LLM/AI Search Audit
Since LLMs like ChatGPT are already surfacing incorrect legacy URLs (as shown in the example in the project brief), manually test prompts that your users are likely to ask:
- "OcNOS Path MTU IPv6 configuration"
- "IP Infusion OcNOS SP 6.3 documentation"
- "OcNOS OSPFv3 configuration guide"
Catalog which URLs appear, note the version, and check whether they are legacy or current. This helps prioritize which URLs to remediate first based on actual user impact.
4. Removing Legacy URLs from the Google Index
Use a layered approach. No single technique is sufficient on its own; they work together and reinforce each other over time.
4.1 Layer 1: Immediate Suppression — Google Search Console Removal Tool
What it does: Hides specific URLs from Google Search results for approximately 180 days (6 months). It does not permanently remove them from the index.
When to use it: For the highest-priority legacy URLs surfacing right now. Use this first for immediate relief while permanent solutions are implemented.
How to use it:
- Log in to Google Search Console for
docs.ipinfusion.com - Navigate to Removals (left sidebar)
- Click New Request → Temporarily remove URL
- Enter individual URLs, or use "Remove all URLs with this prefix" for entire path patterns (e.g.,
/sp-6.3.3-configuration-guide/)
⚠️ Important: The Removals tool is a temporary measure only. Without follow-up permanent actions (Layers 2 and 3 below), removed URLs will reappear after 180 days.
Official documentation: Google Removals Tool Help
4.2 Layer 2: Permanent Deindexing — noindex Meta Tags
What it does: Instructs Googlebot to remove a page from the index permanently on its next crawl. This is the most reliable permanent deindexing signal available.
Since the Legacy server is a flat HTML file system on Digital Ocean, noindex tags can be injected in two ways:
Option A — Add to individual HTML files:
Insert into the <head> section of each file to be deindexed:
<meta name="robots" content="noindex, nofollow">
Option B — Inject via Nginx HTTP header (more scalable):
Configure the Nginx server to emit an X-Robots-Tag header for specific path patterns:
location ~* ^/(sp-6\.3\.3|sp-6\.2|sp-5\.) {
add_header X-Robots-Tag "noindex, nofollow";
}
This approach is highly scalable because it applies to entire version directories without modifying individual HTML files. It requires access to the Nginx configuration on the Digital Ocean server.
Google's official guidance on noindex: Remove Your Site Info from Google
X-Robots-Tag specification: Robots meta tag, data-nosnippet, and X-Robots-Tag
⚠️ Critical note: Do NOT use
robots.txtDisallowas the primary deindexing method. Google dropped support fornoindexdirectives insiderobots.txtin 2019. ADisallowrule prevents crawling, but pages already in the index will remain there because Googlebot can no longer see thenoindextag to process the removal.
4.3 Layer 3: HTTP Status Codes — 410 Gone
For legacy content that has no successor and should simply cease to exist:
410 Gone tells Google definitively that the resource has been permanently removed and will not return. This is stronger than a 404 (which could be interpreted as a temporary absence) and signals to Google to remove the URL from the index faster.
Configure Nginx to return 410 for deprecated version paths:
# Return 410 for EOL documentation versions
location ~* ^/(sp-5\.|sp-4\.) {
return 410;
}
Reference: Google on 404 and 410 status codes
4.4 Layer 4: Robots.txt — Block Future Crawling
While robots.txt alone will not remove already-indexed content, it prevents Googlebot from re-crawling the legacy server and discovering new or re-appearing content.
Add or update /robots.txt on docs.ipinfusion.com:
User-agent: *
Disallow: /
# Allow Googlebot to crawl the robots.txt itself (automatic)
To block by specific version directories only (less aggressive):
User-agent: *
Disallow: /sp-6.3.3-configuration-guide/
Disallow: /sp-6.2/
Disallow: /service-provider/
Note: A full
Disallow: /is appropriate if the Legacy server is intended to serve no public traffic and all content has migrated or been deprecated.
Google Robots.txt documentation: Robots.txt Introduction and Guide
Robots.txt tester (GSC): Available under Settings in Search Console
4.5 Layer 5: 301 Redirects to Current Content
For legacy URLs that correspond to valid, current content on documentation.ipinfusion.com, the ideal solution is a permanent 301 redirect. This:
- Sends users arriving on old links to the correct destination
- Transfers any residual SEO "link equity" from the old URL to the new one
- Signals to Google that the canonical version is now at the new location
Nginx map-based redirect approach (scalable for large URL sets):
# Load a map file of old -> new URL mappings
map $request_uri $new_url {
include /etc/nginx/redirects/ipi_docs_redirects.map;
default "";
}
server {
listen 80;
server_name docs.ipinfusion.com;
if ($new_url != "") {
return 301 $new_url;
}
# Fallback: redirect unmapped URLs to the documentation home
location / {
return 301 https://documentation.ipinfusion.com/home/Content/LibraryPages/Library.htm;
}
}
The .map file would contain entries like:
/sp-6.3.3-configuration-guide/OcNOS-SP/Layer3-Config/OSPF_Configuration.054.07.html https://documentation.ipinfusion.com/ocnos-sp-layer-2-7.0/Content/OSPF-equivalent.htm;
For content without a direct equivalent: Redirect to the product documentation home page or the nearest category landing page.
Nginx redirect documentation: Creating NGINX Rewrite Rules
DigitalOcean Nginx redirect tutorial: How To Create Temporary and Permanent Redirects with Nginx
5. Curating Search Results — Steering Users to Preferred Links
Removing bad links is only half the job. The other half is ensuring the right content ranks well and is clearly authoritative.
5.1 XML Sitemaps for the Current Server
Ensure that documentation.ipinfusion.com has a well-structured, up-to-date XML sitemap submitted to Google Search Console. The sitemap should:
- Include only preferred, current release URLs
- Exclude deprecated or non-canonical pages
- Be updated automatically when new content is published (MadCap Flare may support this natively)
- Use
<lastmod>tags to signal freshness
Submit the sitemap at: GSC → Sitemaps
Sitemap documentation: Build and submit a sitemap
5.2 Canonical Tags
If any legacy content remains accessible (not yet 410'd or redirected), use canonical tags to point Google to the preferred version:
<link rel="canonical" href="https://documentation.ipinfusion.com/ocnos-sp-layer-2-7.0/Content/Home.htm">
This can be added to legacy HTML files or emitted via the Nginx Link header.
5.3 On-Page SEO for Current Documentation
Work upstream in the MadCap Flare publishing workflow (or inject via post-processing) to improve on-page signals on documentation.ipinfusion.com:
- Title tags: Should follow the pattern:
[Topic] — OcNOS [Version] [Product] Documentation | IP Infusion. Avoid generic titles like "Untitled" or "Page 1." - Meta descriptions: 150–160 characters summarizing the content; should include product name, version, and key topic.
- Heading structure: H1 should match the page topic; H2/H3 for sub-sections.
- Structured data / Schema.org: Consider adding
TechArticleorArticleschema markup to documentation pages to improve rich snippet eligibility.
Google's SEO Starter Guide: Search Engine Optimization (SEO) Starter Guide
5.4 Google Search Console — Performance Monitoring
Use the Performance report in GSC for documentation.ipinfusion.com to:
- Identify which queries are driving traffic to current docs
- Identify queries where the legacy site still outranks current docs (these are priority targets)
- Track impressions and click-through rates over time to measure remediation progress
GSC Performance report: Search Console Performance Report
5.5 Sponsored Search / Google Ads (Paid Placement)
The project brief raises the question of paid search as part of the strategy. This is a viable, though ancillary, tactic.
When it makes sense:
- For high-traffic navigational queries (e.g., "OcNOS documentation," "IP Infusion SP 7.0 config guide") where legacy URLs consistently outrank current ones in the interim
- As a short-term bridge while organic remediation takes effect (Google deindexing can take weeks to months)
Approach:
- Bid on exact-match branded queries:
[ip infusion ocnos documentation],[ocnos sp configuration guide] - Use ad copy that clearly indicates the link goes to official, current documentation
- Link directly to preferred version landing pages
Caveats:
- Paid ads are a recurring expense; organic remediation is the durable solution
- Ads do not appear in all contexts (mobile apps, AI-powered search, etc.)
Google Ads Help: Get started with Google Ads
6. LLM Search and Generative AI Optimization
As demonstrated in the project brief (the ChatGPT example surfacing an incorrect docs.ipinfusion.com URL), LLM-powered search tools are increasingly surfacing documentation links. Traditional SEO techniques help here too, but there are emerging LLM-specific strategies.
6.1 llms.txt Standard
The llms.txt file is a proposed open standard (proposed by Jeremy Howard of Answer.AI in September 2024) for helping LLMs discover and navigate website content efficiently. Placed at the root of a domain (e.g., https://documentation.ipinfusion.com/llms.txt), it provides a structured Markdown index of a site's key pages.
Format example:
# IP Infusion OcNOS Documentation
> Official technical documentation for OcNOS network operating system
> by IP Infusion. Covers Service Provider, Data Center, and carrier-grade
> networking configuration.
## Latest Releases
- [OcNOS SP 7.0 Home](https://documentation.ipinfusion.com/ocnos-sp-layer-2-7.0/Content/Home.htm): Service Provider edition, version 7.0
- [OcNOS DC 7.0 Home](https://documentation.ipinfusion.com/ocnos-dc-7.0/Content/Home.htm): Data Center edition, version 7.0
## Configuration Guides
- [Layer 3 Configuration](https://documentation.ipinfusion.com/...): Routing protocols, OSPF, BGP, MPLS
- [Layer 2 Configuration](https://documentation.ipinfusion.com/...): VLANs, spanning tree, switching
Current state of adoption: As of early 2026, major platforms including Anthropic, Cloudflare, and Stripe have implemented llms.txt. GPTBot has been observed fetching llms.txt files. However, no major AI platform has officially confirmed that llms.txt directly influences citation results. Adoption is low-risk and low-effort (a few hours of implementation), and the potential upside increases as LLM-based search grows.
Official
llms.txtproposal: llmstxt.org
Mintlify'sllms.txtimplementation guide: Simplifying docs for AI with /llms.txt
SE Ranking research on adoption: Does LLMs.txt impact AI visibility?
Recommendation: Implement llms.txt on documentation.ipinfusion.com pointing exclusively to current, preferred content. This is a lightweight investment with forward-looking upside.
6.2 Markdown Versions of Documentation Pages
The llms.txt standard also recommends making page content available in clean Markdown at the same URL with .md appended (e.g., documentation.ipinfusion.com/ocnos-sp-7.0/Content/Home.htm.md). This makes content easily ingestible by AI tools without HTML parsing overhead.
Whether this is feasible depends on MadCap Flare's output capabilities. Investigate whether Flare can export Markdown alongside HTML, or whether a post-processing step could convert HTML pages to Markdown.
6.3 Allow AI Crawlers in robots.txt
Ensure that the robots.txt on documentation.ipinfusion.com does not inadvertently block AI crawlers such as:
GPTBot(OpenAI/ChatGPT)ClaudeBot(Anthropic)PerplexityBotGoogle-Extended(Google's AI training crawler)
These crawlers respect robots.txt disallow rules. If you want LLMs to cite your current documentation, these bots must be able to reach it.
7. Typesense Search — Comparison and Fine-Tuning
The existing Typesense self-hosted search solution provides a curated, harmonized search experience across both servers. It can also serve as a data source for evaluating and fine-tuning remediation efforts.
Recommended uses:
- Benchmarking: Run identical queries through Typesense, Google Search, and LLM tools (ChatGPT, Perplexity, etc.) side-by-side. Document which URLs surface for each. This creates a measurable baseline and tracks improvement over time.
- Priority ranking in Typesense: Ensure that Typesense results are weighted to boost current-version content over legacy. Review Typesense ranking rules and apply version-based boosts.
- Coverage gap detection: If a query returns good results in Typesense but poor results in Google, it indicates a Google-specific indexing or authority issue—useful for diagnosing problem areas.
Typesense documentation: Typesense Docs
Typesense ranking and relevance tuning: Ranking and Relevance
8. EOL/EOS Content Policy
The project brief raises whether older content can be removed from docs.ipinfusion.com beyond certain End-of-Life / End-of-Support dates. This is a high-value, high-leverage action.
Recommendation: Establish a formal documentation retention policy:
- Define EOL/EOS dates for each OcNOS release version (this is an IP Infusion internal decision)
- For EOL content: Return HTTP 410 Gone from the Nginx server for all pages in that version's directory. This is the clearest signal to Google that the content is permanently gone.
- For EOS content that still has some value: Redirect to either the nearest equivalent current page or a "version archived" landing page that clearly states the content is no longer maintained and links to the current release.
- Do not simply delete files: Deleting files without configuring Nginx to return 404/410 will result in generic server errors rather than clean deindexing signals.
9. Figuring Out the Source: Google Programmable Search Engine
The project brief specifically asks whether the legacy Google Programmable Search (PSE) could be the source of indexed URLs.
Investigation steps:
- Log in to the Google PSE control panel at programmablesearchengine.google.com
- Check whether any existing PSEs include
docs.ipinfusion.comas an indexed site - If found: Remove
docs.ipinfusion.comfrom the PSE's site list and verify the PSE is pointing only todocumentation.ipinfusion.com
Note that PSE primarily controls results within the custom search widget, not organic Google Search. However, Googlebot may have discovered and indexed linked pages through PSE-connected sitemaps or through crawling patterns associated with PSE configuration. Removing the legacy domain from any PSE is still a recommended hygiene step.
10. Implementation Roadmap
Phase 1 — Immediate (Week 1–2)
| Task | Owner | Notes |
|---|---|---|
| Verify/set up GSC properties for both domains | IP Infusion | docs.ipinfusion.com and documentation.ipinfusion.com |
Run site:docs.ipinfusion.com audit; export URL inventory | Mighty Digital | Use GSC + Screaming Frog |
| Submit GSC Removal requests for top 20–50 highest-visibility legacy URLs | Mighty Digital | 180-day suppression while permanent work proceeds |
| Investigate Google PSE for legacy domain references | IP Infusion / Mighty Digital | Remove if found |
| Manual LLM search audit (ChatGPT, Perplexity, Google AI) | IP Infusion Tech Pubs | Catalog incorrect URL citations |
Phase 2 — Short-Term (Weeks 3–6)
| Task | Owner | Notes |
|---|---|---|
Implement Nginx noindex X-Robots-Tag headers on Legacy server for deprecated version directories | Mighty Digital | Highest-impact permanent deindexing action |
| Implement Nginx 410 Gone for confirmed EOL versions | Mighty Digital | Requires EOL list from IP Infusion |
Add/update robots.txt on docs.ipinfusion.com to block crawling of deprecated paths | Mighty Digital | |
| Build redirect map file: legacy URL → current URL | IP Infusion Tech Pubs + Mighty Digital | Tech Pubs identifies correct destinations; MDG implements |
Submit updated XML sitemap for documentation.ipinfusion.com to GSC | IP Infusion / Flare admin | Ensure only current content included |
Phase 3 — Medium-Term (Weeks 7–12)
| Task | Owner | Notes |
|---|---|---|
| Implement Nginx 301 redirect rules on Legacy server | Mighty Digital | Using map file from Phase 2 |
Implement llms.txt on documentation.ipinfusion.com | Mighty Digital | Current versions and key topics only |
| On-page SEO improvements for CURRENT server (titles, meta descriptions) | IP Infusion Tech Pubs / Mighty Digital | May require Flare template changes |
| Typesense side-by-side benchmark vs Google vs LLM | Mighty Digital | Create regular reporting cadence |
| Evaluate Google Ads for priority branded queries | IP Infusion (decision) / Mighty Digital (execution) | If organic signals still lagging |
Phase 4 — Ongoing
| Task | Owner | Frequency |
|---|---|---|
| Monitor GSC coverage and performance reports | Mighty Digital | Monthly |
Re-run site:docs.ipinfusion.com to track index reduction | Mighty Digital | Monthly |
| Manual LLM audit of priority queries | IP Infusion Tech Pubs | Quarterly |
| Review and update redirect map for newly deprecated content | Both | Per release cycle |
Update llms.txt with new releases | Mighty Digital | Per release cycle |
11. Measurement Criteria
Define success clearly before beginning so progress can be objectively tracked.
| Metric | Baseline (Measure Now) | Target |
|---|---|---|
# of docs.ipinfusion.com URLs in Google index | TBD (run site: audit) | Reduce by 90%+ for EOL versions; 100% for confirmed deprecated content |
| # of incorrect legacy URLs surfaced by LLM queries (test set) | TBD (manual audit) | Reduce to 0 for test query set |
Google Search impressions for docs.ipinfusion.com | TBD (GSC Performance) | Reduce to near 0 |
Google Search impressions for documentation.ipinfusion.com | TBD (GSC Performance) | Stable or increasing |
| Click-through rate on current docs | TBD (GSC) | Improving trend |
| Redirect success rate on Legacy server | N/A | 95%+ of legacy URLs redirect to valid current destination |
12. Division of Labor Summary
| Responsibility | IP Infusion Tech Pubs | Mighty Digital |
|---|---|---|
| EOL/EOS version list and dates | ✅ Primary | Advise |
| URL mapping (legacy → current) | ✅ Primary | Assist with pattern matching |
| GSC property verification and access | ✅ Primary | Assist |
| GSC Removal requests | Approve | ✅ Execute |
| Nginx configuration (robots.txt, noindex headers, 410, 301 redirects) | — | ✅ Execute |
llms.txt creation and maintenance | Content authority | ✅ Implement |
| Typesense relevance tuning | Review | ✅ Execute |
| On-page SEO in MadCap Flare templates | ✅ Primary (Flare access) | Advise |
| Google Ads campaign (if approved) | Approve budget | ✅ Manage |
| Reporting and measurement | Review | ✅ Produce |
13. Key Resources and References
Google Official Documentation
- Remove Your Site Info from Google (Search Central)
- Google Removals Tool (Search Console)
- Robots Meta Tag and X-Robots-Tag
- Robots.txt Introduction
- Build and Submit a Sitemap
- SEO Starter Guide
- Note on Unsupported robots.txt Rules (noindex)
Nginx Configuration
- Creating NGINX Rewrite Rules (Official NGINX Blog)
- DigitalOcean: How To Create Redirects with Nginx
- DigitalOcean: Nginx Rewrite URL Rules
SEO Deindexing Guides
- How to Use GSC's Temporary Removal Tool (gtechme.com)
- How to Bulk Remove URLs from Google's Index (Thrive Agency)
- Conductor: How to Remove URLs from Google Search
LLM Search and llms.txt
- llmstxt.org — Official Proposal
- Mintlify: Simplifying Docs for AI with /llms.txt
- SE Ranking Research: Does llms.txt Impact AI Visibility?
- Search Engine Land: llms.txt Proposed Standard
- Complete Guide to llms.txt (getpublii.com)
Tooling
- Google Search Console
- Google Programmable Search Engine Control Panel
- Screaming Frog SEO Spider
- Sitebulb Site Crawler
- Ahrefs Site Audit
- Semrush Site Audit
- Typesense Documentation
This document is a living plan and should be updated as the inventory audit is completed, EOL decisions are finalized, and implementation phases proceed.