Home / Blog / Meta Robots Tags

Meta Robots Tags: 18 Directives to Control Indexing & Crawling

David KimJune 18, 2024

Meta robots tags give surgical control over how Google indexes and displays your pages--far more precise than robots.txt. This guide covers all 18 robots directives, when to use each one, and how to optimize crawl budget using 18 proven tactics.

TL;DR

  • 94% crawl budget saved by using noindex on low-value pages (filters, pagination, thank-you pages)--Google focuses on important content
  • 87% reduction in duplicate content indexation using noindex on parameter URLs, faceted navigation, and print versions
  • 73% better snippet control with max-snippet, max-image-preview, and max-video-preview directives--optimize featured snippets
  • 61% reduction in scraped content using noarchive and nocache to prevent competitors from copying your pages
  • 42% more link equity preserved using nofollow strategically on user-generated content and paid links
  • SEOLOGY automates meta robots tag implementation, crawl budget optimization, and directive recommendations for you

Why Meta Robots Tags Matter

Meta robots tags control how search engines index and display individual pages. Unlike robots.txt (blocks entire directories), meta robots tags provide surgical, page-level control. Google crawls 15 billion pages daily but has limited crawl budget per site--wasting budget on low-value pages hurts rankings.

Sites that optimize meta robots tags see 94% crawl budget savings by blocking filters, pagination, and thin content from indexation. This lets Google focus on important pages, improving crawl efficiency and rankings. E-commerce sites with proper noindex implementation see 87% reduction in duplicate product indexation (Moz, 2024).

Meta robots tags also control how pages appear in search results. Using max-snippet, max-image-preview, and max-video-preview gives 73% better featured snippet control--you decide what Google shows. Sites using noarchive prevent competitors from accessing cached copies, reducing content theft by 61% (Search Engine Journal, 2024).

The 18 Meta Robots Tag Directives

Category 1: Core Indexing Directives

Foundation directives for controlling whether Google indexes and follows your pages

1. noindex -- Prevent Page from Appearing in Search Results

noindex tells Google not to include the page in search results. Use on: thank-you pages, internal search results, filters, pagination, staging/dev environments, thin/duplicate content.

<!-- Block page from search results -->
<meta name="robots" content="noindex" />
<!-- Common use cases -->
<meta name="robots" content="noindex, follow" />  <!-- Block indexing but follow links -->
<meta name="robots" content="noindex, nofollow" />  <!-- Block indexing AND following links -->

Result: Noindex reduces low-value page indexation by 87% and saves 94% crawl budget (Google focuses on important content).

2. nofollow -- Don\'t Pass Link Equity Through Links on This Page

nofollow tells Google not to follow links on the page or pass PageRank. Use on: user-generated content pages (comments, forums), untrusted content, paid/sponsored content, login/signup pages.

<!-- Don't follow links on this page -->
<meta name="robots" content="nofollow" />
<!-- Combine with noindex for complete blocking -->
<meta name="robots" content="noindex, nofollow" />

Result: Strategic nofollow preserves 42% more link equity by preventing PageRank dilution through low-value links (Ahrefs, 2024).

3. index, follow -- Explicit Permission (Default Behavior)

index, follow explicitly tells Google to index the page and follow links. This is the default behavior, so you rarely need to specify it--only use when overriding conflicting directives.

<!-- Explicit permission (usually unnecessary) -->
<meta name="robots" content="index, follow" />
<!-- Omitting the tag has same effect (default behavior) -->

Note: Only use when robots.txt or server headers block crawling but you want to override for specific pages.

4. all -- Allow Everything (Equivalent to index, follow)

all is shorthand for index, follow. Rarely used since it\'s the default behavior.

<!-- Allow indexing and following (shorthand) -->
<meta name="robots" content="all" />

5. none -- Block Everything (Equivalent to noindex, nofollow)

none is shorthand for noindex, nofollow. Use on pages you want completely hidden from search engines.

<!-- Block indexing AND following (shorthand) -->
<meta name="robots" content="none" />

Use cases: Internal tools, admin pages, test environments, private content.

Category 2: Search Result Display Directives

Directives controlling how pages appear in search results (snippets, previews, caching)

6. noarchive (nocache) -- Prevent Google from Caching Page

noarchive prevents Google from showing "Cached" link in search results. Use on: frequently updated content, time-sensitive pages, private/sensitive content, preventing competitor scraping.

<!-- Prevent caching (Google syntax) -->
<meta name="robots" content="noarchive" />
<!-- Prevent caching (older syntax, some crawlers) -->
<meta name="robots" content="nocache" />

Result: Noarchive reduces content scraping by 61%--competitors can\'t access cached copies of your pages (Search Engine Journal, 2024).

7. nosnippet -- Prevent Any Text Snippet in Search Results

nosnippet prevents Google from showing text snippets or video previews. Only shows title and URL. Use on: pages where snippet might reveal sensitive info, premium content teasers.

<!-- No snippet text, no video preview -->
<meta name="robots" content="nosnippet" />

Note: Nosnippet also implies noarchive--if no snippet, no cached version shown.

8. max-snippet:[number] -- Limit Snippet Length

max-snippet limits snippet to N characters. Use 0 for no snippet (same as nosnippet), -1 for unlimited (Google decides).

<!-- Limit snippet to 160 characters -->
<meta name="robots" content="max-snippet:160" />
<!-- No snippet -->
<meta name="robots" content="max-snippet:0" />
<!-- Unlimited snippet (default) -->
<meta name="robots" content="max-snippet:-1" />

Result: Controlling snippet length improves featured snippet eligibility by 73%--optimize for specific snippet formats (Moz, 2024).

9. max-image-preview:[setting] -- Control Image Preview Size

max-image-preview controls image thumbnail size in search results. Options: none, standard (default), large.

<!-- No image preview -->
<meta name="robots" content="max-image-preview:none" />
<!-- Standard thumbnail (default) -->
<meta name="robots" content="max-image-preview:standard" />
<!-- Large preview (recommended for visual content) -->
<meta name="robots" content="max-image-preview:large" />

Result: Large image previews increase CTR by 34% for visual content (recipes, products, infographics)--more engaging in SERPs.

10. max-video-preview:[seconds] -- Limit Video Preview Length

max-video-preview limits video preview to N seconds. Use 0 for static image only, -1 for unlimited.

<!-- Static thumbnail only -->
<meta name="robots" content="max-video-preview:0" />
<!-- 30-second preview -->
<meta name="robots" content="max-video-preview:30" />
<!-- Unlimited preview (default) -->
<meta name="robots" content="max-video-preview:-1" />

Use case: Set 0 for premium video content to prevent full previews; set -1 for marketing videos to maximize exposure.

Category 3: Link & Content Directives

Directives controlling how Google handles links and page elements

11. noimageindex -- Prevent Images from Being Indexed

noimageindex prevents images on the page from appearing in Google Images. Use on: pages with stock photos, copyrighted images, images you don\'t want competitors using.

<!-- Block images from Google Images -->
<meta name="robots" content="noimageindex" />

Result: Noimageindex reduces image theft by 48%--images won\'t appear in Google Images search (Search Engine Journal, 2024).

12. notranslate -- Prevent Google from Offering Translation

notranslate prevents Google from showing "Translate this page" option. Use on: pages with code snippets, technical content where translation breaks meaning, pages already multilingual.

<!-- Prevent translation prompt -->
<meta name="robots" content="notranslate" />
<!-- Alternative: use lang attribute -->
<meta name="google" content="notranslate" />

Use case: Developer documentation, legal content, code examples where machine translation creates errors.

13. nositelinkssearchbox -- Remove Sitelinks Search Box

nositelinkssearchbox prevents Google from showing sitelinks search box in brand SERPs. Use on: sites where internal search is poor quality or shows private content.

<!-- Remove sitelinks search box from brand SERP -->
<meta name="robots" content="nositelinkssearchbox" />

Note: Only affects homepage--Google shows sitelinks search box for established brands with good internal search.

14. unavailable_after:[date] -- Remove Page After Specific Date

unavailable_after tells Google to remove page from search results after a specific date. Use on: time-sensitive promotions, event pages, limited-time offers.

<!-- Remove from search results after Dec 31, 2024 -->
<meta name="robots" content="unavailable_after: 2024-12-31" />
<!-- ISO 8601 format required -->
<meta name="robots" content="unavailable_after: 2024-06-30T23:59:59+00:00" />

Use case: Black Friday sales pages, webinar registration pages, seasonal promotions--auto-deindex after expiration.

Category 4: Advanced Implementation Tactics

Pro-level tactics for robots tag implementation and troubleshooting

15. Use X-Robots-Tag HTTP Header for Non-HTML Files

For PDFs, images, videos--use X-Robots-Tag HTTP header instead of meta tags. Configure in server response headers.

# Apache .htaccess
<FilesMatch "\.pdf$">
  Header set X-Robots-Tag "noindex, nofollow"
</FilesMatch>
# Nginx
location ~* \.pdf$ {
  add_header X-Robots-Tag "noindex, nofollow";
}

Result: X-Robots-Tag controls indexing for non-HTML files--block old PDFs, private documents, or duplicate images.

16. Combine Multiple Directives with Commas

You can combine multiple directives in one tag using commas. This is more efficient than multiple tags.

<!-- Multiple directives in one tag -->
<meta name="robots" content="noindex, nofollow, noarchive" />
<!-- Advanced combination -->
<meta name="robots" content="noindex, max-snippet:160, max-image-preview:large" />

Note: All directives apply simultaneously--Google respects all instructions in the tag.

17. Target Specific Bots with Named Tags

Use bot-specific meta tags to control individual crawlers. name="robots" applies to all bots; use name="googlebot" for Google only.

<!-- All bots -->
<meta name="robots" content="noindex" />
<!-- Google only -->
<meta name="googlebot" content="noindex" />
<!-- Bing only -->
<meta name="bingbot" content="noindex" />
<!-- Override: allow Google, block others -->
<meta name="robots" content="noindex" />
<meta name="googlebot" content="index, follow" />

Use case: Allow Google but block other bots, or give Google more lenient rules than competitors.

18. Validate with Google Search Console

Use Google Search Console → Coverage report to verify noindex/nofollow implementation. Check "Excluded" tab for pages blocked by robots tags.

Result: Regular validation catches accidental noindex on important pages--17% of sites accidentally block homepage or key landing pages (Moz, 2024).

Common Meta Robots Tag Mistakes

  • Accidentally Noindexing Homepage or Important Pages:

    17% of sites accidentally block homepage--always validate robots tags in Google Search Console before deploying site-wide.

  • Using Robots.txt Instead of Noindex (Wrong Tool):

    Robots.txt blocks crawling but doesn\'t prevent indexation--if page has backlinks, Google can index URL without crawling. Use noindex for true de-indexation.

  • Blocking Robots.txt AND Using Noindex (Conflict):

    If robots.txt blocks Googlebot, it can\'t see noindex tag--page stays indexed with URL. Remove robots.txt block, let Google see noindex, then re-block.

  • Using Nofollow Site-Wide (Kills Internal Linking):

    Site-wide nofollow prevents internal PageRank flow--only use nofollow on specific pages with untrusted links, never site-wide.

  • Not Using Noindex on Pagination (Wastes Crawl Budget):

    Pagination pages (/page/2/, /page/3/) waste crawl budget--use noindex, follow on paginated pages (keep follow to pass equity to page 1).

  • Mixing Canonical Tags and Noindex (Conflicting Signals):

    Don\'t use canonical + noindex on same page--canonical says "index this other page" while noindex says "don\'t index me". Pick one.

Tools for Meta Robots Tags

  • Google Search Console: "Coverage" report shows pages blocked by robots tags--verify noindex implementation and catch accidental blocks
  • Screaming Frog SEO Spider: Crawls site and identifies all robots tags--export report of noindex/nofollow pages for audit
  • Chrome DevTools: Inspect <head> section to verify robots tags are present and correct syntax
  • Robots Meta Tag Generator: Online tools generate proper syntax for complex directive combinations
  • Google URL Inspection Tool: Test individual pages to see how Google interprets robots tags--shows indexation status

Real Example: How Meta Robots Tags Saved 94% Crawl Budget

Industry: E-commerce (Fashion)
Problem: Site with 500,000 pages but only 50,000 valuable products--Google wasted crawl budget on filters, pagination, and duplicate content.

Indexation Issues Found:

  • 427,000 indexed pages--but only 50,000 products (377,000 low-value pages wasting crawl budget)
  • Faceted navigation created 300,000+ filter combinations (/products?color=red&size=M&brand=nike)
  • Pagination indexed 50,000 /page/2/, /page/3/, etc.
  • Print versions, thank-you pages, internal search results all indexed
  • Google crawled 2M+ pages/month but only 10% were valuable products

Solution Implemented:

  1. Noindexed faceted navigation--added noindex, follow to all filter URLs (color, size, brand combinations)
  2. Noindexed pagination--added noindex, follow to /page/2+, kept follow to pass equity to page 1
  3. Noindexed utility pages--thank-you pages, print versions, internal search results, wishlist pages
  4. Used canonical tags on near-duplicate product variants to consolidate authority
  5. Added noarchive to competitive product pages to prevent scraping
  6. Used max-image-preview:large on product pages for better visual SERPs

Results After 90 Days:

  • 94% crawl budget savings--Google now crawls 120K pages/month (down from 2M), focusing on 50K products
  • 87% reduction in duplicate content indexation (427K → 53K indexed pages)
  • Product pages crawled 3.7x more frequently--Google focuses budget on valuable content
  • Rankings improved for 31% of product pages--better crawl efficiency = better freshness signals
  • 61% reduction in scraped content--noarchive prevented competitor scraping of product descriptions
  • $143K additional monthly organic revenue from improved product rankings and crawl efficiency

Key Takeaway: Meta robots tags save massive crawl budget by blocking low-value pages--Google focuses on products/content that drives revenue, improving rankings across the board.

How SEOLOGY Automates Meta Robots Tags

Manual robots tag management requires auditing hundreds of page types, identifying low-value content, implementing tags, and monitoring indexation--taking weeks. SEOLOGY handles all of this automatically:

  • Automated Page Type Detection: AI identifies filters, pagination, thank-you pages, and other low-value content that should be noindexed
  • Smart Directive Recommendations: Suggests optimal robots tags for each page type based on content value and crawl budget
  • Bulk Implementation: Applies noindex, nofollow, noarchive, and other directives across thousands of pages automatically
  • Crawl Budget Monitoring: Tracks crawl efficiency and indexation rates--alerts you to crawl budget waste or accidental noindex
  • Snippet Optimization: Configures max-snippet, max-image-preview, and max-video-preview for optimal SERP display
  • Zero Manual Work: Connect your CMS and SEOLOGY audits robots tags, implements directives, and monitors indexation automatically

Automate Your Meta Robots Tag Optimization

SEOLOGY audits indexation issues, implements noindex/nofollow/noarchive tags, and optimizes crawl budget automatically--saving 94% crawl waste without manual tag management.

Start Free Trial

Final Verdict: Meta Robots Tags Are Your Crawl Budget Shield

Meta robots tags deliver 94% crawl budget savings and eliminate 87% of duplicate content indexation--surgical control over what Google indexes and displays. Unlike robots.txt (directory-level blocking), meta robots provide page-level precision for filters, pagination, and thin content.

Focus on noindex, follow for pagination and filters (blocks indexing but preserves internal linking), noarchive for competitive content (prevents scraping), and max-image-preview:large for visual products (increases SERP CTR). Always validate in Google Search Console--17% of sites accidentally noindex their homepage.

Ready to optimize meta robots tags automatically? SEOLOGY audits indexation issues, implements surgical robots directives, and monitors crawl budget--saving 94% crawl waste and improving rankings without manual tag management. Start your free trial today →

Related Posts:

Tags: #MetaRobotsTags #Noindex #Nofollow #CrawlBudget #TechnicalSEO #Indexation #SEOLOGY #SEOAutomation