Cloudflare Allow Verified Bots

Published: January 8, 202518 min readCloudflare

Cloudflare Allow Verified Bots: Complete Configuration Guide

Cloudflare's Web Application Firewall provides powerful protection against malicious traffic, but its aggressive security measures can inadvertently block legitimate search engine crawlers and other beneficial bots. Learning how to configure Cloudflare to allow verified bots ensures that your site maintains robust security while remaining accessible to Googlebot, Bingbot, and other essential crawlers that drive organic traffic. This comprehensive guide explains Cloudflare's verified bot system, walks through custom rule configuration, and provides advanced strategies for granular bot management that balances security with search engine visibility.

Understanding Cloudflare's Verified Bot System

Cloudflare maintains an extensive directory of verified bots—legitimate crawlers and automated services that have been validated and approved for inclusion in Cloudflare's trusted bot list. This verification system helps site owners distinguish between beneficial bots that should access their content and malicious bots that should be blocked. The verified bot list includes major search engine crawlers like Googlebot and Bingbot, SEO tools like Google Lighthouse and GTmetrix, monitoring services, social media preview generators, and numerous other legitimate automated services.

The verification process ensures that bots meet Cloudflare's strict requirements before being added to the trusted list. Bots must follow Cloudflare's verified bot policy and use one of two validation methods: Web Bot Auth or IP validation. Web Bot Auth involves cryptographic signatures that prove a bot's identity, while IP validation confirms that requests originate from IP addresses officially associated with the bot's operator. This dual-verification approach prevents malicious actors from impersonating legitimate crawlers.

Once Cloudflare verifies a bot, it appears in Cloudflare Radar's public bots and agents directory. Site owners can then reference these verified bots in their WAF custom rules using the cf.client.bot field, which returns true when a request comes from a verified bot. This field provides a simple, reliable way to identify legitimate crawlers without manually maintaining lists of IP addresses or user agents.

Cloudflare categorizes verified bots into sixteen distinct categories including Search Engine Crawler, AI Crawler, SEO, Monitoring & Analytics, Page Preview, Feed Fetcher, Aggregator, Archiver, Security, Accessibility, Academic Research, Advertising & Marketing, Social Media Marketing, Webhooks, AI Assistant, and AI Search. These categories enable granular control over which types of bots can access your site, allowing you to permit search engine crawlers while blocking AI training crawlers, for example.

Why Cloudflare Users Need to Whitelist Bots

Cloudflare's security features operate on a principle of aggressive protection—when in doubt, block the traffic. This approach effectively protects sites from DDoS attacks, scraping attempts, and malicious bot traffic, but it can also block beneficial bots that your business depends on. The consequences of blocking legitimate bots extend far beyond just missing search engine rankings.

When Cloudflare blocks Googlebot, your pages cannot be crawled or indexed, causing your site to gradually disappear from Google search results. New content never gets discovered, existing pages fall out of the index, and your organic search traffic evaporates. Similarly, blocking Bingbot eliminates your visibility in Bing search results, cutting off a significant source of potential traffic.

Beyond search engines, blocking other verified bots creates operational problems. Social media platforms like Facebook, Twitter, and LinkedIn use bots to generate link previews when users share your content. If Cloudflare blocks these preview bots, your shared links appear without images or descriptions, dramatically reducing click-through rates. Monitoring services cannot check your site's uptime, SEO tools cannot audit your pages, and feed readers cannot retrieve your RSS updates.

The challenge lies in distinguishing between beneficial bots and malicious ones. Malicious bots often impersonate legitimate crawlers by using fake user agents or spoofed identities. Simply allowing all traffic that claims to be Googlebot would expose your site to these impersonation attacks. Cloudflare's verified bot system solves this problem by cryptographically validating bot identities, ensuring that only genuine Googlebot traffic receives trusted status.

The cf.client.bot Field Explained

The cf.client.bot field represents the cornerstone of Cloudflare's bot management system. This boolean field evaluates to true when Cloudflare's verification systems confirm that an incoming request originates from a verified bot. Understanding how this field works and how to use it effectively enables sophisticated bot management strategies.

When a request arrives at your Cloudflare-protected site, Cloudflare's bot detection systems analyze multiple signals including IP address, user agent string, request patterns, and cryptographic signatures. If the request matches a verified bot's known characteristics and passes Cloudflare's validation checks, the cf.client.bot field is set to true. This happens automatically without requiring any configuration from site owners.

The power of cf.client.bot lies in its simplicity and reliability. Rather than maintaining complex lists of IP addresses that change frequently, or parsing user agent strings that can be easily spoofed, you simply check whether Cloudflare has verified the bot. This approach remains effective even as bots update their infrastructure, change IP addresses, or modify their user agents.

You can use cf.client.bot in custom rule expressions to create sophisticated bot management logic. The most common pattern involves using the NOT operator to identify requests that are not from verified bots. For example, the expression not cf.client.bot matches all traffic except verified bots, allowing you to apply security measures to human visitors and unverified bots while exempting legitimate crawlers.

Combining cf.client.bot with other fields creates powerful conditional logic. You might challenge requests from specific countries while allowing verified bots: (ip.src.country in {"CN" "RU"} and not cf.client.bot). This expression challenges traffic from China and Russia unless it comes from a verified bot, maintaining security while ensuring search engines can still crawl your site.

Creating Custom Rules to Allow Verified Bots

Cloudflare's WAF custom rules provide the mechanism for implementing bot management policies. These rules evaluate incoming requests against expressions you define and take actions based on whether requests match those expressions. Creating effective custom rules requires understanding rule structure, expression syntax, and action types.

Accessing WAF Custom Rules

To create custom rules for bot management, log into your Cloudflare dashboard and select the domain you want to configure. Navigate to "Security" in the left sidebar, then click "WAF" to access the Web Application Firewall settings. Within the WAF section, click on "Custom rules" to view your existing rules and create new ones.

The custom rules interface displays all rules currently active for your domain, showing their order, expressions, and actions. Rule order matters—Cloudflare evaluates rules sequentially and stops processing when a rule matches and takes a terminal action. Understanding rule order helps you structure your bot management logic effectively.

Basic Rule: Allow All Verified Bots

The simplest bot management rule allows all verified bots while applying security measures to other traffic. This approach works well for most sites that want to ensure search engine crawlers can always access their content.

Click "Create custom rule" to begin building your rule. Give it a descriptive name like "Allow Verified Bots" so you can easily identify its purpose. In the expression builder, you'll create a rule that challenges non-bot traffic while allowing verified bots to pass through.

Set up your rule expression using the following logic:

Field: cf.client.bot
Operator: equals
Value: false

This expression matches requests that are not from verified bots. Alternatively, you can use the expression editor to write: not cf.client.bot

For the action, select "Managed Challenge." This action presents an interactive challenge to human visitors that bots cannot easily solve, while allowing verified bots to bypass the challenge entirely. Managed Challenge provides better user experience than traditional CAPTCHAs while maintaining strong security.

Save your rule and it takes effect immediately. Cloudflare will now challenge all non-bot traffic while allowing verified bots free access to your site. Monitor your security events in the Cloudflare dashboard to see the rule in action and verify that legitimate traffic isn't being incorrectly blocked.

Advanced Rule: Country-Based with Bot Exceptions

Many sites want to apply geo-blocking or country-specific security measures while ensuring search engines can still crawl their content. This requires combining geographical filters with bot verification logic.

Create a new custom rule named "Challenge High-Risk Countries Except Bots." In the expression builder, you'll combine multiple conditions using logical operators to create sophisticated filtering logic.

Use the following expression structure:

(ip.src.country in {"CN" "RU" "KP"} and not cf.client.bot)

This expression matches requests from China, Russia, and North Korea that are not from verified bots. Breaking down the logic:

  • ip.src.country in {"CN" "RU" "KP"} matches requests from the specified countries
  • and requires both conditions to be true
  • not cf.client.bot matches requests that are not from verified bots

The combined expression only matches non-bot traffic from those countries. Verified bots from any country will not match this rule and will be allowed through.

Set the action to "Managed Challenge" to present an interactive challenge to matching requests. This approach blocks most malicious traffic from high-risk countries while ensuring Googlebot, Bingbot, and other legitimate crawlers can still access your site regardless of their geographic origin.

Advanced Rule: Rate Limiting with Bot Exceptions

Rate limiting prevents abuse by restricting how many requests a single IP address can make within a time window. However, search engine crawlers legitimately make many requests in short periods as they crawl your site. Excluding verified bots from rate limits ensures they can crawl efficiently.

Navigate to "Security" > "WAF" > "Rate limiting rules" in your Cloudflare dashboard. Click "Create rate limiting rule" to begin configuring your rate limit with bot exceptions.

Name your rule "Rate Limit Non-Bots" and configure the following settings:

Expression: not cf.client.bot

This ensures the rate limit only applies to non-bot traffic. Verified bots will not be counted against the rate limit and cannot trigger it.

Requests: Set your threshold, for example "100 requests"
Period: Set your time window, for example "10 seconds"

Action: "Managed Challenge"

This configuration challenges any non-bot IP address that makes more than 100 requests in 10 seconds, while allowing verified bots to crawl as quickly as they need without restriction. Adjust the threshold and period based on your site's normal traffic patterns and tolerance for aggressive crawling.

Using Bot Categories for Granular Control

Cloudflare's verified bot categories enable sophisticated bot management strategies that go beyond simple allow/block decisions. The cf.verified_bot_category field lets you create rules that treat different types of bots differently based on their purpose and function.

Allowing Search Engines While Blocking AI Crawlers

The rise of AI training crawlers has created new concerns for content creators who want their content indexed by search engines but don't want it used to train large language models. Cloudflare's bot categories make this distinction possible.

Create a custom rule that allows search engine crawlers while blocking AI crawlers:

Expression:

(cf.verified_bot_category eq "AI Crawler" and cf.client.bot)

Action: Block

This rule specifically blocks verified bots categorized as AI Crawlers, which includes bots like ChatGPT-Bot and Google-Extended that collect data for AI model training. The expression verifies that the request is both from a verified bot and specifically categorized as an AI Crawler before blocking it.

To ensure search engine crawlers remain allowed, create a second rule with higher priority:

Expression:

(cf.verified_bot_category eq "Search Engine Crawler" and cf.client.bot)

Action: Skip (skip all remaining custom rules)

This rule explicitly allows search engine crawlers by skipping all subsequent custom rules that might otherwise block them. Place this rule before your AI Crawler blocking rule in the rule order to ensure search engines always have access.

Allowing SEO Tools and Monitoring Services

SEO tools and monitoring services provide valuable insights about your site's performance, but some site owners want to restrict which tools can access their sites. Bot categories enable selective access.

Create a custom rule that allows specific bot categories:

Expression:

(cf.verified_bot_category in {"Search Engine Crawler" "Search Engine Optimization" "Monitoring & Analytics"} and cf.client.bot)

Action: Skip (skip all remaining custom rules)

This rule allows three categories of bots: search engine crawlers, SEO tools, and monitoring services. All other bot categories will be subject to your normal security rules. This approach works well for sites that want to limit bot access while ensuring essential services remain functional.

Blocking All Bots Except Search Engines

Some sites want to block all automated traffic except search engine crawlers. This aggressive approach maximizes security while maintaining search engine visibility.

Create two rules to implement this policy. First, allow search engine crawlers:

Rule 1 - Allow Search Engines:

Expression:

(cf.verified_bot_category eq "Search Engine Crawler" and cf.client.bot)

Action: Skip (skip all remaining custom rules)

Then block all other verified bots:

Rule 2 - Block Other Bots:

Expression:

cf.client.bot

Action: Block

This two-rule combination allows only search engine crawlers while blocking all other verified bots including AI crawlers, monitoring services, feed fetchers, and social media preview bots. Consider the implications carefully—blocking social media preview bots will prevent link previews when users share your content, potentially reducing social media engagement.

Troubleshooting Cloudflare Bot Management

Even with proper configuration, bot management sometimes produces unexpected results. Understanding common issues and their solutions helps you quickly resolve problems and maintain both security and accessibility.

Issue: Legitimate Bots Still Being Blocked

If verified bots continue being blocked after you've created allow rules, check your rule order. Cloudflare evaluates custom rules sequentially, and a blocking rule earlier in the sequence can prevent later allow rules from being reached.

Navigate to your custom rules list and examine the order. Rules that allow or skip for verified bots should appear before rules that block or challenge traffic. Use the drag handles to reorder rules, placing bot allow rules at the top of your rule list.

Also verify that you're using the correct expression syntax. The expression cf.client.bot evaluates to true for verified bots, so if you want to allow verified bots, you should use a Skip action with the expression cf.client.bot, or challenge non-bots with the expression not cf.client.bot.

Issue: Yandex Bot Intermittently Blocked

Yandex, Russia's primary search engine, occasionally experiences blocking issues when it updates its IP address ranges. Cloudflare's verified bot list may not immediately reflect these changes, causing temporary blocks until the new IPs propagate through Cloudflare's network.

The issue typically resolves automatically within 48 hours as Cloudflare updates its verified bot IP lists. If you need immediate resolution, create a temporary WAF exception that specifically allows Yandex traffic:

Expression:

(http.user_agent contains "Yandex")

Action: Skip (skip all remaining custom rules)

This rule allows any request with "Yandex" in the user agent to bypass your security rules. While this approach is less secure than relying on verified bot status, it provides immediate relief during IP transition periods. Remove this exception once Cloudflare's verified bot list updates.

Issue: Bot Management Rules Not Taking Effect

If your bot management rules don't seem to be working, verify that you're testing from actual bot IP addresses. Testing from your own computer or VPN will not trigger bot verification, even if you modify your user agent string to impersonate a bot.

Use Google Search Console's URL Inspection tool to test whether Googlebot can access your site. This tool sends real Googlebot requests from Google's infrastructure, providing accurate testing of your bot management rules. Similarly, Bing Webmaster Tools offers URL inspection features for testing Bingbot access.

Check your Cloudflare security events log to see how requests are being processed. Navigate to "Security" > "Events" to view recent security events, including which rules matched and what actions were taken. Filter events by bot score or user agent to focus on bot traffic and verify that your rules are matching as expected.

Issue: Legitimate Traffic Incorrectly Identified as Bots

Cloudflare's bot detection occasionally produces false positives, identifying human visitors as bots. This typically occurs with users behind corporate proxies, VPNs, or shared IP addresses that exhibit bot-like behavior patterns.

If users report being incorrectly challenged or blocked, review your custom rule expressions to ensure you're not inadvertently blocking human traffic. Rules that use cf.client.bot should only affect verified bots, not human visitors. However, rules that use bot scores (cf.bot_management.score) can affect humans if the score threshold is set too conservatively.

Consider using Managed Challenge instead of Block for bot-related rules. Managed Challenge allows legitimate users to prove they're human through an interactive challenge, while still blocking automated bot traffic. This approach reduces false positive impact while maintaining security.

Best Practices for Cloudflare Bot Management

Effective bot management requires ongoing attention and strategic planning. Following these best practices ensures your Cloudflare configuration maintains the right balance between security and accessibility.

Start with permissive rules and tighten gradually. Begin by allowing all verified bots and monitoring your security events to understand your bot traffic patterns. After establishing a baseline, create more restrictive rules targeting specific bot categories or behaviors. This incremental approach prevents accidentally blocking beneficial bots while you learn your site's traffic patterns.

Monitor security events regularly. Review your Cloudflare security events log at least weekly to identify trends, catch false positives, and verify that your rules are working as intended. Look for patterns in blocked traffic—if you see legitimate bots being blocked, adjust your rules accordingly.

Document your bot management strategy. Maintain clear documentation explaining which bots you allow, which you block, and why. Include the business rationale for each decision. This documentation helps team members understand your configuration and provides context when troubleshooting issues or making changes.

Test after making changes. Whenever you modify bot management rules, test crawler access using Google Search Console, Bing Webmaster Tools, and other verification tools. Immediate testing catches configuration errors before they impact your search rankings or site functionality.

Stay informed about bot ecosystem changes. New bots emerge regularly, existing bots update their infrastructure, and bot behaviors evolve over time. Subscribe to Cloudflare's blog and security announcements to stay informed about changes to the verified bot system. Review your bot management rules quarterly to ensure they remain aligned with current best practices.

Balance security and functionality. While blocking all bots except search engines maximizes security, it may break legitimate functionality like social media previews, RSS feeds, or monitoring services. Consider the business impact of blocking each bot category before implementing restrictive rules. The goal is appropriate security, not maximum security at the expense of functionality.

Conclusion: Optimize Your Cloudflare Bot Strategy

Configuring Cloudflare to properly allow verified bots represents a critical component of modern web security and SEO strategy. By leveraging Cloudflare's verified bot system and custom rules, you can maintain robust protection against malicious bot traffic while ensuring that search engines, monitoring services, and other beneficial bots can access your content without obstruction.

The key to successful bot management lies in understanding the distinction between different types of bots and implementing appropriate policies for each category. Search engine crawlers require unrestricted access to maintain your search visibility, while AI training crawlers might warrant blocking to protect your content. Social media preview bots enable better engagement when users share your content, while aggressive scrapers should be challenged or blocked.

Remember that bot management is not a set-and-forget configuration. Bot behaviors evolve, new bots emerge, and your business needs change over time. Regular monitoring, testing, and adjustment ensure your Cloudflare bot management rules continue serving your security and business objectives effectively.

Ready to optimize your Cloudflare bot management configuration? Start by auditing your current custom rules, implementing the strategies outlined in this guide, and monitoring your results. Your search rankings, site security, and operational efficiency will all benefit from a well-configured bot management strategy.

Stay Updated on Bot Management

Get the latest tips, guides, and updates on bot whitelisting, Cloudflare configurations, and website security delivered to your inbox.

No spam. Unsubscribe anytime.