How to get AI to recommend your business

A website that takes more than three seconds to load gets left out of AI search results. For years, page load speed was a Google ranking factor: it affected rankings, bounce rates, and conversions. With the rise of generative search engines —ChatGPT, Perplexity, Gemini, Google AI Overviews— the bar for source selection is much higher. And your infrastructure’s speed plays a direct role in whether your content appears in the responses received by millions of users.

In this post, we outline how a CDN can boost AI mentions of your business — and what bot management strategies you can implement to make automated reading easier.

How AI Crawlers Work

AI system crawlers make real-time web queries when building their responses. When a user asks “what’s the best CDN provider in Europe?”, the system fires requests at search results, visits candidate pages, and extracts the content it needs to construct its answer. The entire process takes seconds. And if your server doesn’t respond in time, the crawler moves on to the next source.

The figure that circulates in the GEO (Generative Engine Optimization) space is concrete: AI engine retrieval systems operate with timeouts of between 1 and 3 seconds.

This is not an officially published threshold by OpenAI, Google, or Anthropic. But the technical consensus among industry professionals is clear: if your server takes more than 2–3 seconds to return content, the risk of your page being excluded from the set of sources the system considers increases significantly.

Content delivery networks can significantly improve TTFB — Time to First Byte. We’ve explained in previous posts how CDNs work, but to give you an idea: with an origin server in Central Europe, a crawler in the US can experience a TTFB of over 1 second. With a CDN, that number drops to milliseconds.

Bot Management: Open the Door to AI Crawlers

Reducing TTFB is a necessary condition, but not sufficient. There’s no point in responding in 80ms if your configuration blocks the crawler before it even starts measuring. And this happens more often than you’d think.

Many CDNs, hosting platforms, and CMS systems have started adding restrictive rules by default in response to increased AI bot traffic. The result: sites that have unknowingly blocked GPTBot, ClaudeBot, or PerplexityBot, and whose owners wonder why they don’t appear in generative responses.

The Key Distinction: Search vs. Training

Before configuring anything, it’s important to understand that AI crawlers don’t just do one thing. Most platforms use at least two separate functions with distinct user-agents: the training bot (which uses your content to improve the model) and the real-time search bot (which retrieves your content when a user asks a question).

To appear in generative responses, the search bot is what matters. You can block the training bot if you don’t want your content used as fine-tuning data, and still be cited in real-time responses.

Recommended Configuration

The first level of control is the robots.txt file. An explicit configuration avoids the most common mistake: having a wildcard User-agent: * Disallow: / with exceptions only for Googlebot, so all AI bots get caught in the generic rule unintentionally. Review your robots.txt and make sure AI bots have their own explicit Allow directive.

If you also use a bot management service like our Bot Management, you should keep the ai_scraper flag disabled to allow AI crawler access.

Conclusion

The relationship between web speed and AI mentions is not direct, but it is real and measurable. Generative search engines select sources in real time under strict time constraints. A slow server actively reduces the likelihood of your content appearing in ChatGPT, Perplexity, or Google AI Overviews responses.

But speed alone is not enough. Your robots.txt and CDN configuration determine whether those crawlers can even access your content. These are the two layers of the problem: first, open the door — then respond quickly once they’re in.

With Transparent Edge, both layers are managed in one place: low TTFB in any region thanks to the CDN, and granular control over which bots access your content through Bot Management and its ai_scraper flag.

These are just a few technical recommendations. They should be complemented by other strategic actions recommended in the new era of generative AI. Technology is a necessary condition. Citability is the goal.