Hi there,
This is Thomas, CEO at OtterlyAI. Over the last months, I’ve onboarded dozens of marketing teams to OtterlyAI and AI Search Optimization / Generative Engine Optimization.
In this article, I summarized all the GEO (Generative Engine Optimization) activities that truly matter.
These are proven tactics and efforts I have either tested first-hand, seen implemented by our partners and customers, or at least gathered enough data points to determine if they work.
This is not your manual, guide, or ultimate tutorial. These are MY recommendations and tactics that I would suggest to marketing teams at this point.
Generative Engine Optimization – Key Metrics
Traditional metrics like impressions, click and traffic no longer accurately measure AI search success. Consumer behavior and User Interfaces (UI) are different; therefore, we need to define new KPIs to measure success on ChatGPT, Perplexity, and AI search.
These are my GEO KPIs for measuring AI search success
You can check out the whole Definition of our KPIs here.
Why Do They Matter?
- It’s about how we show up and get recommended as a brand on ChatGPT and AI Search.
- Brand and brand influence define whether we are top of mind among our ideal customers.
- Instead of link tracking, it’s all about citations and whether our content is getting cited on AI Search.
So, now let’s see how we can improve those KPIs for higher AI search visibility.
How to get my website cited on ChatGPT?
These are the key steps to get your website cited on ChatGPT.
Step 1: Does my CDN/Web Host allow AI crawlers to access my website?
Some Content Delivery Networks (CDNs) and web hosts provide you with settings to control which bots can access your website and which cannot. Here’s a screenshot from Cloudflare—a popular CDN—where you can see the settings options:
- Block AI training bots
- Instruct AI bots via the robots.txt file
We found that many CDNs and web hosts block AI crawlers by default. Better be sure and send the support team of your CDN provider an email asking for clarity on this. You can have the best content in the world (or in your industry), yet it might still be invisible to AI search engines because your CDN is blocking their AI crawlers.
You can simulate whether different AI crawlers can access your website with the OtterlyAI Crawler Simulation Tool.
Step 2: Do we block or allow AI crawlers to access our websites via the robots.txt file
Another important housekeeping aspect: Which crawlers are allowed or disallowed via our robots.txt file?
- Allow by default: If a crawler is not mentioned in your
robots.txtfile, it’s effectively allowed to crawl your site. - Use the
Allow:directive: To explicitly permit certain crawlers or sections of your site for AI training, you can use a syntax such as:
User-agent: GPTBot
Allow: /training-data/
This allows the crawler to access only the specified directory.
Important disclaimer: The instructions in robots.txt files cannot enforce crawler behavior on your site; it’s up to the crawler to obey them. Most crawlers obey robots.txt, although it is not enforceable.
Step 3: How much dynamic content do we have on specific URLs of ours?
It’s important to check the ratio between static and dynamic content on our website, or on any particular URL. Why?
AI crawlers do not execute JavaScript. GPTBot and other OpenAI crawlers see raw HTML; they don’t reliably execute client-side JavaScript. If content is behind JavaScript rendering (client-side), that content may be “invisible.” In other words, if your page is a Single Page Application (SPA) or uses heavy client-side rendering such that the HTML skeleton is mostly empty or consists of placeholders and the “real” content comes later via JS, the AI crawlers may only see those placeholders. (source)
When you log into your OtterlyAI account and head over to the GEO audit, you can run any URL audit to check the static vs dynamic content ratio.
As you can see here, it’s important that we have a static ratio score, ensuring that our content is accessible and crawlable by different AI crawlers.
Bonus: Pros vs. Cons: Static HTML vs Dynamic/Client-Side Rendering (for AI Crawler Visibility)
Step 4: On Page – Content Chunk Analysis
Now it’s time to analyze and optimize our content for better retrieval. I currently rely heavily on this CustomGPT “GEO Optimizer” created by Andrei Iunisov. This custom GPT analyzes any content or URL based on the following checks:
Why is this check useful?
It makes our content better and, therefore, improves the chances of our content being retrieved for ChatGPT.
Step 5: On Page Content – Relevance Check
Next, we want to analyze our text for semantic similarity so we can see exactly which paragraph or section aligns best with our target term/target prompt. You can read more about content relevance here.
Personally, I’ve been using Relevance Doctor by iPullRank, a tool for measuring semantic similarity between prompts and content.
The goal is to have text sections that are semantically very similar and closely aligned with our target query. As you can see in the screenshot below, each paragraph is then scored with a similarity score.
Step 6: Content – Are the SEO basics in place?
In addition to the steps outlined above, I always recommend making sure that the following SEO basics are met. Consider them important add-ons to your Generative Engine Optimization strategy.
- A clean HTML structure
- Correct Heading Hierarchy
- Good navigational structure
- Correct Meta data
- Mobile-friendly design
- Clear language tags
Key Takeaways: How to Get Cited in AI Search
Want your content to show up in ChatGPT, Perplexity, Claude, and other AI search engines? These are your must-do moves:
- Check your CDN and hosting setup – Make sure AI crawlers aren’t blocked by default.
- Use
robots.txtintentionally – Explicitly allow AI bots like GPTBot access to your training-worthy pages. - Stick to static HTML for key content – If it’s hidden behind JavaScript, AI crawlers likely won’t see it.
- Structure content into clear, self-contained chunks – Use plain language, direct answers, and avoid clever fluff.
- Ensure high semantic relevance – Use tools to align content closely with the kinds of queries your audience asks.
- Don’t skip SEO basics – Clean code, proper headings, mobile-friendly design, and metadata still matter.
- Get your brand in listicles – These mentions boost both SEO and AI visibility. Link building services provider Editorial.Link helps get them featured.
🎯 Do these six things right, and you’re not just publishing content. You’re making it retrievable, relevant, and ready for the AI era.
GEO Optimization – Frequently Asked Questions (FAQs)
Q: Do all AI crawlers obey the robots.txt file?
A: Most reputable AI crawlers . including GPTBot (OpenAI), ClaudeBot (Anthropic), and PerplexityBot. respect the directives in your robots.txt file. However, this is voluntary. There is no enforcement mechanism beyond crawler compliance.
Q2: Will allowing GPTBot cause any privacy or legal issues?
A: Allowing GPTBot to crawl your site means your public content can be used to train AI / LLM models and inform AI responses. If your pages contain sensitive data, personal information, or regulated content (e.g. health, finance), you should consider excluding those paths via robots.txt. Always consult your legal team for jurisdiction-specific compliance.
Q3: How often should I check my content’s semantic relevance?
A: Ideally, audit your semantic relevance regularly (ideally monthly or quarterly) – especially after major content updates or changes in search behavior. Tools like Relevance Doctor or OtterlyAI’s GEO Audit can show how closely your content aligns with current AI search queries.