Hey @Suvi Lehtovaara,
@mitchell.gordon has a great answer here but let me post my thoughts and some steps that can be taken today to help with this.
I run Product for our AI and Search teams on Customer Community, and this has been on my mind for a while now.
So llms.txt has been discussed as the new standard to drive AI crawlers to site but as this is not yet standard and might never be, still the best place to handle this is what is include in robots.txt. This is not a guarantee but helps or at least does not restrict the AI crawlers from accessing the data found on your community.
Some recommendations for AI Crawler rules inside robots.txt:
# ===========================================
# AI CRAWLERS
# ===========================================
# --- OPENAI ---
User-agent: GPTBot
Allow: /
User-agent: OAI-SearchBot
Allow: /
User-agent: ChatGPT-User
Allow: /
# --- ANTHROPIC ---
User-agent: ClaudeBot
Allow: /
User-agent: Claude-User
Allow: /
User-agent: Claude-SearchBot
Allow: /
# --- PERPLEXITY ---
User-agent: PerplexityBot
Allow: /
# --- GOOGLE ---
User-agent: Google-Extended
Allow: /
# --- APPLE ---
User-agent: Applebot-Extended
Allow: /
# --- COMMON CRAWL ---
User-agent: CCBot
Allow: /
Your robots.txt can be edited in Control and is found here: {your-community-name}.insided.com/seo, however for a use case of maximum AI visibility, I would suggest to use the wildcard. A long explicit list per bot is just noise unless you need specific per-bot rules.
The bottom line is currently, GEO is 80% traditional SEO, so with good technical SEO, fast page speed, applied server-side rendering, use of FAQ / Org schema, with content properly structured for extraction, and fresh content, everything that worked before still should apply to LLMs.
Cheers,
Graeme