Skip to main content

Server Logs Analytics

Server Logs Analytics reveals how AI companies access your website by analyzing your raw web server access logs. This is the ground truth of AI visibility—direct evidence of which AI crawlers are visiting your pages and when. Use it to verify access, find issues with robots.txt, prioritize content, and track trends over time.

Why this matters and where to find your logs

Ansehn - Server Logs Analytics Dashboard for AI Search Monitoring and Optimization

What you can do with Server Logs Analytics

  • Verify AI access: Confirm which AI companies are crawling your site (GPTBot, OAI-SearchBot, ClaudeBot, PerplexityBot, Gemini, Google-Extended, and more).
  • Identify blocked crawlers: Spot missing bots and diagnose robots.txt issues.
  • Track trends: See daily activity by crawler to monitor increases, drops, and spikes.
  • Prioritize content: Focus on pages most visited by AI crawlers.
  • Troubleshoot: Detect access errors or anomalies early.

💡 Pro tip: Compare crawler activity here with your insights in Citation Analytics to understand how crawling correlates with actual citations in AI answers.

How to import your logs

There are two ways to get your server logs into Ansehn:

MethodBest forSetup effort
Manual UploadOne-time analysis, quick checks, any web serverLow — upload a .log or .txt file from your dashboard
AWS CloudFront IntegrationContinuous, automated monitoring of CloudFront distributionsMedium — one-time AWS setup, then fully automated

Supported AI crawlers

The dashboard automatically detects these crawlers (case-insensitive):

  • OpenAI: GPTBot/1.2, OAI-SearchBot/1.0, ChatGPT-User/1.0
  • Anthropic: ClaudeBot, Claude-User, Claude-SearchBot
  • Perplexity: PerplexityBot/1.0, Perplexity-User
  • Google: Gemini, Google-Extended
  • Meta: meta-externalagent (if present in your logs)
  • ... and more

Parsing & filtering rules

The parser automatically:

  • Filters to include only AI crawler traffic
  • Excludes non-200 responses from page statistics
  • Removes /robots.txt requests from page rankings
  • Groups data by date and crawler type

Dashboard overview

Daily Visits from AI Crawlers

Shows the daily volume of requests from AI crawlers and bot types. Each color represents a different crawler, stacked to show total daily activity.

What to look for:

  • Increasing activity: More frequent crawling due to fresh content or rising visibility
  • Sudden spikes: New content discovery, viral mentions, or model retraining cycles
  • Declines: Potential blocking via robots.txt, stale content, or technical issues
  • Missing crawlers: Verify robots.txt and allow rules for key bots

Take action:

  • Publish regularly so crawlers return more often
  • Investigate drops quickly to prevent visibility loss
  • Prioritize high-traffic days/pages for optimization
  • Update sitemaps to accelerate discovery

💡 Pro tip: Cross-reference with Citation Analytics to see if crawl volume correlates with citations in answers.

Top Pages (by AI crawler visits)

Highlights which pages AI crawlers visit most, with a breakdown by bot type. Pages with high visits are likely candidates for training data or citations.

Use this to:

  • Prioritize updates to pages that attract AI crawlers
  • Identify underperforming pages and content gaps
  • Replicate successful structures/topics
  • Detect if critical pages are unintentionally blocked

The dashboard displays up to 100 top pages, paginated 5 at a time.

Top Crawlers

Visualizes relative activity levels across AI crawlers. A balanced shape suggests broad AI visibility; dominance indicates one platform finds your content especially relevant.

Signals to watch:

  • Balanced distribution: Healthy visibility across platforms
  • Missing segments: Possible blocking or limited relevance
  • High "User" activity (ChatGPT-User, Perplexity-User): Real users retrieving your content via AI search

Learn more