LLM Optimizer MCP Server

1

Add it to Claude Code

Run this in a terminal.

Run in terminal
claude mcp add llm-optimizer -- npx -y @jonradoff/llmopt
README.md

An AI visibility intelligence platform for brands.

LLM Optimizer

LLM Optimizer is an AI visibility intelligence platform. It analyzes how large language models and AI search engines perceive, cite, and recommend brands — then provides research-backed optimization strategies to improve that visibility.

ChatGPT, Claude, Gemini, Perplexity, and Google AI Overviews are replacing traditional search for millions of people. The signals that determine whether an AI recommends your brand are fundamentally different from traditional SEO: earned media coverage, transcript quality, content structure, training data frequency, and citation network dynamics matter more than backlinks and keyword density. LLM Optimizer measures these signals across five analysis dimensions and produces a composite AI Visibility Score (0-100) with prioritized, actionable recommendations.

What It Analyzes

LLM Optimizer performs six types of analysis, each grounded in peer-reviewed research:

Answer Engine Optimization — Analyzes your website's content against the optimization strategies validated by the GEO (Generative Engine Optimization) research. Scores pages on quotation density (+41% visibility), statistical evidence (+33%), source citations (+28%), fluency, structural optimization, and machine readability. Produces per-question optimization scores with specific rewrite recommendations.

Video Authority Analysis — Two-phase analysis of YouTube presence. Phase 1 uses a fast model to assess individual videos for transcript quality, keyword alignment, and caption availability. Phase 2 feeds compact assessments into a reasoning model for four-pillar scoring: Transcript Authority, Topical Dominance, Citation Network, and Brand Narrative. Based on research showing YouTube is now the #1 social citation source for LLMs, appearing in 16% of AI answers.

Reddit Authority Analysis — Scrapes Reddit discussions mentioning your brand and analyzes community sentiment, competitive positioning, and training data signal strength. Uses Reddit's public .json endpoints with Cloudflare WARP proxy fallback. Scores four pillars: Presence, Sentiment, Competitive Position, and Training Signal.

Search Visibility Analysis — Evaluates your site's visibility across both Google AI Overviews and standalone LLMs. Checks robots.txt AI crawler policies, structured data, content freshness, brand search momentum, and earned media signals. Based on research showing only 12% overlap between Google top-10 results and ChatGPT/Perplexity citations.

LLM Knowledge Testing — Directly queries multiple LLM providers (Anthropic, OpenAI, Gemini, Grok) with your brand's target queries and analyzes how each model responds. Compares your brand's presence, accuracy, and recommendation likelihood across providers. Supports head-to-head competitor comparison.

Brand Intelligence — Aggregates all analysis dimensions into a composite AI Visibility Score weighted across Optimization (30%), Video Authority (20%), Reddit Authority (20%), Search Visibility (15%), and LLM Test (15%). Generates prioritized action items that track through to completion.

Research Foundation

The analysis methodology is grounded in published research. Key findings that inform the scoring:

  • Content optimization: Embedding authoritative quotations improves AI citation visibility by +41%; adding statistics by +33%; keyword stuffing reduces visibility by -9% (GEO, Princeton/KDD 2024).
  • Training data frequency: Answer accuracy more than doubles from rare (1-5 documents) to high-frequency (51+ documents) in training data. Being in training data AND being retrievable provides a compounding advantage (NanoKnow, 2026).
  • Source preferences: AI search engines cite earned media 72-92% of the time vs. 18-27% for brand-owned content. Only 15-50% overlap with traditional Google results (GEO, Toronto 2025).
  • Video transcripts: A 7B-parameter model trained on YouTube transcripts surpassed 72B models in commentary quality. Transcript quality is the dominant signal for video LLM influence — not production value or view counts (LiveCC, CVPR 2025).
  • Citation concentration: Top 20 news sources capture 28-67% of all AI citations depending on provi

Tools (1)

analyze_visibilityPerforms a comprehensive analysis of brand visibility across answer engines, video, and search.

Configuration

claude_desktop_config.json
{"mcpServers": {"llm-optimizer": {"command": "npx", "args": ["-y", "@jonradoff/llmopt"]}}}

Try it

Analyze my website's visibility for the query 'best enterprise software' and provide rewrite recommendations.
Perform a video authority analysis on my YouTube channel to improve my brand's citation network.
Compare my brand's presence and accuracy across Claude, OpenAI, and Gemini for my target keywords.
Generate a prioritized action plan to improve my AI Visibility Score based on current search signals.

Frequently Asked Questions

What are the key features of LLM Optimizer?

Measures AI visibility across five analysis dimensions including Answer Engine Optimization and Video Authority.. Produces a composite AI Visibility Score (0-100) with actionable recommendations.. Evaluates brand presence across Google AI Overviews, ChatGPT, Perplexity, and other LLMs.. Analyzes YouTube transcript quality and Reddit community sentiment for brand positioning.. Provides research-backed optimization strategies based on GEO (Generative Engine Optimization) findings..

What can I use LLM Optimizer for?

Marketing teams looking to improve brand citation frequency in AI-generated search results.. Content strategists needing to optimize website content for generative engine readability.. Video creators aiming to increase their authority and citation potential within AI search models.. Brand managers tracking competitive positioning across multiple LLM providers..

How do I install LLM Optimizer?

Install LLM Optimizer by running: npx -y @jonradoff/llmopt

What MCP clients work with LLM Optimizer?

LLM Optimizer works with any MCP-compatible client including Claude Desktop, Claude Code, Cursor, and other editors with MCP support.

Turn this server into reusable context

Keep LLM Optimizer docs, env vars, and workflow notes in Conare so your agent carries them across sessions.

Need the old visual installer? Open Conare IDE.
Open Conare