Octopus Energy is a useful AEO case because the problem was not writing more content. The problem was seeing how the brand appeared across AI answers in multiple markets without turning the process into manual spreadsheet work.

TL;DR

Octopus Energy used Ahrefs Brand Radar to replace a tedious manual process for checking AI visibility across ChatGPT, AI Overviews, and other AI-search platforms. The case does not prove a visibility lift, but it does show a more mature operating model: measure mentions, citations, markets, competitors, and source patterns before reshaping content strategy.

What was the situation?

Octopus Energy is a UK-headquartered energy company with retail energy, renewable generation, and low-carbon technology businesses. Ahrefs' case study says it supplies domestic electricity and gas to 7.7 million households across nine countries.

That market structure makes AEO harder than a single-domain audit. A company with multiple countries, older acquired brands, local competitors, and different regulatory contexts can show up differently in each AI answer surface. A prompt about switching energy suppliers in the UK may not behave like a prompt about renewable energy in Germany or Japan.

The core question was simple: when users ask AI systems about energy providers, does Octopus Energy show up, and is the answer accurate enough to trust?

What happened?

Ahrefs says Octopus Energy had been manually checking AI responses before adopting Brand Radar. Laura Iancu, an SEO Growth Specialist at Octopus Energy, described extracting AI responses from ChatGPT, AI Overviews, and other AI-search platforms as tedious.

Brand Radar replaced that manual process with a tool that could surface mentions, citations, and impressions across AI search and related channels. Ahrefs says the tool helped Octopus Energy present clear findings to executives and global marketing teams, which supported buy-in for content strategy and marketing initiatives.

That is the important result. The case is not a clean before-and-after citation-rate lift. It is a process lift: fewer one-off checks, clearer reporting, and better internal alignment around what AI systems were saying.

Why does this work at the answer-engine level?

AI visibility is distributed across prompts, surfaces, and sources. A brand can be mentioned without being cited, cited through a third-party page, omitted from comparison prompts, or described through an outdated source.

Ahrefs' methodology explains why prompt design matters. Brand Radar uses real search demand, People Also Ask style questions, and semantic fanout to create broader prompt coverage. It then stores AI responses so users can search mentions and citations across platforms such as ChatGPT, Perplexity, Gemini, Copilot, Google AI Overviews, and AI Mode.

For AEO, that is a better model than checking five handpicked prompts once. Visibility should be treated as a distribution. The brand's presence can vary by country, wording, competitor set, source freshness, and platform.

What evidence is strongest?

The strongest evidence is operational, not causal. Ahrefs reports that Brand Radar's output matched what Octopus Energy had been finding through manual checks, and that the tool made the work easier to explain and export.

That matters because many AEO tools fail at the internal-use layer. A technically interesting visibility metric is not useful if regional teams, executives, PR, and content leads cannot understand it. Octopus Energy's case shows the value of plain metrics: mentions, citations, impressions, competitors, and cited sources.

The case also has a credible pain point. Multi-market companies really do face brand-history problems. If older acquired names still appear in AI answers, or if different regions have different citation sources, the fix requires more than one global page update.

What is replicable?

The replicable part is the workflow. You do not need Ahrefs to copy the operating model, though a platform can make it easier.

Start with markets and prompt classes. For each market, define brand prompts, competitor prompts, category prompts, and switching or buying prompts. Run them across several answer engines. Record whether the brand appears, whether it is cited, which URL is cited, which competitor appears, and whether the description is accurate.

Then sort gaps by action. If the brand is absent from category prompts, the issue may be authority or third-party mentions. If it is cited from the wrong page, the issue may be documentation or page targeting. If an old brand appears, the issue may be entity cleanup. If competitors dominate comparison prompts, the issue may be product positioning or review-site coverage.

What is uncertain?

The case does not show that Brand Radar directly increased Octopus Energy's AI visibility. Ahrefs describes monitoring, reporting, and strategy support, not a controlled lift in citations or revenue.

That distinction matters. AEO teams should not report "we improved AI visibility" when the evidence only says "we can now measure AI visibility." Measurement is still valuable. It is just a different claim.

There is also tool-methodology risk. Any modeled prompt database reflects assumptions about what users ask, which engines are tested, and how frequently answers are refreshed. Ahrefs is more transparent than many vendors, but the numbers still represent potential visibility rather than exact audience reach.

How should teams apply it?

Teams should use this case to design AI visibility reporting before designing AI content. The order matters. If you publish first and measure later, you cannot tell whether the work changed anything.

Build a visibility dashboard with five columns: prompt, answer engine, brand mention, cited source, and action needed. Add market, language, and business line where relevant. Review it every two weeks with content, PR, product marketing, and regional leads.

The output should not be a vanity score. It should be a work queue. Which page needs rewriting? Which outdated third-party source needs correction? Which comparison query needs a stronger proof asset? Which brand entity is confusing the answer engine?

What to do Monday morning

1. Pick one market and one product line. Do not start global. 2. Create 20 prompts from sales questions, support questions, category terms, and competitor comparisons. 3. Run the same prompts across three answer engines and log mentions, citations, and answer accuracy. 4. Tag each cited source by type: owned page, review site, news article, Reddit, documentation, or official source. 5. Turn the top five gaps into content, PR, documentation, or product-marketing tasks. 6. Repeat the prompt panel every two weeks before expanding to more markets.

Sources