Back to blog
AEOAI CitationsOriginal ResearchState of AI Citations
FogTrail Team··Updated

Linear outranks Monday.com for "best alternative to Jira" in 2 of 5 AI engines. Beehiiv now beats Mailchimp for newsletter queries in 3 of 5, up from 2 in Wave 1. PostHog surged to 5 citations across all 5 engines in Wave 3, the most-cited startup in our entire dataset and the only brand with three consecutive waves of citation growth (2, 3, 5).

These are not flukes. They share a pattern: tight positioning around a specific use case, strong community reputation, and a willingness to own a niche rather than chase broad category dominance. In AI search, that strategy works. But the picture is more nuanced than it first appears. While individual startups are winning, the overall rate at which engines recommend startups first has declined.

Context: Why This Matters

AI search engines do not simply mirror market share. They reward niche positioning, community reputation, and content depth in ways that create specific openings for startups. These findings come from FogTrail's ongoing citation analysis, now spanning three weekly waves. We track 25 B2B SaaS brands across 20 queries and 5 AI search engines (ChatGPT, Perplexity, Gemini, Grok, and Claude), producing 300 engine-query pairs across three waves. The dataset captures how each engine ranks, mentions, and cites brands in real buyer queries. As of March 2026, the results show that AI search engines do not simply mirror market share. They reward something else entirely.

For context on how these engines select what to recommend, see our breakdown of how AI search engines decide what to cite.

The Data: Three Startups, Three Upsets

Three startups, Linear, Beehiiv, and PostHog, each outrank their respective category leaders on multiple AI engines for specific, intent-rich queries. Here is the data from each case.

Case 1: Linear vs. Monday.com (Project Management)

Query: "best alternative to Jira"

Engine#1 Recommendation
PerplexityLinear
ChatGPTLinear
GeminiLinear
GrokClickUp
ClaudeAsana

Linear holds 8 total mentions across the entire dataset, zero formal citations, and a 63% position-1 rate (5 of its 8 mentions are at the top spot). Monday.com has 13 total mentions, 5 citations, and a 15% position-1 rate. Asana has 16 mentions, zero citations, and a 50% position-1 rate.

The numbers are stark: Monday.com gets mentioned 63% more often than Linear across the full query set, and it has actual linked citations to back it up. None of that mattered for this query. ChatGPT described Linear as designed for "speed-obsessed engineering teams." Perplexity also led with Linear despite listing Monday.com, ClickUp, and Asana as alternatives.

Case 2: Beehiiv vs. Mailchimp (Email Marketing)

Query: "what email tool should I use for newsletters"

Engine#1 Recommendation
PerplexityMailchimp
ChatGPTBeehiiv
GeminiMailchimp
GrokBeehiiv
ClaudeMailchimp

Beehiiv has 6 total mentions across the Wave 3 dataset. Mailchimp has 18. That is a significant gap in raw visibility. But Beehiiv achieved a milestone in Wave 3: majority consensus (3 of 5 engines) for the newsletter query, up from 2 of 5 in Waves 1 and 2. Gemini flipped from Mailchimp to Beehiiv, joining ChatGPT and Grok.

ChatGPT recommended Beehiiv first for newsletters, describing it as the best option for "a media-style newsletter, with growth tools built in." Gemini's flip is particularly significant because Gemini had been a consistent Mailchimp-first engine across two previous waves. Beehiiv is now at 3 engines to Mailchimp's 2 for this specific query. But when the query broadens to "email marketing software comparison," Mailchimp and ActiveCampaign reclaim the top spots. The lesson: Beehiiv wins the narrow query, not the broad one.

Case 3: PostHog vs. Amplitude (Analytics)

Query: "best analytics tool for SaaS"

Engine#1 Recommendation
PerplexityAmplitude
ChatGPTPostHog
GeminiAmplitude
GrokAmplitude
ClaudeMixpanel

PostHog's trajectory across three waves tells the strongest startup success story in our dataset. Its citation count grew from 2 to 3 to 5 across three consecutive waves, making it the only brand with three straight weeks of citation increases. In Wave 3, PostHog earned citations from all 5 engines (ChatGPT, Perplexity, Gemini, Grok, and Claude), matching enterprise brand Salesforce's citation count. PostHog has 16 total mentions and 5 citations, a 31% citation rate that is the highest of any startup. By contrast, Linear has 11 mentions and 1 citation (9%). PostHog's open-source model, strong documentation, and active community presence are generating the signals that AI engines increasingly reward with direct URL links, not just text mentions.

The Startup-Friendliness Gap Across Engines

Not all engines treat startups equally. The size-based analysis reveals a significant spread, and the trend across three waves shows that startup-friendliness is volatile, not rising.

EngineStartup at #1 (W1)W2W3Trend
ChatGPT5 (25%)5 (25%)3 (15%)Declining
Claude2 (10%)3 (15%)1 (5%)Declining
Gemini1 (5%)2 (10%)2 (10%)Stable
Grok1 (5%)2 (10%)2 (10%)Stable
Perplexity0 (0%)1 (5%)1 (5%)Stable

ChatGPT was the most startup-friendly engine in Waves 1 and 2, placing startup brands at position 1 in 25% of queries. In Wave 3, that rate dropped to 15%. Claude declined even more sharply, from 15% to 5%. No engine now exceeds 15% startup-at-#1 rate. The average startup brand gets 7.1 mentions across the Wave 3 dataset, compared to 17.3 for enterprise brands and 15.4 for midmarket. Startups appear on an average of 3.5 engines versus 5.0 for enterprise brands.

SizeBrandsAvg Mentions/Brand (W3)Avg Engines/BrandTotal Citations (W3)
Enterprise (6)HubSpot, Salesforce, Mailchimp, Asana, GA, Monday.com17.35.018
Midmarket (8)Pipedrive, Mixpanel, Amplitude, ActiveCampaign, ClickUp, ConvertKit, Vercel, Netlify15.44.918
Startup (11)Close, Attio, Linear, Height, Beehiiv, Loops, PostHog, Heap, Railway, Render, Fly.io7.13.512

Startups get roughly 41% of the visibility that enterprise brands get on a per-brand basis. But when they do break through, they tend to land at position 1, not position 3. The three case studies above are not anomalies. They represent the characteristic startup pattern in AI search: rarely mentioned, but when mentioned, often recommended first.

The nuance from three waves of data: while individual startups like PostHog and Beehiiv are winning in their niches, the overall startup-at-#1 rate is not trending upward. The enterprise advantage in AI recommendations is resilient. Startup success in AI search is specific, not systematic.

What These Three Startups Have in Common

The shared traits are not subtle.

Tight use-case positioning. Linear is not a "project management tool." It is a tool for engineering teams that want speed. Beehiiv is not an "email marketing platform." It is a newsletter platform. PostHog is not "analytics." It is open-source product analytics for startups. Each brand owns a specific job-to-be-done rather than competing across the full category.

This mirrors what worked in traditional SEO with long-tail keywords. Broad category terms ("best CRM") go to incumbents. Specific intent queries ("best alternative to Jira for engineering teams") go to whoever owns that niche most credibly. AI engines appear to operate on a similar principle, amplified by the fact that they synthesize community sentiment and product positioning rather than just link authority.

Strong developer and community reputation. All three brands have outsized presence in developer communities, open-source ecosystems, or niche creator communities. PostHog is open source. Linear has a near-cult following among developers. Beehiiv has become the default among independent newsletter creators. AI models, trained on vast amounts of community discussion, absorb these reputational signals. For more on the mechanics of this, see our analysis of how LLMs decide what to cite.

They are not trying to be everything. Monday.com positions itself across project management, CRM, operations, and marketing. Mailchimp now encompasses email, websites, social, and SMS. Amplitude covers analytics, experimentation, and CDP. Broader positioning means more total mentions, but it dilutes the signal for any specific query. The startups win precisely because they are narrower.

What This Does Not Mean

Startup wins on specific queries do not translate to broad category dominance. Incumbents still control generic category queries decisively. In our dataset, incumbents still dominate "alternative to" queries 87% of the time, stable across two consecutive waves. Mailchimp leads in its email marketing mentions overall. HubSpot and Salesforce control CRM. Vercel holds an 88% position-1 rate across Dev Tools (down from 100% in Wave 1, but still dominant). And ChatGPT's startup-at-#1 rate dropped from 25% to 15% in Wave 3.

Startups win in specific, intent-rich queries where their positioning aligns exactly with what the user asked for. That is a meaningful but bounded advantage. If you are a startup, the opportunity is real, but only if you know which queries you can win and which engines give you the best shot.

What You Can Do About It

  • Identify your winnable queries. Find the specific use-case queries where your positioning is strongest. "Best newsletter platform" is winnable for Beehiiv. "Best email marketing software" is not. Map your product to the exact job-to-be-done queries where you have a credible claim.

  • Prioritize ChatGPT for early traction, but diversify. ChatGPT had a 25% startup-at-#1 rate in our first two waves but dropped to 15% in Wave 3. It still links to brand websites in 18% of its citations, more than most engines. Optimize your site content, pricing pages, and documentation for ChatGPT, but do not rely on a single engine. Startup-friendliness is volatile.

  • Invest in community, not just content. AI models absorb community reputation. Reddit threads, GitHub discussions, and niche community praise all feed the training data and real-time search context that engines use. This is not traditional content marketing. It is reputation marketing.

  • Monitor across all five engines. Your startup might be invisible on Perplexity while holding position 1 on ChatGPT. Single-engine monitoring gives you an incomplete picture. Track your position across ChatGPT, Perplexity, Gemini, Grok, and Claude to understand your actual AI visibility.

  • Do not chase broad category terms. Enterprise brands have structural advantages in broad queries. Compete where your specificity is an asset, not a liability.

Methodology

This analysis is based on FogTrail's ongoing citation study: 20 queries across 5 B2B SaaS categories, sent to 5 AI search engines (ChatGPT, Perplexity, Gemini, Grok, Claude) via real-time API calls simulating actual user searches, repeated across three weekly waves in March 2026. 25 brands were tracked across 300 engine-query pairs. Brand sizes were defined prior to data collection based on company stage and revenue. Wave-over-wave trends are noted throughout.

Frequently Asked Questions

Do startups need more total mentions to compete in AI search?

No. Linear has 8 total mentions versus Monday.com's 13, yet Linear holds position 1 in 63% of its appearances. Total mentions measure presence, not prominence. Position and context matter more than volume.

Which AI engine should startups focus on first?

ChatGPT remains the most startup-friendly engine, though its startup-at-#1 rate dropped from 25% to 15% in our third wave. It links to brand websites in 18% of its citations and still elevates startup brands above incumbents for specific use-case queries. But diversifying across engines is important, as startup-friendliness rates are volatile.

Does this mean market share does not matter for AI search rankings?

Market share still matters for broad category queries. HubSpot, Salesforce, and Mailchimp dominate their respective categories overall. But for narrow, intent-specific queries, positioning and community reputation can override market share.

How often do AI search engine rankings change?

AI citation rankings are not static. Engine models update regularly, and the sources they draw from shift over time. See our research on how often AI search engines update citations for more detail.

Can these results be replicated?

Query results from AI engines are non-deterministic, meaning identical queries can produce slightly different responses. The patterns we report are based on structured, controlled queries run at the same time across all engines, but individual responses may vary on repeat runs.

Related Resources