- Mar 24, 2026
- 6
- 0
- Awards
- 3
- First Name
- Joe
I have to give you massive credit here, your diagnosis of the legacy tech bloat is spot on. A 70% 'Poor' PageSpeed rating across the industry is embarrassing, and building a clean, SSR-first infrastructure with pristine schema is exactly what the industry needs to fix the crawl budget issues. I am genuinely looking forward to seeing those technical benchmarks for @DealerInt.
Where I think the architecture still hits a wall is your assumption about how non-Google models acquire data, specifically regarding retail velocity versus index churn.
You mentioned that GMC feeds and SSR schema are how Perplexity and OpenAI 'know' a car arrived. GMC is fantastic for Google's ecosystem, but Google isn't sharing that structured feed with OpenAI or Anthropic. For those models to discover inventory without an aggregator, they are entirely reliant on their own web crawlers hitting your SSR schema.
Even with a lightning fast site, standard web crawling is fundamentally incompatible with automotive retail velocity. While an average unit might sit for 30 to 60 days, the highly desirable, aggressively priced inventory....the exact cars users are actively querying AI for, often move in a matter of days. If a foundational model's crawler only indexes a specific VDP once a week, the AI is going to confidently send shoppers to 404 pages and sold vehicles. Models cannot tolerate that level of hallucination risk.
The 'Aggregator Tax' isn't just a visibility tax; it's a data-licensing reality. Foundational models are striking massive enterprise data deals with centralized hubs precisely because they need a real-time API firehose, not because they want to rely on crawling decentralized local domains, no matter how fast or clean your pipe is.
I completely agree that your infrastructure will absolutely crush legacy platforms on standard Google crawlability. But until OpenAI decides to trust and query thousands of decentralized MCP endpoints instead of buying a clean, normalized data feed from a centralized network, the aggregators still hold the keys to the non-Google LLM intelligence layers.
Where I think the architecture still hits a wall is your assumption about how non-Google models acquire data, specifically regarding retail velocity versus index churn.
You mentioned that GMC feeds and SSR schema are how Perplexity and OpenAI 'know' a car arrived. GMC is fantastic for Google's ecosystem, but Google isn't sharing that structured feed with OpenAI or Anthropic. For those models to discover inventory without an aggregator, they are entirely reliant on their own web crawlers hitting your SSR schema.
Even with a lightning fast site, standard web crawling is fundamentally incompatible with automotive retail velocity. While an average unit might sit for 30 to 60 days, the highly desirable, aggressively priced inventory....the exact cars users are actively querying AI for, often move in a matter of days. If a foundational model's crawler only indexes a specific VDP once a week, the AI is going to confidently send shoppers to 404 pages and sold vehicles. Models cannot tolerate that level of hallucination risk.
The 'Aggregator Tax' isn't just a visibility tax; it's a data-licensing reality. Foundational models are striking massive enterprise data deals with centralized hubs precisely because they need a real-time API firehose, not because they want to rely on crawling decentralized local domains, no matter how fast or clean your pipe is.
I completely agree that your infrastructure will absolutely crush legacy platforms on standard Google crawlability. But until OpenAI decides to trust and query thousands of decentralized MCP endpoints instead of buying a clean, normalized data feed from a centralized network, the aggregators still hold the keys to the non-Google LLM intelligence layers.