{"id":1130,"date":"2026-01-09T09:47:23","date_gmt":"2026-01-09T01:47:23","guid":{"rendered":"http:\/\/longzhuplatform.com\/?p=1130"},"modified":"2026-01-09T09:47:23","modified_gmt":"2026-01-09T01:47:23","slug":"state-of-ai-search-optimization-2026-via-sejournal-kevin_indig","status":"publish","type":"post","link":"http:\/\/longzhuplatform.com\/?p=1130","title":{"rendered":"State Of AI Search Optimization 2026 via @sejournal, @Kevin_Indig"},"content":{"rendered":"<p><\/p> <div id=\"narrow-cont\"> <p data-pm-slice=\"1 1 [\" list=\"\"><em>Boost your skills with Growth Memo\u2019s weekly expert insights. Subscribe for free!<\/em><\/p> <p>Every year, after the winter holidays, I spend a few days ramping up by gathering the context from last year and reminding myself of where my clients are at. I want to use the opportunity to share my understanding of where we are with AI Search, so you can quickly get back into the swing of things.<\/p> <p>As a reminder, the vibe around ChatGPT turned a bit sour at the end of 2025:<\/p> <ul> <li>Google released the superior Gemini 3, causing Sam Altman to announce a Code Red (ironically, three years after Google did the same at the launch of ChatGPT 3.5).<\/li> <li>OpenAI made a series of circular investments that raised eyebrows and questions about how to finance them.<\/li> <li>ChatGPT, which sends the majority of all LLMs, reaches at most 4% of the current organic (mostly Google) referral traffic.<\/li> <\/ul> <p>Most of all, we still don\u2019t know the value of a mention in an AI response. However, the topic of AI and LLMs couldn\u2019t be more important because the Google user experience is turning from a list of results to a definitive answer.<\/p> <p>A big \u201cthank you\u201d to Dan Petrovic\u00a0and Andrea Volpini\u00a0for reviewing my draft and adding meaningful concepts.<\/p> <figure id=\"attachment_564369\" class=\"wp-caption aligncenter\" style=\"width: 1536px\"><img decoding=\"async\" src=\"https:\/\/cdn.searchenginejournal.com\/wp-content\/uploads\/2026\/01\/ai-search-927.jpg\" alt=\"AI Search Optimization\" width=\"1536\" height=\"1024\" class=\"size-full wp-image-564369\" srcset=\"https:\/\/cdn.searchenginejournal.com\/wp-content\/uploads\/2026\/01\/ai-search-927-384x256.jpg 384w, https:\/\/cdn.searchenginejournal.com\/wp-content\/uploads\/2026\/01\/ai-search-927-425x283.jpg 425w, https:\/\/cdn.searchenginejournal.com\/wp-content\/uploads\/2026\/01\/ai-search-927-480x320.jpg 480w, https:\/\/cdn.searchenginejournal.com\/wp-content\/uploads\/2026\/01\/ai-search-927-680x453.jpg 680w, https:\/\/cdn.searchenginejournal.com\/wp-content\/uploads\/2026\/01\/ai-search-927-768x512.jpg 768w, https:\/\/cdn.searchenginejournal.com\/wp-content\/uploads\/2026\/01\/ai-search-927-850x567.jpg 850w, https:\/\/cdn.searchenginejournal.com\/wp-content\/uploads\/2026\/01\/ai-search-927-1024x683.jpg 1024w, https:\/\/cdn.searchenginejournal.com\/wp-content\/uploads\/2026\/01\/ai-search-927-1280x720.jpg 1280w, https:\/\/cdn.searchenginejournal.com\/wp-content\/uploads\/2026\/01\/ai-search-927-1300x680.jpg 1300w, https:\/\/cdn.searchenginejournal.com\/wp-content\/uploads\/2026\/01\/ai-search-927.jpg 1536w\" sizes=\"auto, (max-width: 1536px) 100vw, 1536px\" loading=\"lazy\" title=\"State Of AI Search Optimization 2026 via @sejournal, @Kevin_Indig\u63d2\u56fe\" \/><figcaption class=\"wp-caption-text\">Image Credit: Kevin Indig<\/figcaption><\/figure> <h2><strong>Retrieved \u2192 Cited \u2192 Trusted<\/strong><\/h2> <p>Optimizing for AI search visibility follows a pipeline similar to the classic \u201ccrawl, index, rank\u201d for search engines:<\/p> <ol> <li>Retrieval systems decide which pages enter the candidate set.<\/li> <li>The model selects which sources to cite.<\/li> <li>Users decide which citation to trust and act on.<\/li> <\/ol> <p>Caveats:<\/p> <ol> <li>A lot of the recommendations overlap strongly with common SEO best practices. Same tactics, new game.<\/li> <li>I don\u2019t pretend to have an exhaustive list of everything that works.<\/li> <li>Controversial factors like schema or llms.txt are not included.<\/li> <\/ol> <h3><strong>Consideration: Getting Into The Candidate Pool<\/strong><\/h3> <p>Before any content enters the model\u2019s consideration (grounding) set, it must be crawled, indexed, and fetchable within milliseconds during real-time search.<\/p> <p>The factors that drive consideration are:<\/p> <ul> <li>Selection Rate and Primary Bias.<\/li> <li>Server response time.<\/li> <li>Metadata relevance.<\/li> <li>Product feeds (in ecommerce).<\/li> <\/ul> <h4><strong>1. Selection Rate And Primary Bias<\/strong><\/h4> <ul> <li><strong>Definition:<\/strong>\u00a0Primary bias measures the brand-attribute associations a model holds before grounding in live search results. Selection Rate measures how frequently the model chooses your content from the retrieval candidate pool.<\/li> <li><strong>Why it matters:<\/strong> LLMs are biased by training data. Models develop confidence scores for brand-attribute relationships (e.g., \u201ccheap,\u201d \u201cdurable,\u201d \u201cfast\u201d) independent of real-time retrieval. These pre-existing associations influence citation likelihood even when your content enters the candidate pool.<\/li> <li><strong>Goal:<\/strong>\u00a0Understand which attributes the model associates with your brand and how confident it is in your brand as an entity. Systematically strengthen those associations through targeted on-page and off-page campaigns.<\/li> <\/ul> <h4><strong>2. Server Response Time<\/strong><\/h4> <ul> <li><strong>Definition:<\/strong>\u00a0The time between a crawler request and the server\u2019s first byte of response data (TTFB = Time To First Byte).<\/li> <li><strong>Why it matters:<\/strong> When models need web results for reasoning answers (RAG), they need to retrieve the content like a search engine crawler. Even though retrieval is mostly index-based, faster servers help with rendering, agentic workflows, and freshness, and compound query fan-out. LLM retrieval operates under tight latency budgets during real-time search. Slow responses prevent pages from entering the candidate pool because they miss the retrieval window. Consistently slow response times trigger crawl rate limiting.<\/li> <li><strong>Goal:<\/strong>\u00a0Maintain server response times\u00a0&lt;200ms. Sites with &lt;1s load times receive \u00a03x more\u00a0Googlebot requests than sites &gt;3s. For LLM crawlers (GPTBot, Google-Extended), retrieval windows are even tighter than traditional search.<\/li> <\/ul> <h4><strong>3. Metadata Relevance<\/strong><\/h4> <ul> <li><strong>Definition:<\/strong>\u00a0Title tags, meta descriptions, and URL structure that LLMs parse when evaluating page relevance during live retrieval.<\/li> <li><strong>Why it matters:<\/strong> Before picking content to form AI answers, LLMs parse titles for topical relevance, descriptions as document summaries, and URLs as context clues for page relevance and trustworthiness.<\/li> <li><strong>Goal:<\/strong>\u00a0Include target concepts in titles\u00a0<em>and<\/em>\u00a0descriptions (!) to match user prompt language. Create keyword-descriptive URLs, potentially even including the current year to signal freshness.<\/li> <\/ul> <h4><strong>4. Product Feed Availability (Ecommerce)<\/strong><\/h4> <ul> <li><strong>Definition:<\/strong>\u00a0Structured product catalogs submitted directly to LLM platforms with real-time inventory, pricing, and attribute data.<\/li> <li><strong>Why it matters:<\/strong>\u00a0Direct feeds bypass traditional retrieval constraints and enable LLMs to answer transactional shopping queries (\u201dwhere can I buy,\u201d \u201cbest price for\u201d) with accurate, current information.<\/li> <li><strong>Goal:<\/strong>\u00a0Submit merchant-controlled product feeds to ChatGPT\u2019s merchant program (chatgpt.com\/merchants) in JSON, CSV, TSV, or XML format with complete attributes (title, price, images, reviews, availability, specs). Implement ACP (Agentic Commerce Protocol) for agentic shopping.<\/li> <\/ul> <h3><strong>Relevance: Being Selected For Citation<\/strong><\/h3> <p>\u201cThe Attribution Crisis in LLM Search Results\u201d (Strauss et al., 2025) reports low citation rates even when models access relevant sources.<\/p> <ul> <li>24% of ChatGPT (4o) responses are generated without explicitly fetching any online content.<\/li> <li>Gemini provides no clickable citation in 92% of answers.<\/li> <li>Perplexity visits about 10 relevant pages per query but cites only three to four.<\/li> <\/ul> <p>Models can only cite sources that enter the context window. Pre-training mentions often go\u00a0unattributed. Live retrieval adds a URL, which enables attribution.<\/p> <h4><strong>5. Content Structure<\/strong><\/h4> <ul> <li><strong>Definition:<\/strong>\u00a0The semantic HTML hierarchy, formatting elements (tables, lists, FAQs), and fact density that make pages machine-readable.<\/li> <li><strong>Why it matters:<\/strong>\u00a0LLMs extract and cite specific passages. Clear structure makes pages easier to parse and excerpt. Since prompts average\u00a05x the length of keywords, structured content answering multi-part questions outperforms single-keyword pages.<\/li> <li><strong>Goal:<\/strong>\u00a0Use semantic HTML with clear H-tag hierarchies, tables for comparisons, and lists for enumeration. Increase\u00a0fact and concept density\u00a0to maximize snippet contribution probability.<\/li> <\/ul> <h4><strong>6. FAQ Coverage<\/strong><\/h4> <ul> <li><strong>Definition:<\/strong> Question-and-answer sections that mirror the conversational phrasing users employ in LLM prompts.<\/li> <li><strong>Why it matters:<\/strong>\u00a0FAQ formats align with how users query LLMs (\u201dHow do I\u2026,\u201d \u201cWhat\u2019s the difference between\u2026\u201d). This structural and linguistic match increases citation and mention likelihood compared to keyword-optimized content.<\/li> <li><strong>Goal:<\/strong>\u00a0Build FAQ libraries from real customer questions (support tickets, sales calls, community forums) that capture emerging prompt patterns. Monitor FAQ freshness through lastReviewed or DateModified schema.<\/li> <\/ul> <h4><strong>7. Content Freshness<\/strong><\/h4> <ul> <li><strong>Definition:<\/strong>\u00a0Recency of content updates as measured by \u201clast updated\u201d timestamps and actual content changes.<\/li> <li><strong>Why it matters:<\/strong>\u00a0LLMs parse last-updated metadata to assess source recency and prioritize recent information as more accurate and relevant.<\/li> <li><strong>Goal:<\/strong> Update content within the past three months for maximum performance. Over\u00a070% of pages\u00a0cited by ChatGPT were updated within 12 months, but content updated in the\u00a0last three months\u00a0performs best across all intents.<\/li> <\/ul> <h4><strong>8. Third-Party Mentions (\u201dWebutation\u201d)<\/strong><\/h4> <ul> <li><strong>Definition:<\/strong>\u00a0Brand mentions, reviews, and citations on external domains (publishers, review sites, news outlets) rather than owned properties.<\/li> <li><strong>Why it matters:<\/strong>\u00a0LLMs weigh external validation more heavily than self-promotion the closer user intent comes to a purchase decision. Third-party content provides independent verification of claims and establishes category relevance through co-mentions with recognized authorities. They increase the entitithood inside large context graphs.<\/li> <li><strong>Goal:<\/strong>\u00a085% of brand mentions\u00a0in AI search for high purchase intent prompts come from third-party sources. Earn\u00a0contextual backlinks\u00a0from authoritative domains and maintain complete profiles on category\u00a0review platforms.<\/li> <\/ul> <h4><strong>9. Organic Search Position<\/strong><\/h4> <ul> <li><strong>Definition:<\/strong>\u00a0Page ranking in traditional search engine results pages (SERPs) for relevant queries.<\/li> <li><strong>Why it matters:<\/strong>\u00a0Many LLMs use search engines as retrieval sources. Higher organic rankings increase the probability of entering the LLM\u2019s candidate pool and receiving citations.<\/li> <li><strong>Goal:<\/strong>\u00a0Rank in Google\u2019s top 10 for fan-out query variations around your core topics, not just head terms. Since LLM prompts are conversational and varied, pages ranking for many long-tail and question-based variations have higher citation probability. Pages in the top 10 show a\u00a0strong correlation\u00a0(~0.65) with LLM mentions, and\u00a076% of AI Overview citations\u00a0pull from these positions. Caveat: Correlation varies by LLM. For example, overlap is\u00a0high for AI Overviews\u00a0but\u00a0low for ChatGPT.<\/li> <\/ul> <h3><strong>User Selection: Earning Trust And Action<\/strong><\/h3> <p>Trust is critical because we\u2019re dealing with a single answer in AI search, not a list of search results. Optimizing for trust is similar to optimizing for click-through rates in classic search, just that it takes longer and is harder to measure.<\/p> <h4><strong>10. Demonstrated Expertise<\/strong><\/h4> <ul> <li><strong>Definition:<\/strong>\u00a0Visible credentials, certifications, bylines, and verifiable proof points that establish author and brand authority.<\/li> <li><strong>Why it matters:<\/strong>\u00a0AI search delivers single answers rather than ranked lists. Users who click through require stronger\u00a0trust signals\u00a0before taking action because they\u2019re validating a definitive claim.<\/li> <li><strong>Goal:<\/strong>\u00a0Display author credentials, industry certifications, and verifiable proof (customer logos, case study metrics, third-party test results, awards) prominently. Support marketing claims with evidence.<\/li> <\/ul> <h4><strong>11. User-Generated Content Presence<\/strong><\/h4> <ul> <li><strong>Definition:<\/strong>\u00a0Brand representation in community-driven platforms (Reddit, YouTube, forums) where users share experiences and opinions.<\/li> <li><strong>Why it matters:<\/strong>\u00a0Users validate synthetic AI answers against human experience. When\u00a0AI Overviews\u00a0appear, clicks on Reddit and YouTube grow from 18% to 30% because users seek social proof.<\/li> <li><strong>Goal:<\/strong>\u00a0Build positive presence in category-relevant subreddits, YouTube, and forums. YouTube and Reddit are\u00a0consistently\u00a0in the top 3 most cited domains\u00a0across LLMs.<\/li> <\/ul> <h2><strong>From Choice To Conviction<\/strong><\/h2> <p>Search is moving from abundance to synthesis. For two decades, Google\u2019s ranked list gave users a choice. AI search delivers a single answer that compresses multiple sources into one definitive response.<\/p> <p>The mechanics differ from early 2000s SEO:<\/p> <ul> <li>Retrieval windows replace crawl budgets.<\/li> <li>Selection rate replaces PageRank.<\/li> <li>Third-party validation replaces anchor text.<\/li> <\/ul> <p>The strategic imperative is identical: earn visibility in the interface where users search. Traditional SEO remains foundational, but AI visibility demands different content strategies:<\/p> <ul> <li>Conversational query coverage matters more than head-term rankings.<\/li> <li>External validation matters more than owned content.<\/li> <li>Structure matters more than keyword density.<\/li> <\/ul> <p>Brands that build systematic optimization programs now will compound advantages as LLM traffic scales. The shift from ranked lists to definitive answers is irreversible.<\/p> <hr\/> <p><em>Featured Image: Paulo Bobita\/Search Engine Journal<\/em><\/p> <\/div> <p>SEO#State #Search #Optimization #sejournal #Kevin_Indig1767923243<\/p> ","protected":false},"excerpt":{"rendered":"<p>Boost your skills with Growth Memo\u2019s weekly expert insights. Subscribe for free! Every year, after the winter holidays, I spend a few days ramping up by gathering the context from last year and reminding myself of where my clients are at. I want to use the opportunity to share my understanding of where we are [&hellip;]<\/p>\n","protected":false},"author":1,"featured_media":1131,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[16],"tags":[555,554,95,80,519],"class_list":["post-1130","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-accessibility","tag-kevin_indig","tag-optimization","tag-search","tag-sejournal","tag-state"],"acf":[],"_links":{"self":[{"href":"http:\/\/longzhuplatform.com\/index.php?rest_route=\/wp\/v2\/posts\/1130","targetHints":{"allow":["GET"]}}],"collection":[{"href":"http:\/\/longzhuplatform.com\/index.php?rest_route=\/wp\/v2\/posts"}],"about":[{"href":"http:\/\/longzhuplatform.com\/index.php?rest_route=\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"http:\/\/longzhuplatform.com\/index.php?rest_route=\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"http:\/\/longzhuplatform.com\/index.php?rest_route=%2Fwp%2Fv2%2Fcomments&post=1130"}],"version-history":[{"count":0,"href":"http:\/\/longzhuplatform.com\/index.php?rest_route=\/wp\/v2\/posts\/1130\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"http:\/\/longzhuplatform.com\/index.php?rest_route=\/wp\/v2\/media\/1131"}],"wp:attachment":[{"href":"http:\/\/longzhuplatform.com\/index.php?rest_route=%2Fwp%2Fv2%2Fmedia&parent=1130"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"http:\/\/longzhuplatform.com\/index.php?rest_route=%2Fwp%2Fv2%2Fcategories&post=1130"},{"taxonomy":"post_tag","embeddable":true,"href":"http:\/\/longzhuplatform.com\/index.php?rest_route=%2Fwp%2Fv2%2Ftags&post=1130"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}