{"id":1819,"date":"2026-01-20T22:11:10","date_gmt":"2026-01-20T14:11:10","guid":{"rendered":"http:\/\/longzhuplatform.com\/?p=1819"},"modified":"2026-01-20T22:11:10","modified_gmt":"2026-01-20T14:11:10","slug":"perplexity-ai-interview-explains-how-ai-search-works-via-sejournal-martinibuster","status":"publish","type":"post","link":"http:\/\/longzhuplatform.com\/?p=1819","title":{"rendered":"Perplexity AI Interview Explains How AI Search Works via @sejournal, @martinibuster"},"content":{"rendered":"<p><\/p> <div id=\"narrow-cont\"> <p>I recently spoke with Jesse Dwyer of Perplexity about SEO and AI search about what SEOs should be focusing on in terms of optimizing for AI search. His answers offered useful feedback about what publishers and SEOs should be focusing on right now.<\/p> <h2>AI Search Today<\/h2> <p>An important takeaway that Jesse shared is that personalization is completely changing<\/p> <blockquote> <p>\u201cI\u2019d have to say the biggest\/simplest thing to remember about AEO vs SEO is it\u2019s no longer a zero sum game. Two people with the same query can get a different answer on commercial search, if the AI tool they\u2019re using loads personal memory into the context window (Perplexity, ChatGPT).<\/p> <p>A lot of this comes down to the technology of the index (why there actually is a difference between GEO and AEO). But yes, it is currently accurate to say (most) traditional SEO best practices still apply.\u201d<\/p> <\/blockquote> <p>The takeaway from Dwyer\u2019s response is that search visibility is no longer about a single consistent search result. Personal context as a role in AI answers means that two users can receive significantly different answers to the same query with possibly different underlying content sources.<\/p> <p>While the underlying infrastructure is still a classic search index, SEO still plays a role in determining whether content is eligible to be retrieved at all. Perplexity AI is said to use a form of PageRank, which is a link-based method of determining the popularity and relevance of websites, so that provides a hint about some of what SEOs should be focusing on.<\/p> <p>However, as you\u2019ll see, what is retrieved is vastly different than in classic search.<\/p> <p><em>I followed up with the following question:<\/em><\/p> <p>So what you\u2019re saying (and correct me if I\u2019m wrong or slightly off) is that Classic Search tends to reliably show the same ten sites for a given query. But for AI search, because of the contextual nature of AI conversations, they\u2019re more likely to provide a different answer for each user.<\/p> <p><em>Jesse answered:<\/em><\/p> <blockquote> <p>\u201cThat\u2019s accurate yes.\u201d<\/p> <\/blockquote> <h2>Sub-document Processing: Why AI Search Is Different<\/h2> <p>Jesse continued his answer by talking about what goes on behind the scenes to generate an answer in AI search.<\/p> <p><em>He continued:<\/em><\/p> <blockquote> <p>\u201cAs for the index technology, the biggest difference in AI search right now comes down to whole-document vs. \u201csub-document\u201d processing.<\/p> <p>Traditional search engines index at the whole document level. They look at a webpage, score it, and file it.<\/p> <p>When you use an AI tool built on this architecture (like ChatGPT web search), it essentially performs a classic search, grabs the top 10\u201350 documents, then asks the LLM to generate a summary. That\u2019s why GPT search gets described as \u201c4 Bing searches in a trenchcoat\u201d \u2014the joke is directionally accurate, because the model is generating an output based on standard search results.<\/p> <p>This is why we call the optimization strategy for this GEO (Generative Engine Optimization). That whole-document search is essentially still algorithmic search, not AI, since the data in the index is all the normal page scoring we\u2019re used to in SEO. The AI-first approach is known as \u201csub-document processing.\u201d<\/p> <p>Instead of indexing whole pages, the engine indexes specific, granular snippets (not to be confused with what SEO\u2019s know as \u201cfeatured snippets\u201d). A snippet, in AI parlance, is about 5-7 tokens, or 2-4 words, except the text has been converted into numbers, (by the fundamental AI process known as a \u201ctransformer\u201d, which is the T in GPT). When you query a sub-document system, it doesn\u2019t retrieve 50 documents; it retrieves about 130,000 tokens of the most relevant snippets (about 26K snippets) to feed the AI.<\/p> <p>Those numbers aren\u2019t precise, though. The actual number of snippets always equals a total number of tokens that matches the full capacity of the specific LLM\u2019s context window. (Currently they average about 130K tokens). The goal is to completely fill the AI model\u2019s context window with the most relevant information, because when you saturate that window, you leave the model no room to \u2018hallucinate\u2019 or make things up.<\/p> <p>In other words, it stops being a creative generator and delivers a more accurate answer. This sub-document method is where the industry is moving, and why it is more accurate to be called AEO (Answer Engine Optimization).<\/p> <p>Obviously this description is a bit of an oversimplification. But the personal context that makes each search no longer a universal result for every user is because the LLM can take everything it knows about the searcher and use that to help fill out the full context window. Which is a lot more info than a Google user profile.<\/p> <p>The competitive differentiation of a company like Perplexity, or any other AI search company that moves to sub-document processing, takes place in the technology between the index and the 26K snippets. With techniques like modulating compute, query reformulation, and proprietary models that run across the index itself, we can get those snippets to be more relevant to the query, which is the biggest lever for getting a better, richer answer.<\/p> <p>Btw, this is less relevant to SEO\u2019s, but this whole concept is also why Perplexity\u2019s search API is so legit. For devs building search into any product, the difference is night and day.\u201d<\/p> <\/blockquote> <p><strong>Dwyer contrasts two fundamentally different indexing and retrieval approaches:<\/strong><\/p> <ul> <li>Whole-document indexing, where pages are retrieved and ranked as complete units.<\/li> <li>Sub-document indexing, where meaning is stored and retrieved as granular fragments.<\/li> <\/ul> <p>In the first version, AI sits on top of traditional search and summarizes ranked pages. In the second, the AI system retrieves fragments directly and never reasons over full documents at all.<\/p> <p>He also described that answer quality is constrained by context-window saturation, that accuracy emerges from filling the model\u2019s entire context window with relevant fragments. When retrieval succeeds at saturating that window, the model has little capacity to invent facts or hallucinate.<\/p> <p>Lastly, he says that \u201cmodulating compute, query reformulation, and proprietary models\u201d is part of their secret sauce for retrieving snippets that are highly relevant to the search query.<\/p> <p><em>Featured Image by Shutterstock\/Summit Art Creations<\/em><\/p> <\/div> <p>#Perplexity #Interview #Explains #Search #Works #sejournal #martinibuster1768918270<\/p> ","protected":false},"excerpt":{"rendered":"<p>I recently spoke with Jesse Dwyer of Perplexity about SEO and AI search about what SEOs should be focusing on in terms of optimizing for AI search. His answers offered useful feedback about what publishers and SEOs should be focusing on right now. AI Search Today An important takeaway that Jesse shared is that personalization [&hellip;]<\/p>\n","protected":false},"author":1,"featured_media":1820,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[18],"tags":[211,4274,415,4273,95,80,359],"class_list":["post-1819","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-careers","tag-explains","tag-interview","tag-martinibuster","tag-perplexity","tag-search","tag-sejournal","tag-works"],"acf":[],"_links":{"self":[{"href":"http:\/\/longzhuplatform.com\/index.php?rest_route=\/wp\/v2\/posts\/1819","targetHints":{"allow":["GET"]}}],"collection":[{"href":"http:\/\/longzhuplatform.com\/index.php?rest_route=\/wp\/v2\/posts"}],"about":[{"href":"http:\/\/longzhuplatform.com\/index.php?rest_route=\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"http:\/\/longzhuplatform.com\/index.php?rest_route=\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"http:\/\/longzhuplatform.com\/index.php?rest_route=%2Fwp%2Fv2%2Fcomments&post=1819"}],"version-history":[{"count":0,"href":"http:\/\/longzhuplatform.com\/index.php?rest_route=\/wp\/v2\/posts\/1819\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"http:\/\/longzhuplatform.com\/index.php?rest_route=\/wp\/v2\/media\/1820"}],"wp:attachment":[{"href":"http:\/\/longzhuplatform.com\/index.php?rest_route=%2Fwp%2Fv2%2Fmedia&parent=1819"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"http:\/\/longzhuplatform.com\/index.php?rest_route=%2Fwp%2Fv2%2Fcategories&post=1819"},{"taxonomy":"post_tag","embeddable":true,"href":"http:\/\/longzhuplatform.com\/index.php?rest_route=%2Fwp%2Fv2%2Ftags&post=1819"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}