{"id":3718,"date":"2026-02-18T07:23:09","date_gmt":"2026-02-17T23:23:09","guid":{"rendered":"http:\/\/longzhuplatform.com\/?p=3718"},"modified":"2026-02-18T07:23:09","modified_gmt":"2026-02-17T23:23:09","slug":"googles-jeff-dean-ai-search-relies-on-classic-ranking-and-retrieval","status":"publish","type":"post","link":"http:\/\/longzhuplatform.com\/?p=3718","title":{"rendered":"Google\u2019s Jeff Dean: AI Search relies on classic ranking and retrieval"},"content":{"rendered":"<p><\/p> <div id=\"articleContent\" itemprop=\"articlebody\"> <div class=\"bialty-container\"> <p>Jeff Dean says Google\u2019s AI Search still works like classic Search: narrow the web to relevant pages, rank them, then let a model generate the answer.<\/p> <p>In an interview on Latent Space: The AI Engineer Podcast, Google\u2019s chief AI scientist explained how Google\u2019s AI systems work and how much they rely on traditional search infrastructure.<\/p> <p><strong>The architecture: filter first, reason last<\/strong>. Visibility still depends on clearing ranking thresholds. Content must enter the broad candidate pool, then survive deeper reranking before it can be used in an AI-generated response. Put simply, AI doesn\u2019t replace ranking. It sits on top of it.<\/p> <p>Dean said an LLM-powered system doesn\u2019t read the entire web at once. It starts with Google\u2019s full index, then uses lightweight methods to identify a large candidate pool \u2014 tens of thousands of documents. Dean said:<\/p> <ul class=\"wp-block-list\"> <li>\u201cYou identify a subset of them that are relevant with very lightweight kinds of methods. You\u2019re down to like 30,000 documents or something. And then you gradually refine that to apply more and more sophisticated algorithms and more and more sophisticated sort of signals of various kinds in order to get down to ultimately what you show, which is the final 10 results or 10 results plus other kinds of information.\u201d<\/li> <\/ul> <p>Stronger ranking systems narrow that set further. Only after multiple filtering rounds does the most capable model analyze a much smaller group of documents and generate an answer. Dean said:<\/p> <ul class=\"wp-block-list\"> <li>\u201cAnd I think an LLM-based system is not going to be that dissimilar, right? You\u2019re going to attend to trillions of tokens, but you\u2019re going to want to identify what are the 30,000-ish documents that are with the maybe 30 million interesting tokens. And then how do you go from that into what are the 117 documents I really should be paying attention to in order to carry out the tasks that the user has asked me to do?\u201d<\/li> <\/ul> <p>Dean called this the \u201cillusion\u201d of attending to trillions of tokens. In practice, it\u2019s a staged pipeline: retrieve, rerank, synthesize. Dean said:<\/p> <ul class=\"wp-block-list\"> <li>\u201cGoogle search gives you \u2026 not the illusion, but you are searching the internet, but you\u2019re finding a very small subset of things that are relevant.\u201d<\/li> <\/ul> <p><strong>Matching: from keywords to meaning<\/strong>. Nothing new here, but we heard another reminder that covering a topic clearly and comprehensively matters more than repeating exact-match phrases.<\/p> <p>Dean explained how LLM-based representations changed how Google matches queries to content.<\/p> <p>Older systems relied more on exact word overlap. With LLM representations, Google can move beyond the idea that particular words must appear on the page and instead evaluate whether a page \u2014 or even a paragraph \u2014 is topically relevant to a query. Dean said:<\/p> <ul class=\"wp-block-list\"> <li>\u201cGoing to an LLM-based representation of text and words and so on enables you to get out of the explicit hard notion of particular words having to be on the page. But really getting at the notion of this topic of this page or this page paragraph is highly relevant to this query.\u201d<\/li> <\/ul> <p>That shift lets Search connect queries to answers even when wording differs. Relevance increasingly centers on intent and subject matter, not just keyword presence.<\/p> <p><strong>Query expansion didn\u2019t start with AI<\/strong>. Dean pointed to 2001, when Google moved its index into memory across enough machines to make query expansion cheap and fast. Dean said:<\/p> <ul class=\"wp-block-list\"> <li>\u201cOne of the things that really happened in 2001 was we were sort of working to scale the system in multiple dimensions. So one is we wanted to make our index bigger, so we could retrieve from a larger index, which always helps your quality in general. Because if you don\u2019t have the page in your index, you\u2019re going to not do well. <\/li> <li>\u201cAnd then we also needed to scale our capacity because we were, our traffic was growing quite extensively. So we had a sharded system where you have more and more shards as the index grows, you have like 30 shards. Then if you want to double the index size, you make 60 shards so that you can bound the latency by which you respond for any particular user query. And then as traffic grows, you add more and more replicas of each of those.<\/li> <li>And so we eventually did the math that realized that in a data center where we had say 60 shards and 20 copies of each shard, we now had 1,200 machines with disks. And we did the math and we\u2019re like, Hey, one copy of that index would actually fit in memory across 1,200 machines. So in 2001, we \u2026 put our entire index in memory and what that enabled from a quality perspective was amazing.<\/li> <\/ul> <p>Before that, adding terms was expensive because it required disk access. Once the index lived in memory, Google could expand a short query into dozens of related terms \u2014 adding synonyms and variations to better capture meaning. Dean said:<\/p> <ul class=\"wp-block-list\"> <li>\u201cBefore, you had to be really careful about how many different terms you looked at for a query, because every one of them would involve a disk seek.<\/li> <li>\u201cOnce you have the whole index in memory, it\u2019s totally fine to have 50 terms you throw into the query from the user\u2019s original three- or four-word query. Because now you can add synonyms like restaurant and restaurants and cafe and bistro and all these things.<\/li> <li>\u201cAnd you can suddenly start \u2026 getting at the meaning of the word as opposed to the exact semantic form the user typed in. And that was \u2026 2001, very much pre-LLM, but really it was about softening the strict definition of what the user typed in order to get at the meaning.\u201d<\/li> <\/ul> <p>That change pushed Search toward intent and semantic matching years before LLMs. AI Mode (and its other AI experiences) continues Google\u2019s ongoing shift toward meaning-based retrieval, enabled by better systems and more compute.<\/p> <p><strong>Freshness as a core advantage<\/strong>. Dean said one of Search\u2019s biggest transformations was update speed. Early systems refreshed pages as rarely as once a month. Over time, Google built infrastructure that can update pages in under a minute. Dean said:<\/p> <ul class=\"wp-block-list\"> <li>\u201cIn the early days of Google, we were growing the index quite extensively. We were growing the update rate of the index. So the update rate actually is the parameter that changed the most.\u201d<\/li> <\/ul> <p>That improved results for news queries and affected the main search experience. Users expect current information, and the system is designed to deliver it. Dean said:<\/p> <ul class=\"wp-block-list\"> <li>\u201cIf you\u2019ve got last month\u2019s news index, it\u2019s not actually that useful.\u201d<\/li> <\/ul> <p>Google uses systems to decide how often to crawl a page, balancing how likely it is to change with how valuable the latest version is. Even pages that change infrequently may be crawled often if they\u2019re important enough. Dean said:<\/p> <ul class=\"wp-block-list\"> <li>\u201cThere\u2019s a whole \u2026 system behind the scenes that\u2019s trying to decide update rates and importance of the pages. So, even if the update rate seems low, you might still want to recrawl important pages quite often because the likelihood they change might be low, but the value of having updated is high.\u201d<\/li> <\/ul> <p><strong>Why we care<\/strong>. AI answers don\u2019t bypass ranking, crawl prioritization, or relevance signals. They depend on them. Eligibility, quality, and freshness still determine which pages are retrieved and narrowed. LLMs change how content is synthesized and presented \u2014 but the competition to enter the underlying candidate set remains a search problem.<\/p> <p><strong>The interview.<\/strong> Owning the AI Pareto Frontier \u2014 Jeff Dean<\/p> <figure class=\"wp-block-embed is-type-video is-provider-youtube wp-block-embed-youtube wp-embed-aspect-16-9 wp-has-aspect-ratio\"> <div class=\"wp-block-embed__wrapper\"> <noscript><iframe loading=\"lazy\" title=\"The AI Frontier: from Gemini 3 Deep Think distilling to Flash \u2014 Jeff Dean\" width=\"640\" height=\"360\" src=\"https:\/\/www.youtube.com\/embed\/F_1oDPWxpFQ?feature=oembed\" frameborder=\"0\" allow=\"accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share\" referrerpolicy=\"strict-origin-when-cross-origin\" allowfullscreen><\/iframe><\/noscript> <\/div> <\/figure> <\/div> <hr\/> <p class=\"article-disclosure\"> <em>Search Engine Land is owned by Semrush. We remain committed to providing high-quality coverage of marketing topics. Unless otherwise noted, this page\u2019s content was written by either an employee or a paid contractor of Semrush Inc.<\/em> <\/p> <hr\/> <div class=\"author-about py-4\"> <div class=\"card bg-light\"> <div class=\"row gx-0\"> <div class=\"col-12 col-lg-auto\"> <div class=\"authorImage p-2\"> <img loading=\"lazy\" decoding=\"async\" class=\"img-fluid rounded avatar-border\" src=\"data:image\/svg+xml,%3Csvg%20xmlns=\" http:=\"\" alt=\"Danny Goodwin\" width=\"140\" height=\"140\" data-lazy-src=\"https:\/\/searchengineland.com\/wp-content\/seloads\/2024\/07\/Danny-Goodwin-scaled.jpeg.webp\" title=\"Google\u2019s Jeff Dean: AI Search relies on classic ranking and retrieval\u63d2\u56fe\" \/><img loading=\"lazy\" decoding=\"async\" class=\"img-fluid rounded avatar-border\" src=\"https:\/\/searchengineland.com\/wp-content\/seloads\/2024\/07\/Danny-Goodwin-scaled.jpeg.webp\" alt=\"Danny Goodwin\" width=\"140\" height=\"140\" title=\"Google\u2019s Jeff Dean: AI Search relies on classic ranking and retrieval\u63d2\u56fe1\" \/> <\/div> <\/p><\/div> <div class=\"col-12 col-lg\"> <div class=\"card-body author-body p-2\"> <div id=\"authorBio-23878\" class=\"author-desc\"> <p>Danny Goodwin is Editorial Director of Search Engine Land &amp; Search Marketing Expo &#8211; SMX. He joined Search Engine Land in 2022 as Senior Editor. In addition to reporting on the latest search marketing news, he manages Search Engine Land\u2019s SME (Subject Matter Expert) program. He also helps program U.S. SMX events. <\/p> <p>Goodwin has been editing and writing about the latest developments and trends in search and digital marketing since 2007. He previously was Executive Editor of Search Engine Journal (from 2017 to 2022), managing editor of Momentology (from 2014-2016) and editor of Search Engine Watch (from 2007 to 2014). He has spoken at many major search conferences and virtual events, and has been sourced for his expertise by a wide range of publications and podcasts.<\/p> <\/p><\/div> <\/p><\/div> <\/p><\/div> <\/p><\/div> <\/p><\/div> <\/div> <p> <!-- START SIDEBAR LOWER SPACE --><\/p> <p><!-- END SIDEBAR LOWER SPACE --><\/p><\/div> <p>News#Googles #Jeff #Dean #Search #relies #classic #ranking #retrieval1771370589<\/p> ","protected":false},"excerpt":{"rendered":"<p>Jeff Dean says Google\u2019s AI Search still works like classic Search: narrow the web to relevant pages, rank them, then let a model generate the answer. In an interview on Latent Space: The AI Engineer Podcast, Google\u2019s chief AI scientist explained how Google\u2019s AI systems work and how much they rely on traditional search infrastructure. [&hellip;]<\/p>\n","protected":false},"author":1,"featured_media":3719,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[18],"tags":[6603,12611,179,7924,83,175,12612,88,95],"class_list":["post-3718","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-careers","tag-classic","tag-dean","tag-googles","tag-jeff","tag-news","tag-ranking","tag-relies","tag-retrieval","tag-search"],"acf":[],"_links":{"self":[{"href":"http:\/\/longzhuplatform.com\/index.php?rest_route=\/wp\/v2\/posts\/3718","targetHints":{"allow":["GET"]}}],"collection":[{"href":"http:\/\/longzhuplatform.com\/index.php?rest_route=\/wp\/v2\/posts"}],"about":[{"href":"http:\/\/longzhuplatform.com\/index.php?rest_route=\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"http:\/\/longzhuplatform.com\/index.php?rest_route=\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"http:\/\/longzhuplatform.com\/index.php?rest_route=%2Fwp%2Fv2%2Fcomments&post=3718"}],"version-history":[{"count":0,"href":"http:\/\/longzhuplatform.com\/index.php?rest_route=\/wp\/v2\/posts\/3718\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"http:\/\/longzhuplatform.com\/index.php?rest_route=\/wp\/v2\/media\/3719"}],"wp:attachment":[{"href":"http:\/\/longzhuplatform.com\/index.php?rest_route=%2Fwp%2Fv2%2Fmedia&parent=3718"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"http:\/\/longzhuplatform.com\/index.php?rest_route=%2Fwp%2Fv2%2Fcategories&post=3718"},{"taxonomy":"post_tag","embeddable":true,"href":"http:\/\/longzhuplatform.com\/index.php?rest_route=%2Fwp%2Fv2%2Ftags&post=3718"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}