How Search Generative Experience works and why retrieval … – Search Engine Land
Search, as we know it, has been irrevocably changed by generative AI.
The rapid improvements in Googles Search Generative Experience (SGE) and Sundar Pichais recent proclamations about its future suggest its here to stay.
The dramatic change in how information is considered and surfaced threatens how the search channel (both paid and organic) performs and all businesses that monetize their content. This is a discussion of the nature of that threat.
While writing The Science of SEO, Ive continued to dig deep into the technology behind search. The overlap between generative AI and modern information retrieval is a circle, not a Venn diagram.
The advancements in natural language processing (NLP) that started with improving search have given us Transformer-based large language models (LLMs).LLMs have allowed us to extrapolate content in response to queries based on data from search results.
Lets talk about how it all works and where the SEO skillset evolves to account for it.
Retrieval-augmented generation (RAG) is a paradigm wherein relevant documents or data points are collected based on a query or prompt and appended as a few-shot prompt to fine-tune the response from the language model.
Its a mechanism by which a language model can be grounded in facts or learn from existing content to produce a more relevant output with a lower likelihood of hallucination.
While the market thinks Microsoft introduced this innovation with the new Bing, the Facebook AI Research team first published the concept in May 2020 in the paper Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks, presented at the NeurIPS conference.However, Neeva was the first to implement this in a major public search engine by having it power its impressive and highly specific featured snippets.
This paradigm is game-changing because, although LLMs can memorize facts, they are information-locked based on their training data.For example, ChatGPTs information has historically been limited to a September 2021 information cutoff.
The RAG model allows new information to be considered to improve the output. This is what youre doing when using the Bing Search functionality or live crawling in a ChatGPT plugin like AIPRM.
This paradigm is also the best approach to using LLMs to generate stronger content output. I expect more will follow what were doing at my agency when they generate content for their clients as the knowledge of the approach becomes more commonplace.
Imagine that you are a student who is writing a research paper. You have already read many books and articles on your topic, so you have the context to broadly discuss the subject matter, but you still need to look up some specific information to support your arguments.
You can use RAG like a research assistant: you can give it a prompt, and it will retrieve the most relevant information from its knowledge base. You can then use this information to create more specific, stylistically accurate, and less bland output. LLMs allow computers to return broad responses based on probabilities. RAG allows that response to be more precise and cite its sources.
A RAG implementation consists of three components:
To make this less abstract, think about ChatGPTs Bing implementation. When you interact with that tool, it takes your prompt, performs searches to collect documents and appends the most relevant chunks to the prompt and executes it.
All three components are typically implemented using pre-trained Transformers, a type of neural network that has been shown to be very effective for natural language processing tasks. Again, Googles Transformer innovation powers the whole new world of NLP/U/G these days. Its difficult to think of anything in the space that doesnt have the Google Brain and Research teams fingerprints on it.
The Input Encoder and Output Generator are fine-tuned on a specific task, such as question answering or summarization. The Neural Retriever is typically not fine-tuned, but it can be pre-trained on a large corpus of text and code to improve its ability to retrieve relevant documents.
RAG is typically done using documents in a vector index or knowledge graphs. In many cases, knowledge graphs (KGs) are the more effective and efficient implementation because they limit the appended data to just the facts.
The overlap between KGs and LLMs shows a symbiotic relationship that unlocks the potential of both. With many of these tools using KGs, now is a good time to start thinking about leveraging knowledge graphs as more than a novelty or something that we just provide data to Google to build.
The benefits of RAG are pretty obvious; you get better output in an automated way by extending the knowledge available to the language model. What is perhaps less obvious is what can still go wrong and why. Lets dig in:
Retrieval is the make or break moment
Look, if the retrieval part of RAG isnt on point, were in trouble. Its like sending someone out to pick up a gourmet cheesesteak from Barclay Prime, and they come back with a veggie sandwich from Subway not what you asked for.
If its bringing back the wrong documents or skipping the gold, your outputs gonna be a bit well lackluster. Its still garbage in, garbage out.
Its all about that data
This paradigms got a bit of a dependency issue and its all about the data. If youre working with a dataset thats as outdated as MySpace or just not hitting the mark, youre capping the brilliance of what this system can do.
Echo chamber alert
Dive into those retrieved documents, and you might see some dj vu. If theres overlap, the models going to sound like that one friend who tells the same story at every party.
Youll get some redundancy in your results, and since SEO is driven by copycat content, you may get poorly researched content informing your results.
Prompt length limits
A prompt can only be so long, and while you can limit the size of the chunks, it may still be like trying to fit the stage for Beyonces latest world tour into a Mini-Cooper. To date, only Anthropics Claude supports a 100,000 token context window. GPT 3.5 Turbo tops out at 16,000 tokens.
Going off-script
Even with all your Herculean retrieval efforts, that doesnt mean that the LLM is going to stick to the script. It can still hallucinate and get things wrong.
I suspect these are some reasons why Google did not move on this technology sooner, but since they finally got in the game, lets talk about it.
Get the daily newsletter search marketers rely on.
Numerous articles will tell you what SGE is from a consumer perspective, including:
For this discussion, well talk about how SGE is one of Googles implementations of RAG; Bard is the other.
(Sidebar: Bards output has gotten a lot better since launch. You should probably give it another try.)
The SGE UX is still very much in flux. As I write this, Google has made shifts to collapse the experience with Show more buttons.
Lets zero in on the three aspects of SGE that will change search behavior significantly:
Historically, search queries are limited to 32 words. Because documents were considered based on intersecting posting lists for the 2 to 5-word phrases in those terms, and the expansion of those terms,
Google did not always understand the meaning of the query. Google has indicated that SGE is much better at understanding complex queries.
The AI snapshot is a more robust form of the featured snippet with generative text and links to citations. It often takes up the entirety of the above-the-fold content area.
The follow-up questions bring the concept of context windows in ChatGPT into search. As the user moves from their initial search to subsequent follow-up searches, the consideration set of pages narrows based on the contextual relevance created by the preceding results and queries.
All of this is a departure from the standard functionality of Search. As users get used to these new elements, there is likely to be a significant shift in behavior as Google focuses on lowering the Delphic costs of Search. After all, users always wanted answers, not 10 blue links.
The market believes that Google built SGE as a reaction to Bing in early 2023. However, the Google Research team presented an implementation of RAG in their paper, "Retrieval-Augmented Language Model Pre-Training (REALM)," published in August 2020.
The paper talks about a method of using the masked language model (MLM) approach popularized by BERT to do open-book question answering using a corpus of documents with a language model.
REALM identifies full documents, finds the most relevant passages in each, and returns the single most relevant one for information extraction.
During pre-training, REALM is trained to predict masked tokens in a sentence, but it is also trained to retrieve relevant documents from a corpus and attend to these documents when making predictions. This allows REALM to learn to generate more factually accurate and informative text than traditional language models.
Googles DeepMind team then took the idea further with Retrieval-Enhanced Transformer (RETRO). RETRO is a language model that is similar to REALM, but it uses a different attention mechanism.
RETRO attends to the retrieved documents in a more hierarchical way, which allows it to better understand the context of the documents. This results in text that is more fluent and coherent than text generated by REALM.
Following RETRO, The teams developed an approach called Retrofit Attribution using Research and Revision (RARR) to help validate and implement the output of an LLM and cite sources.
RARR is a different approach to language modeling. RARR does not generate text from scratch. Instead, it retrieves a set of candidate passages from a corpus and then reranks them to select the best passage for the given task. This approach allows RARR to generate more accurate and informative text than traditional language models, but it can be more computationally expensive.
These three implementations for RAG all have different strengths and weaknesses. While whats in production is likely some combination of innovations represented in these papers and more, the idea remains that documents and knowledge graphs are searched and used with a language model to generate a response.
Based on the publicly shared information, we know that SGE uses a combination of the PaLM 2 and MuM language models with aspects of Google Search as its retriever. The implication is that Googles document index and Knowledge Vault can both be used to fine-tune the responses.
Bing got there first, but with Googles strength in Search, there is no organization as qualified to use this paradigm to surface and personalize information.
Googles mission is to organize the worlds information and make it accessible. In the long term, perhaps well look back at the 10 blue links the same way we remember MiniDiscs and two-way pagers. Search, as we know it, is likely just an intermediate step until we arrive at something much better.
ChatGPTs recent launch of multimodal features is the "Star Trek" computer that Google engineers have often indicated they want to be. Searchers have always wanted answers, not the cognitive load of reviewing and parsing through a list of options.
A recent opinion paper titled Situating Search challenges the belief, stating that users prefer to do their research and validate, and search engines have charged ahead.
So, heres what is likely to happen as a result.
As users move away from queries composed of newspeak, their queries will get longer.
As users realize that Google has a better handle on natural language, it will change how they phrase their searches. Head terms will shrink while chunky middle and long-tail queries will grow.
The 10 blue links will get fewer clicks because the AI snapshot will push the standard organic results down. The 30-45% click-through rate (CTR) for Position 1 will likely drop precipitously.
However, we currently dont have true data to indicate how the distribution will change. So, the chart below is only for illustrative purposes.
Rank tracking tools have had to render the SERPs for various features for some time. Now, these tools will need to wait more time per query.
Most SaaS products are built on platforms like Amazon Web Service (AWS), Google Cloud Platform (GCP) and Microsoft Azure, which charge for compute costs based on the time used.
While rendered results may have come back in 1-2 seconds, now it may need to wait much longer, thereby causing the costs for rank tracking to increase.
Follow-up questions will give users Choose Your Own Adventure-style search journeys. As the context window narrows, a series of hyper-relevant content will populate the journey where each individual would have otherwise yielded more vague results.
Effectively, searches become multidimensional, and the onus is on content creators to make their content fulfill multiple stages to remain in the consideration set.
In the example above, Geico would want to have content that overlaps with these branches so they remain in the context window as the user progresses through their journey.
We dont have data on how user behavior has changed in the SGE environment. If you do, please reach out (looking at you, SimilarWeb).
What we do have is some historical understanding of user behavior in search.
We know that users take an average of 14.66 seconds to choose a search result. This tells us that a user will not wait for an automatically triggered AI snapshot with a generation time of more than 14.6 seconds. Therefore, anything beyond that time range does not immediately threaten your organic search traffic because a user will just scroll down to the standard results rather than wait.
We also know that, historically, featured snippets have captured 35.1% of clicks when they are present in the SERPs.
These two data points can be used to inform a few assumptions to build a model of the threat of how much traffic could be lost from this rollout.
Lets first review the state of SGE based on available data.
Since theres no data on SGE, it would be great if someone created some. I happened to come across a dataset of roughly 91,000 queries and their SERPs within SGE.
For each of these queries, the dataset includes:
The queries are also segmented into different categories so we can get a sense of how different things perform. I dont have enough of your attention left to go through the entirety of the dataset, but here are some top-level findings.
AI snapshots now take an average of 6.08 seconds to generate
When SGE was first launched, and I started reviewing load times of the AI snapshot, it took 11 to 30 seconds for them to appear. Now I'm seeing a range of 1.8 to 17.2 seconds for load times. Automatically triggered AI snapshots load between 2.9 and 15.8 seconds.
As you can see from the chart, most load times are well below 14.6 seconds at this point. Its pretty clear that the 10 blue link traffic for the overwhelming majority of queries will be threatened.
The average varies a bit depending on the keyword category. With the Entertainment-Sports category having a much higher load time than all other categories, this may be a function of how heavy the source content for pages typically is for each given vertical.
Snapshot type distribution
While there are many variants of the experience, I have broadly segmented the snapshot types into Informational, Local, and Shopping page experiences. Within my 91,000 keyword set, the breakdown is 51.08% informational, 31.31% local, and 17.60% shopping.
60.34% of queries did not feature an AI snapshot
In parsing the page content, the dataset identifies two cases to verify whether there is a snapshot on the page. It looks for the autotriggered snapshot and the Generate button. Reviewing this data indicates that 39.66% of queries in the dataset have triggered AI snapshots.
The top 10 results are often used but not always
In the dataset Ive reviewed, Positions 1, 2, and 9 get cited the most in the AI snapshots carousel.
The AI snapshot most often uses six results out of the top 10 to build its response. However, 9.48% of the time, it does not use any of the top 10 results in the AI snapshot.
Based on my data, it rarely uses all the results from the top 10.
Highly relevant chunks often appear earlier in the carousel
Lets consider the AI snapshot for the query [bmw i8]. The query returns seven results in the carousel. Four of them are explicitly referenced in the citations.
Clicking on a result in the carousel often takes you to one of the fraggles (the term for passage ranking links that the brilliant Cindy Krum coined) that drop you on a specific sentence or paragraph.
The implication is that these are the paragraphs or sentences that inform the AI snapshot.
Naturally, our next step is to try to get a sense of how these results are ranked because they are not presented in the same order as the URLs cited in the copy.
I assume that this ranking is more about relevance than anything else.
To test this hypothesis, I vectorized the paragraphs using the Universal Sentence Encoder and compared them to the vectorized query to see if the descending order holds up.
Id expect the paragraph with the highest similarity score would be the first one in the carousel.
The results are not quite what I expected. Perhaps there may be some query expansion at play where the query Im comparing is not the same as what Google might be comparing.
More here:
How Search Generative Experience works and why retrieval ... - Search Engine Land
- 70+ PPC and Google Adwords Interview Questions and Answers for 2025 - Simplilearn - November 16th, 2024 [November 16th, 2024]
- Reframing SEO: Why training search engines is the new game in the age of AI - Search Engine Land - August 29th, 2024 [August 29th, 2024]
- Redefining SEO: How training search engines is shaping the future of digital content - Tech Edition - August 29th, 2024 [August 29th, 2024]
- SEO University Partners with Salterra to Launch Advanced Schema - WICZ - August 25th, 2024 [August 25th, 2024]
- SEO University Partners with Salterra to Launch Advanced Schema Course, Empowering SEO Professionals with Expert Training - Barchart - August 20th, 2024 [August 20th, 2024]
- SEO University Partners with Salterra to Launch Advanced Schema - openPR - August 20th, 2024 [August 20th, 2024]
- Top Websites to Learn SEO in 2024 - Analytics Insight - July 26th, 2024 [July 26th, 2024]
- What is the process to Learn SEO Step by Step? - INSCMagazine - January 30th, 2024 [January 30th, 2024]
- Park Seo-joon Mentions V's Photo At Army Training Center, He Wore The Same Raincoat As I Did 15 Years Go - KBIZoom - December 17th, 2023 [December 17th, 2023]
- The Bicycle Coalition Attends the Vision Zero Cities 2023 Conference - Bicycle Coalition of Greater Philadelphia - October 27th, 2023 [October 27th, 2023]
- The 40 best crime movies of all time - Entertainment Weekly News - October 27th, 2023 [October 27th, 2023]
- 50 Remote Jobs That Pay Over $50000 a Year: Part Two Jobs ... - Medium - October 23rd, 2023 [October 23rd, 2023]
- ONE: Radzuan responds to Stamp rematch talk, impressed by title win - South China Morning Post - October 23rd, 2023 [October 23rd, 2023]
- California Law Limits Bitcoin ATM Transactions to $1,000 to Thwart ... - Slashdot - October 23rd, 2023 [October 23rd, 2023]
- Tech CEO Sentenced To 5 Years in IP Address Scheme - Slashdot - October 23rd, 2023 [October 23rd, 2023]
- Is Digital Marketing Training Worth it - Kings of War - October 3rd, 2023 [October 3rd, 2023]
- The 2023 Nonprofit Power 100 - City & State - October 3rd, 2023 [October 3rd, 2023]
- 'Embarrassing' Court Document Google Wanted to Hide Finally ... - Slashdot - October 3rd, 2023 [October 3rd, 2023]
- H&R Block, Meta, and Google Slapped With RICO Suit, Allegedly ... - Slashdot - October 3rd, 2023 [October 3rd, 2023]
- FBI Indicts Goldman Sachs Analyst Who Tried Using Xbox Chat for ... - Slashdot - October 3rd, 2023 [October 3rd, 2023]
- 8 top marketing certifications and courses for 2023 - TechTarget - July 17th, 2023 [July 17th, 2023]
- How to win SEO allies and influence the brand guardians - Search Engine Land - July 17th, 2023 [July 17th, 2023]
- How relying on LLMs can lead to SEO disaster - Search Engine Land - July 17th, 2023 [July 17th, 2023]
- Become the next generation of multimedia content creators and ... - Education Times - July 17th, 2023 [July 17th, 2023]
- A Week in My Life: Fiona Brindle, Head of SEO, TrunkBBI - Prolific North - July 17th, 2023 [July 17th, 2023]
- Preparing the underserved: Five Auburn University alumni ... - Office of Communications and Marketing - July 17th, 2023 [July 17th, 2023]
- Should You Have a Go at Search Engine Optimization (SEO)? - Printing Impressions - June 9th, 2023 [June 9th, 2023]
- Chris Raulf of Boulder SEO Marketing to Give Masterclass on Micro ... - Digital Journal - June 9th, 2023 [June 9th, 2023]
- Augmented Reality Training Simulator Market 2031 Key Insights and ... - KaleidoScot - June 9th, 2023 [June 9th, 2023]
- Training Software Market 2023 Trends with Analysis on Key Players ... - KaleidoScot - June 9th, 2023 [June 9th, 2023]
- Training Outsourcing Market 2023 Trends with Analysis on Key ... - KaleidoScot - June 9th, 2023 [June 9th, 2023]
- COVID-19 Impact Analysis of Education Market 2031 | Key Players ... - KaleidoScot - June 9th, 2023 [June 9th, 2023]
- MarTechBot: Insights from real-world usage (so far) - MarTech - June 9th, 2023 [June 9th, 2023]
- Cognitive Assessment and Training Healthcare Market 2031 Growth ... - KaleidoScot - June 9th, 2023 [June 9th, 2023]
- Prestige whisky brand appoints Wild PR to support business growth - Bdaily News - June 9th, 2023 [June 9th, 2023]
- Erling Haaland Names Toughest Opponent He's Faced This Year ... - Sports Lens - June 9th, 2023 [June 9th, 2023]
- Local Brand Advisor Proves Its Worth As Leading and Results ... - Digital Journal - May 29th, 2023 [May 29th, 2023]
- Family: The Unbreakable Bond - K-drama Episode 10 Recap ... - TheReviewGeek - May 29th, 2023 [May 29th, 2023]
- Salesbop: The AI-Powered Sales Coach and Trainer ... - Digital Journal - May 29th, 2023 [May 29th, 2023]
- Career Technical Educational Opportunities for Students Attending ... - Demopolis Times - May 29th, 2023 [May 29th, 2023]
- Doctor Cha Episode 13 Twitter Reactions: Cliffhanger Over ... - Leisure Byte - May 29th, 2023 [May 29th, 2023]
- The National Eating Disorder Helpline Replaced Its Staff With a ... - The Mary Sue - May 29th, 2023 [May 29th, 2023]
- Brendan Johnston: A 15 year pro-racing quest with a gravel resolution - Cyclingnews - May 29th, 2023 [May 29th, 2023]
- Business Briefing: Apple Blossom Holistic, business news and ... - Laois Today - May 29th, 2023 [May 29th, 2023]
- How the media is covering ChatGPT - Columbia Journalism Review - May 29th, 2023 [May 29th, 2023]
- BSM to Host a Complimentary Webinar Entitled "AI and SEO. The ... - Digital Journal - May 18th, 2023 [May 18th, 2023]
- Developing Skills to Stay Competitive - ATD - May 18th, 2023 [May 18th, 2023]
- The biggest challenges facing small businesses and how to ... - Arizona Big Media - May 18th, 2023 [May 18th, 2023]
- Priyanka Chopra Jonas On Husband Nick Jonas' 'Mean' Martini, Her ... - ELLE UK - May 18th, 2023 [May 18th, 2023]
- The Idaho Towns Bankrolling Donald Trump's Campaign - News Radio 1310 KLIX - May 18th, 2023 [May 18th, 2023]
- Online Stable Startup: Tips and Tricks for Launching a Horse Business - Everything Horse UK - May 18th, 2023 [May 18th, 2023]
- ReKommendations: My Perfect Stranger, Duty After School, and more; K-dramas to catch up with this weekend - PINKVILLA - May 18th, 2023 [May 18th, 2023]
- The Full Cast of Netflix's 'Black Knight' - We Got This Covered - May 18th, 2023 [May 18th, 2023]
- Thanet business news: CAMRA awards, Thanet Earth, Dirtee Feast ... - The Isle of Thanet News - May 18th, 2023 [May 18th, 2023]
- Top 100: New to the List Fast Action Pest Control - PCT Magazine - May 18th, 2023 [May 18th, 2023]
- We are in content marketing era, the opportunities are diverse - Capital FM Kenya - May 14th, 2023 [May 14th, 2023]
- 25+ Best Remote Jobs Without Degree or Experience in 2023 - Southwest Journal - May 14th, 2023 [May 14th, 2023]
- SEO Fight Club Episode 198 Explores AI Training Corpus And AI ... - Digital Journal - May 12th, 2023 [May 12th, 2023]
- Various Advantages of HubSpot - CIOReview - May 12th, 2023 [May 12th, 2023]
- How to Start and Grow a Successful Real Estate Business: Business ... - RealtyBizNews - May 12th, 2023 [May 12th, 2023]
- Small Business, Big Results: Rely on Top SEO Company in Ahmedabad - The Week - May 12th, 2023 [May 12th, 2023]
- How to Get Google's Attention with AI-Generated Content - PR News - For Smart Communicators - May 12th, 2023 [May 12th, 2023]
- Meet the next Leadership Academy for Women in Media cohort in ... - Poynter - May 12th, 2023 [May 12th, 2023]
- Republic of Korea and U.S. Navy Conduct Combined Maritime ... - Pacific Command - May 10th, 2023 [May 10th, 2023]
- Boostly introduces ChatGPT integration for direct booking websites - Short Term Rentalz - May 10th, 2023 [May 10th, 2023]
- YACSS Announces Panel of Speakers for the First Annual YACSS SEO Conference - Yahoo Finance - May 10th, 2023 [May 10th, 2023]
- Google On Protecting Anchor Text Signal From Spam Site Influence - Search Engine Journal - May 10th, 2023 [May 10th, 2023]
- How To Start A Business In 11 Steps (2023 Guide) - Forbes - May 10th, 2023 [May 10th, 2023]
- Ocean Tomo, a part of J.S. Held Welcomes Delegation from Korea ... - PR Web - May 10th, 2023 [May 10th, 2023]
- Lionel Messi Returns To PSG Training After Suspension Lifted - Sports Lens - May 10th, 2023 [May 10th, 2023]
- Engaging Consumers in a Generative AI World - BCG - May 10th, 2023 [May 10th, 2023]
- Alyse Anderson has been training with Rose Namajunas - Asian MMA - May 8th, 2023 [May 8th, 2023]
- 12 questions to ask SEO platform vendors during the demo - MarTech - May 8th, 2023 [May 8th, 2023]
- How To Write ChatGPT Prompts To Get The Best Results - Search Engine Journal - May 8th, 2023 [May 8th, 2023]
- Roses and thorns: 5-6-23 - The Commercial Dispatch - May 8th, 2023 [May 8th, 2023]
- Achieving success in your own terms through the eyes of six Filipino ... - Manila Bulletin - May 8th, 2023 [May 8th, 2023]
- Rethinking SEO Strategy: Mindset Coach Helps Businesses Achieve ... - BusinessMole - May 2nd, 2023 [May 2nd, 2023]
- Achieving SEO Success: Mindset Coach Offers Innovative Problem ... - Business Manchester - May 2nd, 2023 [May 2nd, 2023]
- Megan Bridgeman named SEO editor based on the West Coast - The Washington Post - May 2nd, 2023 [May 2nd, 2023]
- SEO agency Megantic promotes one and hires another - AdNews - May 2nd, 2023 [May 2nd, 2023]