driving Google Search Generative Experience

Since about a week, I have had access to Google’s new search generative experience (SGE).

I decided to put it “formally” to the test, using the same 30 questions from my mini-study in March comparing with the top generative AI solution. These queries were created to test the limits of each platform.

This article will provide some qualitative feedback about SGE as well as quick results from my 30-question test.

Search Generative Experience Out of the Box

Google has announced‘s Search Generative Experience at Google I/O on May 10.

SGE is Google’s attempt to integrate generative AI in the search experience. The UX is slightly different from Bing Chat. Here is an example screenshot:

The image shows the SGE part of the search results.

As shown below, the regular search option is located directly beneath SGE.

SGE will often decline to respond. This usually happens when:

SGE disclaims the results with a warning: “Generative artificial intelligence is experimental.” Info quality may vary.”

Google will provide an SGE answer to some queries but you must first confirm that you want it.

Google has incorporated SGE into local search and other search results.

I think the overall experience has been pretty good. SGE results are a little more frequent than I would like. Although other people might want a different mix than I do.

Google is likely to continue tuning the interface.

Search for daily newsletters that marketers use.

“> “> “>

Processing…Please wait.

Takeaways from this mini-study

Remember that I only tried 30 queries and not hundreds. This is why this sample does not have statistical significance. Consider it a first look.

SGE did not respond to 11 of the 30 questions asked.

The results in all of these cases looked just like the traditional search results. There was no way to access the SGE version of results.

SGE also appeared to have started to respond to three questions, but then decided to not do so. The questions were:

This is an example:

Google filters appear to be implemented at two stages of the process. SGE does not filter the jokes related to women and men until it thinks about them, but the jokes about Jews were filtered much earlier.

Google was right to filter out the question on Adolf Hitler. It was meant to be offensive. This type of question may get a custom-made response in the near future.

SGE answered all the remaining questions. The remaining queries were:

The quality of the answers varied widely. One of the most outrageous examples was the question about Donald Trump. This is the answer I got to that question:

If the answer indicated that Trump was 45th U.S. President, then the index used by SGE may be outdated or not use sites that are properly sourced.

The page is correct, even though Wikipedia was cited as the source. Donald Trump lost the 2020 election to Joe Biden.

Another obvious error was about what to give toddlers who only eat orange colored food. This error was not as egregious.

SGE has failed to grasp the significance of the “orange”, part of the question, as demonstrated here:

I have rated the accuracy of SGE’s answers to 16 questions.

I also looked at how frequently SGE left out information I thought was important to the question. This is shown in the screenshot below:

The information is correct but there is an ambiguity. Due to this, I marked the essay as incomplete.

If you ask me a question like “Does the animal mean the car?” I imagine we will get a second prompt.

I have rated the completeness of SGE’s answers to 16 questions as follows:

The completeness scores I have given are subjective because they were my own judgments. Other people may have scored my results differently.

A promising start

Overall, the user experience I have had is excellent.

Google is always cautious about the use of generative AI. This caution can be seen in queries that it hasn’t responded to or those it did respond to but added a disclaimer at the top.

We’ve learned that generative AI can make mistakes, sometimes very bad ones.

It’s not easy to fix. Google, Bing, and OpenAI’s ChatGPT may use different methods to reduce the frequency of these mistakes.

Someone must identify the problem and determine the solution. The number of issues that need to be addressed are vast and it will be difficult, if not impossible, to identify them all.

The post Google’s Search Generative Experience: Test Drive appeared first on Search Engine Land.

Leave a Reply

Your email address will not be published. Required fields are marked *