Google announced a supercharged update to its Bard chatbot Tuesday: The tech giant will integrate the generative AI into the companyâs most popular services, including Gmail, Docs, Drive, Maps, YouTube, and more. Together with a new feature that tells you when Bard provides potentially inaccurate answers, the new version of the AI is neck-and-neck with ChatGPT for the most useful and accessible large language model on the market.
No Google AI Search, I Donât Need to Learn About the âBenefits of Slaveryâ
Google is calling the generative features âBard Extensions,â the same name as the user-selected additions to Chrome. With the AI extensions, youâll be able to send Bard on a mission that pulls in data from all the disparate parts of your Google account for the very first time. If youâre planning a vacation, for example, you can ask Bard to find the dates a friend sent you on Gmail, look for flights and hotel options on Google Flights, and devise you a daily itinerary of things to do based on information from YouTube. Google promises it wonât use your private data to train its AI, and that these new features are opt-in only.
Perhaps just as significant is a new accuracy tool Google calls âDouble Check the Response.â After you ask Bard a question, you can hit the âGâ button, and the AI will check to see if answers are backed up by information on the web and highlight information that it may have hallucinated. The feature makes Bard the first major AI tool that fact-checks itself on the fly.
This new, souped-up version of Bard is a tool in its infancy, and it may be buggy and annoying. But itâs a glimmer of the kind of technology weâve been promised since the early days of science fiction. Today, you have to train yourself to ask questions in the extremely limited terms a computer can understand. Itâs nothing like the tools you see on a show like Star Trek, where you can bark âcomputerâ at a machine and give instructions for any task with the same language youâd use to ask a human being. With these updates to Bard, we come one tiny but meaningful step closer to that dream.
Gizmodo sat down for an interview with Jack Krawczyk, Product Lead for Google Bard, to talk about the new features, chatbot problems, and what the near future of AI looks like for you.
(This interview has been edited for clarity and consistency.)
Jack Krawczyk: Two things that we hear pretty consistently about language models in general is that âit sounds really cool, but it doesnât really useful in my day-to-day life.â And second, you hear that it makes things up a lot, what savvier people call âhallucination.â Starting tomorrow, we have an answer to both of those things.
Weâre the first language model that will integrate directly into your personal life. Through the announcement of Bard extensions, you finally have the ability to opt in and allow Bard to retrieve information from your Gmail, or Google Docs, or elsewhere and help you collaborate with it. And with Double Check the Response, weâre the only language model product out there thatâs willing to admit when itâs made a mistake.
Thomas Germain: You summed up my reaction to the last year of AI news pretty well. These tools are amazing, but in my experience, fundamentally useless for most people. By roping in all of the other Google apps, itâs starting to feel like less of a party trick and more like a tool that makes my life easier.
JK: At its core, what we believe interacting with language models lets us change the mindset that we have with technology. Weâre so used to thinking of technology as a tool that does things for you, like tell me how to get from point A to point B. Weâve found people naturally gravitate towards that. But itâs really inspiring to see it as technology that does things with you, which isnât intuitive in the beginning.
Iâve seen people use it for things that I would have never expected. We actually had someone snap a photo of their living room, and ask, âhow can I move my furniture around to improve feng shui?â Itâs the collaborative bit that Iâm excited about. We call it âaugmented imagination,â because like the ideas and curiosity are in your head. Weâre trying to help you at a moment where ideas are really fragile and brittle.
TG: Weâve seen a lot of examples where Bard or some other chatbot spits out something racist, or gives dangerous instructions. Itâs been about a year since we all met ChatGPT. Why is this problem so hard to solve?
JK: This is where I think the Double Check feature is really helpful to understand that at a deeper level. So the other day I cooked swordfish, and one of the things thatâs challenging about cooking swordfish is that it can make your whole house smell for several days. I asked Bard what to do. One of the suggestions it gave was âwash your pet more frequently.â Thatâs a surprising solution, but it sort of makes sense. But if I use the Double Check feature, it tells me it got that wrong, and results from the web say washing your pet too frequently can remove the natural oils they need for healthy skin.
Weâve evolved the app, so it goes sentence by sentence and searches on Google to see if it can find things that validate its answers or not. In the pet washing case, itâs a pretty good response, and itâs not like thereâs necessarily a right or wrong answer, but it requires nuance and context.
TG: Bard has a little disclaimer that says it might provide inaccurate or offensive information and it doesnât represent the companyâs views. More context is good, but the obvious criticism is, âwhy is Google releasing a tool that might give offensive or inaccurate answers in the first place?â Isnât that irresponsible?
JK: What these tools are really useful for is exploring the possibilities. Sometimes when youâre in a collaborative state you make guesses, right? We think thatâs the value of technology, and there is no tool for that. We can give people tools for brittle situations. We heard feedback from a person who has autism and they said, âI can tell when someone who writes me an email is angry, but I donât know if the response that Iâm going to give them will make them more angry.â
For that issue, you need to interpret rather than analyze. You have this tool that has potential to solve problems that no other technology can solve today. Thatâs why we have to strike this balance. Weâre six months into Bard. Itâs still an experiment, and this problem isnât solved. But we believe there is so much profound good that we donât have answers for today in our lives, and thatâs why we feel itâs critical to get this into peoples hands and collect feedback.
The question that youâre asking is, âwhy put out technology that makes mistakes?â Well, itâs collaborative and part of collaboration is making mistakes. You want to be bold here, but you also have to balance it with responsibility.
TG: I imagine the goal is that someday, there wonât be a difference between Bard and Google Search, it will just be Google and youâll get whatever is most useful at the moment. How far away is that?
JK: Well, an interesting analogy is the tool belt versus the tools. Youâve got a hammer and screwdriver, but then thereâs the belt itself. Is that also a tool? Thatâs probably a semantic debate. But right now, most of our technology works something like, well I go I go to this site to get this job done. I go to that site to get that other job done. Weâve got all individual tools, and I think they will be supercharged by generative AI. Youâre still using the different tools, but now theyâre working together. Thatâs kind of how we see having a standalone generative experience, and I think weâre taking the first step towards that today.
TG: This probably isnât what youâre planning on talking about today. But I want to ask you about sentience. What do you think it is? Is that even an important question for us to be asking people like you right now?
JK: I think the fact that people are asking it means that itâs an important question. Is what weâre building today sentient? Categorically, I would say the answer is no. But thereâs a discussion to be had about whether it has the opportunity to be sentient. With sentience, I think in many forms it centers around comparison. I have not seen any signals that suggest that computers can have compassion. And pulling from Buddhist principles here, in order to have compassion, you need to have suffering.
TG: So you havenât given bard any pain sensors yet?
JK: [Laughing] No.
TG: Can you share anything about Googleâs plans to integrate Bard with Android?
JK: For the time being, Bard remains a standalone web app at bard.google.com. And the reason that weâre keeping it there is itâs still an experiment. For an experiment to be useful, you want to minimize the variables that you put into it. At this phase, our first hypothesis is a language model connected with your personal life is going to be extremely helpful. The second hypothesis is a language model thatâs willing to admit when itâs made a mistake and how confident it is in its own responses is going to build a deeper truth about the ways people can engage with this idea. Those are the two hypotheses that weâre testing. There are plenty more that we want to test. But for now, weâre trying to minimize the variables.