小程序
传感搜
传感圈

AI Chatbots Are Coming to Search Engines. Can You Trust Them?

2023-02-19
关注

Months after the chatbot ChatGPT wowed the world with its uncanny ability to write essays and answer questions like a human, artificial intelligence (AI) is coming to Internet search.

Three of the world’s biggest search engines — Google, Bing and Baidu — last week said they will be integrating ChatGPT or similar technology into their search products, allowing people to get direct answers or engage in a conversation, rather than merely receiving a list of links after typing in a word or question. How will this change the way people relate to search engines? Are there risks to this form of human–machine interaction?

Microsoft’s Bing uses the same technology as ChatGPT, which was developed by OpenAI of San Francisco, California. But all three companies are using large language models (LLMs). LLMs create convincing sentences by echoing the statistical patterns of text they encounter in a large database. Google’s AI-powered search engine, Bard, announced on 6 February, is currently in use by a small group of testers. Microsoft’s version is widely available now, although there is a waiting list for unfettered access. Baidu’s ERNIE Bot will be available in March.

Before these announcements, a few smaller companies had already released AI-powered search engines. “Search engines are evolving into this new state, where you can actually start talking to them, and converse with them like you would talk to a friend,” says Aravind Srinivas, a computer scientist in San Francisco who last August co-founded Perplexity — an LLM-based search engine that provides answers in conversational English.

Changing trust

The intensely personal nature of a conversation — compared with a classic Internet search — might help to sway perceptions of search results. People might inherently trust the answers from a chatbot that engages in conversation more than those from a detached search engine, says Aleksandra Urman, a computational social scientist at the University of Zurich in Switzerland.

A 2022 study1 by a team based at the University of Florida in Gainesville found that for participants interacting with chatbots used by companies such as Amazon and Best Buy, the more they perceived the conversation to be human-like, the more they trusted the organization.

That could be beneficial, making searching faster and smoother. But an enhanced sense of trust could be problematic given that AI chatbots make mistakes. Google’s Bard flubbed a question about the James Webb Space Telescope in its own tech demo, confidently answering incorrectly. And ChatGPT has a tendency to create fictional answers to questions to which it doesn’t know the answer — known by those in the field as hallucinating.

A Google spokesperson said Bard’s error “highlights the importance of a rigorous testing process, something that we’re kicking off this week with our trusted-tester programme”. But some speculate that, rather than increasing trust, such errors, assuming they are discovered, could cause users to lose confidence in chat-based search. “Early perception can have a very large impact,” says Sridhar Ramaswamy, a computer scientists based in Mountain View, California and chief executive of Neeva, an LLM-powered search engine launched in January. The mistake wiped $100 billion from Google’s value as investors worried about the future and sold stock.

Lack of transparency

Compounding the problem of inaccuracy is a comparative lack of transparency. Typically, search engines present users with their sources — a list of links — and leave them to decide what they trust. By contrast, it’s rarely known what data an LLM trained on — is it Encyclopaedia Britannica or a gossip blog?

“It’s completely untransparent how [AI-powered search] is going to work, which might have major implications if the language model misfires, hallucinates or spreads misinformation,” says Urman.

If search bots make enough errors, then, rather than increasing trust with their conversational ability, they have the potential to unseat users’ perceptions of search engines as impartial arbiters of truth, Urman says.

She has conducted as-yet unpublished research that suggests current trust is high. She examined how people perceive existing features that Google uses to enhance the search experience, known as ‘featured snippets’, in which an extract from a page that is deemed particularly relevant to the search appears above the link, and ‘knowledge panels’ — summaries that Google automatically generates in response to searches about, for example, a person or organization. Almost 80% of people Urman surveyed deemed these features accurate, and around 70% thought they were objective.

Chatbot-powered search blurs the distinction between machines and humans, says Giada Pistilli, principal ethicist at Hugging Face, a data-science platform in Paris that promotes the responsible use of AI. She worries about how quickly companies are adopting AI advances: “We always have these new technologies thrown at us without any control or an educational framework to know how to use them.”

This article is reproduced with permission and was first published on February 13 2023.

参考译文
人工智能聊天机器人即将进入搜索引擎。你能相信他们吗?
几个月前,聊天机器人ChatGPT以其像人类一样写文章和回答问题的神奇能力震惊了世界,如今,人工智能(AI)即将进入互联网搜索领域。全球三大搜索引擎谷歌、必应和百度上周表示,他们将把ChatGPT或类似技术集成到他们的搜索产品中,让人们可以直接得到答案或参与对话,而不仅仅是在输入一个单词或问题后收到一串链接。这将如何改变人们与搜索引擎的关系?这种形式的人机交互有风险吗?微软的必应使用了与ChatGPT相同的技术,后者是由加州旧金山的OpenAI公司开发的。但这三家公司都在使用大型语言模型(LLMs)。llm通过在大型数据库中遇到的文本的统计模式来创建令人信服的句子。谷歌的人工智能搜索引擎Bard于2月6日宣布,目前有一小群测试人员正在使用。微软的版本现在已经广泛使用,不过要想无限制地使用还需要排队等候。百度的ERNIE Bot将于3月上市。在这些声明发布之前,一些小公司已经发布了人工智能搜索引擎。旧金山的计算机科学家Aravind Srinivas说:“搜索引擎正在演变成一种新的状态,在这种状态下,你实际上可以开始与它们交谈,并像与朋友交谈一样与它们交谈。”他在去年8月联合创立了Perplexity——一个基于llm的搜索引擎,以会话式英语提供答案。与经典的网络搜索相比,对话中强烈的个人性质可能有助于改变人们对搜索结果的看法。瑞士苏黎世大学(University of Zurich)的计算社会科学家亚历山德拉·乌尔曼(Aleksandra Urman)表示,人们可能天生就更相信参与对话的聊天机器人的答案,而不是来自独立搜索引擎的答案。位于盖恩斯维尔的佛罗里达大学的一个团队在2022年进行的一项研究发现,对于与亚马逊和百思买等公司使用的聊天机器人互动的参与者来说,他们越觉得对话与人类类似,就越信任该组织。这可能是有益的,可以使搜索更快、更流畅。但考虑到人工智能聊天机器人会犯错,信任感的增强可能会有问题。谷歌的巴德在自己的技术演示中搞砸了一个关于詹姆斯·韦伯太空望远镜的问题,自信地回答错误。ChatGPT倾向于为它不知道答案的问题创造虚构的答案——被该领域的人称为幻觉。谷歌的一位发言人表示,巴德的错误“凸显了严格测试过程的重要性,我们本周将启动可信测试程序”。但一些人推测,如果这些错误被发现,可能会导致用户对基于聊天的搜索失去信心,而不是增加信任。加州山景城(Mountain View)的计算机科学家斯里达尔•拉马斯瓦米(Sridhar Ramaswamy)表示:“早期的感知可能会产生非常大的影响。”他同时也是今年1月推出的Neeva搜索引擎的首席执行官。这一错误使谷歌的市值蒸发了1000亿美元,因为投资者对未来感到担忧,纷纷抛售股票。使不准确问题更加复杂的是透明度的相对缺乏。通常,搜索引擎会向用户提供他们的来源——一列链接——并让他们自己决定信任什么。相比之下,很少有人知道LLM训练的数据是什么——是大英百科全书还是八卦博客?乌尔曼说:“(人工智能搜索)将如何工作是完全不透明的,如果语言模型失灵、产生幻觉或传播错误信息,这可能会产生重大影响。” 乌尔曼说,如果搜索机器人犯了足够多的错误,那么,它们的对话能力不仅不会增加用户的信任,而且有可能颠覆用户对搜索引擎作为公正的真相仲裁者的认知。她进行了一项尚未发表的研究,表明目前的信任度很高。她研究了人们如何感知谷歌用来增强搜索体验的现有功能,即所谓的“特色片段”,即从页面中摘录的被认为与搜索特别相关的内容出现在链接上方,以及“知识面板”——谷歌在搜索时自动生成的摘要,例如关于一个人或一个组织的搜索。在Urman的调查中,近80%的人认为这些特征是准确的,约70%的人认为它们是客观的。巴黎数据科学平台hug Face的首席伦理学家吉亚达•皮斯蒂利(Giada Pistilli)表示,聊天机器人驱动的搜索模糊了机器和人类之间的区别。该平台旨在促进负责任地使用人工智能。她担心企业采用人工智能技术的速度有多快:“我们总是在没有任何控制或教育框架的情况下,把这些新技术扔给我们。”本文经许可转载,首次发表于2023年2月13日。
您觉得本篇内容如何
评分

评论

您需要登录才可以回复|注册

提交评论

提取码
复制提取码
点击跳转至百度网盘