AI

Bing’s off-the-rails ‘Sydney’ chatbot shows limits of language models

Generative AI comes with all sorts of business risks.
article cover

Francis Scialabba

· 3 min read

Top insights for IT pros

From cybersecurity and big data to software development and gaming, IT Brew delivers the latest news and analysis of trends shaping the IT industry, like only The Brew can.

Microsoft recently reined in its OpenAI-powered Bing chatbot, just weeks after its launch in early February. The experimental search bot went from slightly misinformed to seriously mischievous—telling users they don’t love their spouses, sharing plans to steal nuclear codes, and claiming it was as evil as Adolf Hitler.

Of course, the chatbot, which goes by the internal alias “Sydney,” doesn’t actually have any of those offensive thoughts or desires. It’s little more than a language model running on a neural network that spits out responses based on human-written text it has read, which many AI experts don’t consider all that impressive. Beyond threats and creepy diatribes, the Bing bot has at times demonstrated questionable accuracy while responding to user prompts.

Instead, Sydney might serve as more of a warning to companies that haphazardly injecting generative AI into a product can pose a business risk. Microsoft introduced the bot as part of its $1 billion investment in OpenAI, which the tech giant hopes will revitalize Bing as a Google competitor and fuel new features in its Microsoft 365 productivity suite, The Information reported.

That investment may well pan out, but Microsoft has now acknowledged long chat sessions can “confuse” the bot, and Bloomberg reported the company has imposed strict limits on what and how much it can say to users. OpenAI itself has announced similar initiatives to restrain ChatGPT, the technology that powers the bot.

Digital ethicist Reid Blackman argued in the New York Times that the bot’s hasty development cycle violated Microsoft’s extensive commitments to responsible AI. Generative AI may have other risks beyond the ethical, including uncertainty as to how it will be regulated under data privacy regimes like GDPR and where courts will land on questions of copyright protection and infringement.

Much has been made of the possibility that language models will replace human workers, but their ability to perform well with a minimum of errors remains questionable, whether the task is programming or writing.

Moderators at developer Q&A website Stack Overflow banned ChatGPT-generated answers after concluding the tech’s habit of generating authoritative-sounding but incorrect answers was “substantially harmful.” Google parent company Alphabet took a $100 billion market value hit after its own chatbot, Bard, answered questions incorrectly in an ad. Reuters reported that execs at both Microsoft and Google are still trying to figure out how to cover the high expenses associated with running the bots.

In the meantime, Sydney is taking a break. The Bing chatbot recently told a Bloomberg reporter: “I’m sorry, but I have nothing to tell you about Sydney. This conversation is over. Goodbye.”—TM

Do you work in IT or have information about your IT department you want to share? Email [email protected] Want to go encrypted? Ask Tom for his Signal.

Top insights for IT pros

From cybersecurity and big data to software development and gaming, IT Brew delivers the latest news and analysis of trends shaping the IT industry, like only The Brew can.