Why AI will get even better at conversation (but not at telling the truth)

TECH 2
Noam Shazeer (left) and Daniel De Freitas, the founders of Character.AI, a company that allows users of its website to converse with sophisticated chatbots that are trained to talk like humans, in Palo Alto, California, on January 6, 2023. (Photo: NYTimes)
On a recent afternoon, Jonas Thiel, a socioeconomics major at a college in northern Germany, spent more than an hour chatting online with some of the left-wing political philosophers he had been studying. These were not the actual philosophers but virtual re-creations, brought to conversation, if not quite life, by sophisticated chatbots on a website called Character.AI.اضافة اعلان

Thiel’s favorite was a bot that imitated Karl Kautsky, a Czech-Austrian socialist who died before World War II. When Thiel asked Kautsky’s digital avatar to provide some advice for modern-day socialists struggling to rebuild the workers’ movement in Germany, the Kautsky-bot suggested that they publish a newspaper. “They can use it not only as a means of spreading socialist propaganda, which is in short supply in Germany for the time being, but also to organize working-class people,” the bot said.

The Kautsky-bot went on to argue that the working classes would eventually “come to their senses” and embrace a modern-day Marxist revolution. “The proletariat is at a low point in their history right now,” it wrote. “They will eventually realize the flaws in capitalism, especially because of climate change.”
For the moment, these and other advanced chatbots are a source of entertainment. And they are quickly becoming a more powerful way of interacting with machines.
Over the course of several days, Thiel met with other virtual scholars, including G.A. Cohen and Adolph Reed Jr. But he could have picked almost anyone — living or dead, real or imagined. At Character.AI, which emerged this summer, users can chat with reasonable facsimiles of people as varied as Queen Elizabeth II, William Shakespeare, Billie Eilish, or Elon Musk (there are several versions). Anyone you want to invoke, or concoct, is available for conversation. The company and site, founded by Daniel De Freitas and Noam Shazeer, two former Google researchers, is among the many efforts to build a new kind of chatbot. These bots cannot chat exactly like a human, but they often seem to.

In late November, OpenAI, a San Francisco artificial intelligence lab, unveiled a bot called ChatGPT that left more than 1 million people feeling as if they were chatting with another human being. Similar technologies are under development at Google, Meta, and other tech giants. Some companies have been reluctant to share the technology with the wider public. Because these bots learn their skills from data posted to the internet by real people, they often generate untruths, hate speech, and language that is biased against women and people of color. If misused, they could become a more efficient way of running the kind of misinformation campaign that has become commonplace in recent years.

“Without any additional guardrails in place, they are just going to end up reflecting all the biases and toxic information that is already on the web,” said Margaret Mitchell, a former AI researcher at Microsoft and Google, where she helped start its Ethical AI team. She is now with AI startup Hugging Face.

But other companies, including Character.AI, are confident that the public will learn to accept the flaws of chatbots and develop a healthy distrust of what they say. Thiel found that the bots at Character.AI had both a talent for conversation and a knack for impersonating real-life people. “If you read what someone like Kautsky wrote in the 19th century, he does not use the same language we use today,” he said. “But the AI can somehow translate his ideas into ordinary modern English.”

For the moment, these and other advanced chatbots are a source of entertainment. And they are quickly becoming a more powerful way of interacting with machines. Experts are still debating whether the strengths of these technologies will outweigh their flaws and potential for harm, but they agree on one point: The believability of make-believe conversation will continue to improve.

The art of conversationIn 2015, De Freitas, then working as a software engineer at Microsoft, read a research paper published by scientists at Google Brain, the flagship artificial intelligence lab at Google. Detailing what it called “A Neural Conversational Model”, the paper showed how a machine could learn the art of conversation by analyzing dialogue transcripts from hundreds of movies.
“You could tell this bot could generalize,” he said. “What it said did not look like what was in a movie script.”
The paper described what AI researchers call a neural network, a mathematical system loosely modeled on the web of neurons in the brain. This same technology also translates between Spanish and English on services like Google Translate and identifies pedestrians and traffic signs for self-driving cars navigating city streets.

A neural network learns skills by pinpointing patterns in enormous amounts of digital data. By analyzing thousands of cat photos, for instance, it can learn to recognize a cat.

When De Freitas read the paper, he was not yet an AI researcher; he was a software engineer working on search engines. But what he really wanted was to take Google’s idea to its logical extreme.

“You could tell this bot could generalize,” he said. “What it said did not look like what was in a movie script.”

He moved to Google in 2017. Officially, he was an engineer for YouTube, the company’s video-sharing site. But for his “20 percent time” project — a Google tradition that lets employees explore new ideas alongside their daily obligations — he began building his own chatbot.

The idea was to train a neural network using a much larger collection of dialogue: reams of chat logs culled from social media services and other sites across the internet. The idea was simple, but it would require enormous amounts of computer processing power. Even a supercomputer would need weeks or even months to analyze all that data.

As a Google engineer, he held a few credits that allowed him to run experimental software across the company’s vast network of computer data centers. But these credits would grant only a small fraction of the computing power needed to train his chatbot. So he started borrowing credits from other engineers; as the system analyzed more data, its skills would improve by leaps and bounds.

Initially, he trained his chatbot using what is called an LSTM, for Long Short-Term Memory — a neural network designed in the 1990s specifically for natural language. But he soon switched to a new kind of neural network called a transformer, developed by a team of Google AI researchers that included Noam Shazeer.

Unlike an LSTM, which reads text one word at a time, a transformer can use multiple computer processors to analyze an entire document in a single step.

Google, OpenAI and other organizations were already using transformers to build what are called “large language models”, systems suited for a wide range of language tasks, from writing Twitter messages to answering questions. Still working on his own, De Freitas focused the idea on conversation, feeding his transformer as much dialogue as possible.

It was an exceedingly simple approach. But as De Freitas likes to say: “Simple solutions for incredible results.”

The result in this case was a chatbot that he called Meena. It was so effective that Google Brain hired De Freitas and turned his project into an official research effort. Meena became LaMDA, short for Language Model for Dialogue Applications.

The project spilled into the public consciousness early in the summer when another Google engineer, Blake Lemoine, told The Washington Post that LaMDA was sentient. This assertion was an exaggeration, to say the least. But the brouhaha showed how quickly chatbots were improving inside top labs like Google Brain and OpenAI.
“These systems are not designed for truth,” Shazeer said. “They are designed for plausible conversation.”
Google was reluctant to release the technology, worried that its knack for misinformation and other toxic language could damage the company brand. But by this time De Freitas and Shazeer had left Google, determined to get this kind of technology into the hands of as many people as possible through their new company, Character.AI.

“The technology is useful today — for fun, for emotional support, for generating ideas, for all kinds of creativity,” Shazeer said.

Designed for open-ended exchangesChatGPT, the bot released by OpenAI to much fanfare in late November, was designed to operate as a new kind of question-and-answer engine. It is pretty good in this role, but the user never knows when the chatbot will just make something up. It may tell you that the official currency of Switzerland is the euro (it is actually the Swiss franc) or that Mark Twain’s Celebrated Jumping Frog of Calaveras County could not only jump but talk. AI researchers call this generation of untruths “hallucination”.

In building Character.AI, De Freitas and Shazeer had a different objective: open-ended conversation. They believe that today’s chatbots are better suited to this kind of service, for now a means of entertainment, factual or not. As the site notes, “Everything Characters say is made up!”

“These systems are not designed for truth,” Shazeer said. “They are designed for plausible conversation.”

De Freitas, Shazeer, and their colleagues did not build one bot that imitates Musk, another that mimics Queen Elizabeth, and a third that parrots Shakespeare. They built a single system that can imitate all those people and others. It has learned from reams of dialogue, articles, books, and digital text describing people like Musk, the queen, and Shakespeare.

Sometimes, the chatbot gets things right. Sometimes, it does not. When Thiel chatted with an avatar meant to imitate Reed, the 20th-century American political thinker, it turned him into “some kind of militant Maoist, which is definitely not right”.

Like with Google, OpenAI, and other top labs, De Freitas, Shazeer, and their colleagues plan on training their system with ever-larger amounts of digital data. This training could take months and millions of dollars; it could also sharpen the skills of the artificial conversationalist.

Researchers say that the rapid improvement will last only so long. Richard Socher, former chief scientist in charge of AI at Salesforce who now runs a startup called You.com, believes these exponential improvements will begin to level off over the next few years, when language models reach a point when they have analyzed virtually all the text on the internet.

But Shazeer said the runway is much longer: “There are billions of people in the world generating text all the time. People will keep spending more and more money to train smarter and smarter systems. We are nowhere near the end of that trend.”


Read more Technology
Jordan News