Jun 22, 2022 • 1HR 5M

Nothing More Than a Magic Trick (w/Gary Marcus)

On Dead Cat, we discuss Google's artificial intelligence, LaMDA, with a critic of the current approach to generalized artificial intelligence

Open in playerListen on);

Appears in this episode

Eric Newcomer
Ahead of the 2nd Cerebral Valley AI Summit on Nov. 15, hosts Eric Newcomer, Max Child, and James Wilsterman bring you a six-part podcast series on the rise of generative artificial intelligence.
Episode details
Source: Gary Marcus’s book with Ernest Davis, Rebooting AI

Are we nearing a time when we are going to get to have real, meaningful conversations with artificial intelligence?

Nitasha Tiku got the world wondering just that with her story in the Washington Post about a Google engineer who believes that the company’s LaMDA artificial intelligence might be sentient. Google engineer Blake Lemoine carried out a series of seemingly personal conversations with the artificial intelligence and walked away believing that there was a sort of person behind the messages he was receiving.

Artificial intelligence expert Gary Marcus thinks the idea that artificial intelligence systems are anywhere close to sentience is patently absurd. He wrote on his Substack:

Neither LaMDA nor any of its cousins (GPT-3) are remotely intelligent. All they do is match patterns, draw from massive statistical databases of human language. The patterns might be cool, but language these systems utter doesn’t actually mean anything at all. And it sure as hell doesn’t mean that these systems are sentient.

On Dead Cat, Tom Dotan and I talked to Marcus about artificial intelligence, how tech companies should frame these text generating machines to their users, and the media’s failure to cover speculative technologies skeptically. (In the post we make reference to Marcus’s post Does AI really need a paradigm shift?)

Give it a listen.

Listen on Apple

Listen on Spotify

Read the automated transcript.