bit.ly/3HtHQD0 MIT Museum
314 Main St.
Cambridge, MA 02142
Innovation, Lectures & Conferences, Tech, University
This past summer a Google engineer made international headlines claiming that Google’s AI chatbot system LaMDA is sentient—capable of thinking, feeling and experiencing—because it can generate grammatical, meaningful language. The news generated a number of articles and posts by people in the broader community including researchers in AI, brain and cognitive science, psychology and linguistics.
The Conversation’s article, “Google’s powerful AI spotlights a human cognitive glitch: Mistaking fluent speech for fluent thought” explores how the human mind perceives intelligence in others and how a smart-sounding AI says as much about us as it does about technology. As written in their introductory paragraph, “People are so accustomed to assuming that fluent language comes from a thinking, feeling human that evidence to the contrary can be difficult to wrap your head around. How are people likely to navigate this relatively uncharted territory? Because of a persistent tendency to associate fluent expression with fluent thought, it is natural – but potentially misleading – to think that if an AI model can express itself fluently, that means it thinks and feels just like humans do.”
Join us on December 15 when we explore this question with Josh Tenenbaum and Anna Ivanova at the intersection of language, cognitive science and machine learning.
$10 advance purchase and registration
$15 day of or registration at the door
$5 advanced registration with MIT ID
Seating is limited. Advance purchase recommended.