Large language models (LLMs) — e.g., BERT, GPT-3, — and chatbots like ChatGPT, are quite impressive. In the coming years, they will likely lead to a number of transformative tools and applications. They may well herald a coming AI revolution. But, how much closer are they getting us to the kind of general intelligence that, for the foreseeable future, only humans have? How can we best sort out their true significance and value from the accompanying hype? Let me discuss how two different articles addressed these questions.
“The success of the large neural language models on many NLP [Natural Language Processing] tasks is exciting” wrote linguistic professors Emiliy Bender and Alexander Koller in their 2020 paper “Climbing towards NLU: On Meaning, Form, and Understanding in the Age of Data.” “This has led to claims, in both academic and popular publications, that such models understand or comprehend natural language or learn its meaning. From our perspective, these are overclaims caused by a misunderstanding of the relationship between linguistic form and meaning.”
In their paper, Bender and Koller explain why LLMs like GPT-3 are likely to become innovative language tools, — kind of like highly advanced spell checkers or word processors, — while dismissing claims that they have the ability to reason and understand the meaning of the language they’re generating. Their explanations are couched in linguistic concepts which they carefully define: form, communicative intent, meaning, and understanding.
Continue reading "Are GPT-3 and ChatGPT Truly Intelligent?" »