Author: Tony, Researcher at Zhuiyi Technology AI Lab
Everyone knows that language models like BERT have been widely used in natural language processing. However, a question sometimes arises: do these language models truly understand language? Experts and scholars have different opinions on this. The author of this article elaborates on this topic from several aspects, including the essence of language, the relationship between language symbols and meaning, and what language models actually learn from language.
-
unlockable = un+lockable / unlock + able
-
Welcome new teachers to dine = Welcome + new teachers + to dine / Welcome + new + teachers + to dine
-
Shinomiya Kaguya is my wife = I am Shinomiya Kaguya’s husband
-
To understand the symbols and their combination rules within the language;
-
To understand the meaning of language, that is, to understand the form of sense and its reference, as well as how both change with context. However, existing language models are trained based on textual corpora. The data they see iscompliant with language rules, in a certaincontext (a specific time and place, a certain mental state, based on certain knowledge, having made certain inferences, etc.) produced byhuman outputs of a pile of language symbols. If we consider the meaning and context that humans intend to express as input, and the text as output, then language is a function that maps meaning and context to texts that comply with language rules. And it is impossible for a language model to infer the form of the function, that is, the correspondence between input and output, solely based on the possible outputs of the function. Therefore, according to the definition of this article, language models cannotfully understand language.
[1] https://thegradient.pub/nlp-imagenet/
[2] http://faculty.washington.edu/ebender/papers/GeneralPurposeNLU.pdf