Summary: Artificial intelligence can understand complex words and concepts by representing the meaning of words in a similar way that correlates with human judgments.
Source: UCLA
In โThrough the Looking Glass,โ Humpty Dumpty says scornfully, โWhenย Iย use a word, it means just what I choose it to mean โ neither more nor less.โ Alice replies, โThe question is whether youย canย make words mean so many different things.โ
The study of what words really mean is ages old. The human mind must parse a web of detailed, flexible information and use sophisticated common sense to perceive their meaning.
Now, a newer problem related to the meaning of words has emerged: Scientists are studying whether artificial intelligence can mimic the human mind to understand words the way people do. A new study by researchers at UCLA, MIT andย the National Institutes of Healthย addresses that question.
The paper, published in the journalย Nature Human Behaviour, reports that artificial intelligence systems can indeed learn very complicated word meanings, and the scientists discovered a simple trickย to extract that complex knowledge.
They found that the AI system they studied represents the meanings of words in a way that strongly correlates with human judgment.
The AI system the authors investigated has been frequently used in the past decade to study word meaning. It learns to figure out word meanings by โreadingโ astronomical amounts of content on the internet, encompassing tens of billions of words.
When words frequently occur together โ โtableโ and โchair,โ for example โ the system learns that their meanings are related. And if pairs of words occur together very rarely โ like โtableโ and โplanet,โ โ it learns that they have very different meanings.
That approach seems like a logical starting point, but consider how well humans would understand the world if the only way to understand meaning was to count how often words occur near each other, without any ability to interact with other people and our environment.
Idan Blank, a UCLA assistant professor of psychology and linguistics, and the studyโs co-lead author, said the researchers set out to learn what the system knows about the words it learns, and what kind of โcommon senseโ it has.
Before the research began, Blank said, the system appeared to have one major limitation: โAs far as the system is concerned, every two words have only one numerical value that represents how similar they are.โ
In contrast, human knowledge is much more detailed and complex.
โConsider our knowledge of dolphins and alligators,โ Blank said. โWhen we compare the two on a scale of size, from โsmallโ to โbig,โ they are relatively similar. In terms of their intelligence, they are somewhat different. In terms of the danger they pose to us, on a scale from โsafeโ to โdangerous,โ they differ greatly. So a wordโs meaning depends on context.
โWe wanted to ask whether this system actually knows these subtle differences โ whether its idea of similarity is flexible in the same way it is for humans.โ
To find out, the authors developed a technique they call โsemantic projection.โ One can draw a line between the modelโs representations of the words โbigโ and โsmall,โ for example, and see where the representations of different animals fall on that line.
Using that method, the scientists studied 52 word groups to see whether the system could learn to sort meanings โ like judging animals by either their size or how dangerous they are to humans, or classifying U.S. states by weather or by overall wealth.
Among the other word groupings were terms related to clothing, professions, sports, mythological creatures and first names. Each category was assigned multiple contexts or dimensions โ size, danger, intelligence, age and speed, for example.
The researchers found that, across those many objects and contexts, their method proved very similar to human intuition. (To make that comparison, the researchers also asked cohorts of 25 people each to make similar assessments about each of the 52 word groups.)
Remarkably, the system learned to perceive that the names โBettyโ and โGeorgeโ are similar in terms of being relatively โold,โ but that they represented different genders. And that โweightliftingโ and โfencingโ are similar in that both typically take place indoors, but different in terms of how much intelligence they require.
โIt is such a beautifully simple method and completely intuitive,โ Blank said. โThe line between โbigโ and โsmallโ is like a mental scale, and we put animals on that scale.โ
Blank said he actually didnโt expect the technique to work but was delighted when it did.
โIt turns out that this machine learning system is much smarter than we thought; it contains very complex forms of knowledge, and this knowledge is organized in a very intuitive structure,โ he said. โJust by keeping track of which words co-occur with one another in language, you can learn a lot about the world.โ
The studyโs co-authors are MIT cognitive neuroscientist Evelina Fedorenko, MIT graduate student Gabriel Grand, and Francisco Pereira, whoย leads the machine learning team at theย National Institutes of Healthโs National Institute of Mental Health.
Funding: The research was funded in part by the Office of the Director of National Intelligence, Intelligence Advanced Research Projects Activity through the Air Force Research Laboratory.
About this AI and language research news
Author: Stuart Wolpert
Source: UCLA
Contact: Stuart Wolpert – UCLA
Image: The image is credited to Idan Blank/UCLA
Original Research: Open access.
“Semantic projection recovers rich human knowledge of multiple object features from word embeddings” by Idan Blank et al. Nature Human Behavior
Abstract
Semantic projection recovers rich human knowledge of multiple object features from word embeddings
How is knowledge about word meaning represented in the mental lexicon?
Current computational models infer word meanings from lexical co-occurrence patterns. They learn to represent words as vectors in a multidimensional space, wherein words that are used in more similar linguistic contextsโthat is, are more semantically relatedโare located closer together.
However, whereas inter-word proximity captures only overall relatedness, human judgements are highly context dependent. For example, dolphins and alligators are similar in size but differ in dangerousness.
Here, we use a domain-general method to extract context-dependent relationships from word embeddings: โsemantic projectionโ of word-vectors onto lines that represent features such as size (the line connecting the words โsmallโ and โbigโ) or danger (โsafeโ to โdangerousโ), analogous to โmental scalesโ. This method recovers human judgements across various object categories and properties.
Thus, the geometry of word embeddings explicitly represents a wealth of context-dependent world knowledge.


