13
13

Rotate King to get Queen: Word Relationships as Orthogonal Transformations in Embedding Space

Abstract

A notable property of word embeddings is that word relationships can exist as linear substructures in the embedding space. For example, gender\textit{gender} corresponds to womanman\vec{\textit{woman}} - \vec{\textit{man}} and queenking\vec{\textit{queen}} - \vec{\textit{king}}. This, in turn, allows word analogies to be solved arithmetically: kingman+womanqueen\vec{\textit{king}} - \vec{\textit{man}} + \vec{\textit{woman}} \approx \vec{\textit{queen}}. This property is notable because it suggests that models trained on word embeddings can easily learn such relationships as geometric translations. However, there is no evidence that models exclusively\textit{exclusively} represent relationships in this manner. We document an alternative way in which downstream models might learn these relationships: orthogonal and linear transformations. For example, given a translation vector for gender\textit{gender}, we can find an orthogonal matrix RR, representing a rotation and reflection, such that R(king)queenR(\vec{\textit{king}}) \approx \vec{\textit{queen}} and R(man)womanR(\vec{\textit{man}}) \approx \vec{\textit{woman}}. Analogical reasoning using orthogonal transformations is almost as accurate as using vector arithmetic; using linear transformations is more accurate than both. Our findings suggest that these transformations can be as good a representation of word relationships as translation vectors.

View on arXiv
Comments on this paper