PDFLINK |

# The Structure of Meaning in Language: Parallel Narratives in Linear Algebra and Category Theory

## Introduction

Categories for AI, an online program about category theory in machine learning, unfolded over several months beginning in the fall of 2022. As described on their website https://cats.for.ai, the “Cats for AI” organizing committee, which included several researchers from industry including two from DeepMind, felt that the machine learning community ought to be using more rigorous compositional tools and that category theory has “great potential to be a cohesive force” in science in general and in artificial intelligence in particular. While this article is by no means a comprehensive report on that event, the popularity of “Cats for AI”—the five introductory lectures have been viewed thousands of times—signals the growing prevalence of category theoretic tools in AI.

One way that category theory is gaining traction in machine learning is by providing a formal way to discuss how learning systems can be put together. This article has a different and somewhat narrow focus. It’s about how a fundamental piece of AI technology used in language modeling can be understood, with the aid of categorical thinking, as a process that extracts structural features of language from purely syntactical input. The idea that structure arises from form may not be a surprise for many readers—category theoretic ideas have been a major influence in pure mathematics for three generations—but there are consequences for linguistics that are relevant for some of the ongoing debates about artificial intelligence. We include a section that argues that the mathematics in these pages rebut some widely accepted ideas in contemporary linguistic thought and support a return to a structuralist approach to language.

The article begins with a fairly pedantic review of linear algebra which sets up a striking parallel with the relevant category theory. The linear algebra is then used to review how to understand word embeddings, which are at the root of current large language models (LLMs). When the linear algebra is replaced, *Mad Libs* style, with the relevant category theory, the output becomes not word embeddings but a lattice of formal concepts. The category theory that gives rise to the concept lattice is a particularly simplified piece of enriched category theory and suggests that by simplifying a little less, even more of the structure of language could be revealed.

## Objects Versus Functions on Objects

When considering a mathematical object that has little or incomplete structure, one can replace by something like “functions on which will have considerably more structure than ” Usually, there is a natural embedding . so that when working with one is working with all of , and more.

The first example that comes to mind is the set of functions on a set , valued in a field which forms a vector space. The embedding , is defined by sending to the indicator function of defined by When . is finite, there is a natural inner product on defined by making into a Hilbert space. The physics “ket” notation for the indicator function of nicely distinguishes the element from the vector and reminds us that there is an inner product. The image of in defines an orthonormal spanning set and if the elements are ordered, then the vectors become an ordered basis for and each basis vector has a *one-hot coordinate vector*: a column vector consisting of all zeroes except for a single in the This, by the way, is the starting point of quantum information theory. Classical bits -entry. are replaced by quantum bits which comprise an orthonormal basis of a two-dimensional complex Hilbert space There might not be a way to add elements in the set . or average two of them for example, but those operations can certainly be performed in , In coordinates, for instance, if . then the sum , will have all zeroes with in both the s and - and the sum -entries has all zeroes and a in the -entry.

When the ground field is the field with two elements the vector space structure seems a little weak. Scalar multiplication is trivial, but there are other notable structures on , Elements of . can be thought of as subsets of the correspondence being between characteristic functions and the sets on which they are supported. So , has all the structure of a Boolean algebra: the join and the meet of two vectors correspond to the union and intersection of the two subsets defined by and and neither the meet nor the join coincide with vector addition. Every vector has a “complement” defined by interchanging , the vectors in , are partially ordered by containment, every vector has a cardinality defined by the number of nonzero entries, and so on.

Another closely related example comes from category theory. By replacing a category by the set-valued functors on , one obtains a category with significantly more structure. Here the “op” indicates the variance of the functors in question (contravariant, in this case), a technical point that isn’t very important here, but is included for accuracy. It’s common to call a functor , in a *presheaf* on The Yoneda lemma provides an embedding . of the original category as a full subcategory of Given an object . in Grothendieck used , to denote a *representable presheaf* which is defined by mapping an object to the set of morphisms from to the object represting the presheaf. In this notation, the Yoneda embedding is defined on objects by And just as the vector space . has more structure than the category , has more structure than For any category . the category , of presheaves is complete and cocomplete, meaning that the categorical limits and colimits of small diagrams exist in the category, and more. It is also an example of what is called a *topos* which is a natural place in which to do geometry and logic.

As an illustration, consider a finite set which can be viewed as a discrete category , that is, a category whose only morphisms are identity morphisms. In this case , and a presheaf , on assigns a set to every object in If the elements of . are ordered, then can be thought of as a column vector whose entries are sets, with the set in the The representable functor -entry. can be thought of as a column vector whose entries are all the empty set except for a one-point set in the Using -entry. for and for produces the same arrays as the one-hot basis vectors that span the vector space Notably, the categorical coproduct . does not exist in the category but the coproduct , of the representable functors and *does* exist in the category of presheaves on If . then is a column consisting of empty sets except for a one-point set in the and - the coproduct -entries; consists of all empty sets except for a two point set in the And just as the indicator functions form a basis of the vector space -entry. every functor , is constructed from representable functors. When is a finite set, every vector in is a linear combination of basis vectors, and analogously every presheaf in is a colimit of representables.

In this article, it will be helpful to consider *enriched category theory*, which is the appropriate version of category theory to work with when the set of morphisms between two objects is no longer a *set*. That is, it may be a partially ordered set, or an Abelian group, or a topological space, or something else. So, enriched category theory amounts to replacing with a different category. This is analogous to changing the base field of the vector space and if the new base category has sufficiently nice structure, then most everything said about replacing , by goes over nearly word-for-word. For example, replacing by the category which is a category with two objects , and and one nonidentity morphism results in the category , of presheaves on -valued In the case when . is a set viewed as a discrete category the presheaves in , are exactly the same as functions on -valued which are the same as subsets of , The structure on . afforded by it being a category of presheaves is the Boolean algebra structure on the subsets of -enriched previously described. The categorical coproduct of presheaves -enriched and is the join (union) and the categorical product is the meet (intersection) , So for any set . the set of functions , can either be viewed as a vector space over the field with two elements, or as enriched presheaves on , valued in depending on whether we think of , as a field, or as a category and we get different structures depending on which point of view is taken. ,

Now, before going further, notice that replacing an object by a free construction on that object can’t immediately reveal much about the underlying object. Whether it’s the free vector space on a finite set resulting in or the free cocompletion of a category resulting in the structures one obtains are ,*free* and employ the underlying object as little more than an indexing set. The structures on the “functions” on are owed, essentially, to the structure of what the functions are valued *in*. For example, the source of the completeness and cocompleteness of is the completeness and cocompleteness of the category of sets. Similarly, vector addition and scalar multiplication in arise from addition and multiplication in the field The point is that passing to a free construction on . provides some extra room in which to investigate and in the theory of vector spaces, for instance, things become interesting when linear transformations are involved. As another example, passing from a finite group , to the free vector space doesn’t tell you much about the group, until that is, you involve the elements of the group as operators The result is the regular representation for . which among its many beautiful properties, decomposes into the direct sum of irreducible representations with every irreducible representation of , included as a term with meaningful multiplicity.

This brings us back to the strategy suggested in the first line of this section. When only a little bit is known about the internal structure of an object an approach to learn more is to replace , by something like functions on and study how the limited known structure of interacts with the freely defined structures on the functions of A choice is required of what, specifically, to value the functions in and how mathematically that target is viewed. The remaining sections of this article can be interpreted simply as working through the details in an example with linguistic importance for a couple of natural choices of what to value the functions in. .

## Embeddings in Natural Language Processing

In the last decade, researchers in the field of computational linguistics and natural language processing (NLP) have taken the step of replacing words, which at the beginning only have the structure of a set, ordered alphabetically, by vectors. One gets the feeling that there is structure in words—words appear to be used in language with purpose and meaning; dictionaries relate each word to other words; words can be labelled with parts of speech; and so on—though the precise mathematical nature of the structure of words and their usage is not clear. Language would appear to represent a significant real-world test of the strategy to uncover structure described in the previous section. While the step of replacing words by vectors constituted one of the main drivers of current advances in the field of artificial intelligence, it is not readily recognizable as an instance of replacing a set by functions on that set. This is because replacing words by vectors is typically performed implicitly by NLP tools, which are mathematically obscured by their history—a history which we now briefly review.

Following the surprisingly good results obtained in domains such as image and sound processing, researchers working to process natural language in the 2010s became interested in the application of deep neural network (DNN) models. As a reminder, in its most elementary form, a DNN can be described as a function explicitly expressed as a composition:

1where the are matrices, the are “biases”, the is a (nonlinear) “activation” function, and is an output function. After fixing the activation and output functions, a DNN lives in a moduli space of functions parametrized by the entries of the matrices and the bias vectors Training a DNN is the process of searching through the moduli space for suitable . and to find a function that performs a desired task, usually by minimizing a cost function defined on the moduli space. In the particular case of natural language tasks, this typically requires feeding linguistic data into the model and setting the optimization objective to the minimization of the error between the actual and intended outputs, with the optimization performed by a form of gradient descent.

Significantly, in this setting, linguistic data (typically words) would be represented as vectors—the domain of a DNN is a vector space. A natural first choice for practitioners was then to represent words as one-hot vectors. Thus, if one has a vocabulary consisting of, say, words, then embeds words as the standard basis vectors in a real vector space. -dimensional

Likewise, if the output is to be decoded as a word, then the output function is to be interpreted as a probability distribution over the target vocabulary, which is usually the same though it could certainly be different in applications like translation. ,

At the time, a DNN was thought of as an end-to-end process: whatever happens between the input and the output was treated as a black-box and left for the optimization algorithm to handle. However, a surprising circumstance arose. If the first layer (namely, in Equation 1) of a model trained for a given linguistic task was used as the first layer of another DNN aimed at a different linguistic task, there would be a significant increase in performance of the second task. It was not long thereafter that researchers began to train that single layer independently as the unique hidden layer of a model that predicts a word given other words in the context. Denoting that single layer by one can then obtain ,

3which embeds in a vector space of much lower dimension, typically two or three hundred. Take, for instance, the vector representations made available by ea, where the word is mapped to a vector with components:

In this way, the images of the initial one-hot basis vectors under the map could be used as low-dimensional dense word vector representations to be fed as inputs across multiple DNN models. The word “dense” here is not a technical term but is used in contrast to the original one-hot word vectors in which are “sparse” in the sense that all entries were 0 except for a single , On the other hand, the word vectors in . generally have nonzero entries.

The set of word vector representations produced in this way—also known as “word embeddings”—were found to have some surprising capabilities. Not only did the performance of models across different tasks increase substantially, but also unexpected linguistic significance was found in the vector space operations of the embedded word vectors. In particular, the inner product between two vectors shows a high correlation with semantic similarity. As an example, the vector representations for , , , and , are among the ones with the largest inner product with that of Even more surprisingly, addition and subtraction of vectors in the embedding space correlate with analogical relations between the words they represent. For instance, the vector for . minus the vector for is numerically very near the vector for minus the vector for MSC+13, all of which suggests that the word vectors live in something like a *space of meanings* into which individual words embed as points. Subtracting the vector for from the vector for does not result in a vector that corresponds to any dictionary word. Rather, the difference of these vectors is more like the concept of a “capital city”, not to be confused with the vector for the word which is located elsewhere in the meaning space. ,

Word vector representations such as the one described are now the standard input to current neural linguistic models, including LLMs which are currently the object of so much attention. And the fact, as suggested by these findings, that semantic properties can be extracted from the formal manipulation of pure syntactic properties—that meaning can emerge from pure form—is undoubtedly one of the most stimulating ideas of our time. We will later explain that such an idea is not new but has, in fact, been present in linguistic thought for at least a century.

But first it is important to understand why word embeddings illustrate the utility of passing from to introduced in the previous section. The semantic properties of word embeddings are not present in the one-hot vectors embedded in Indeed, the inner product of the one-hot vectors corresponding to . and is zero, as it is for any two orthogonal vectors, and the vector space operations in are not linguistically meaningful. The difference between the one-hot vectors for and is as far away from the difference between the one-hot vectors for and as it is from the difference between the one-hot vectors for and or any other two one-hot vectors. Linguistically significant properties emerge only after composing with the embedding map in 3 that was obtained through neural optimization algorithms, which are typically difficult to interpret.

In the specific case of word embeddings, however, the algorithm has been scrutinized and shown to be performing an implicit factorization of a matrix comprised of information about how words are used in language. To elaborate, the optimization objective can be shown to be equivalent to factorizing a matrix where the , - is a linguistically relevant measure of the term-context association between words -entry and That measure is based on the pointwise mutual information between both words, which captures the probability that they appear near each other in a textual corpus .LG14. The map is then an optimal low-rank approximation of That is, one finds . and of sizes and respectively, with such that , is mimimal. The upshot is that neural embeddings are just low-dimensional approximations to the columns of Therefore, .*the surprising properties exhibited by embeddings are less the consequence of some magical attribute of neural models than the algebraic structure underlying linguistic data found in corpora of text*. Indeed, it has since been shown that one can obtain results comparable to those of neural word embeddings by directly using the columns of or a low-dimensional approximation thereof, as explicit word-vector representations ,LGD15. Interestingly, the other factor which is readable as the second layer of the trained DNN, is typically discarded although it does contain relevant linguistic information. ,

In summary, the math story of word embeddings goes like this: first, pass from the set of vocabulary words to the free vector space While there is no meaningful linguistic information in . it provides a large, structured setting in which a limited amount of information about the structure of , can be placed. Specifically, this limited information is a matrix consisting of rough statistical data about how words go with other words in a corpus of text. Now, the columns of or better yet, the columns of a low-rank factorization of , then interact with the vector space structure to reveal otherwise hidden syntactic and semantic information in the set , of words. Although a matrix of statistical data seems more mathematically casual than, say, a matrix representing the multiplication table of a group, it has the appeal of assuming nothing about the structure that might possess. Rather, it is purely a witness of how has been used in a particular corpus. It’s like a set of observations is the input, and a more formal structure is an output.

So, if word embeddings achieved the important step of finding a linguistically meaningful *space* in which words live, then the next step is to better understand what is the *structure* underlying that space. Post facto realizations about vector subtraction reflecting certain semantic analogies hint that even more could be discovered. For this next discussion, it is important to understand that there is an exact solution for the low-rank factorization of a matrix using the truncated singular value decomposition (SVD), which has a beautiful analogue in category theory. To fully appreciate the analogy, it will be helpful to review matrices from an elementary perspective.

## From the Space of Meanings to the Structure of Meanings

In this section, keep in mind the comparison between functions on a set valued in a field and functors on a category valued in or another enriching category. Now, let’s consider matrices.

Given finite sets and an , - matrix valued in a field is a function By simple currying, . defines functions and defined by and Ordering the elements of . and the function , can be represented as a rectangular array of numbers with rows and columns with the value being the number in the row and -th column. The function -th is then identified with the row of the matrix, which has as many entries as the elements of -th and defines a function on sending to the entry in the row. Similarly, the -th column of -th represents the function Linearly extending the maps . and produces linear maps and which of course are the linear maps associated with the matrix , and its transpose Here is a diagram: .

Now, the compositions and are linear operators with special properties. If we fix the ground field to be the real numbers we can apply the spectral theorem to obtain orthonormal bases , of and of consisting of eigenvectors of and respectively with shared nonnegative real eigenvalues where , This data can be refashioned into a factorization of . as This is the so-called singular value decomposition of . The . are the columns of the , are the rows of and , is the diagonal matrix whose entry is -th The matrices . and satisfy and In SVD terminology, the nonnegative real numbers . are the *singular values* of and the vectors , and are the *left and right singular vectors* of In other words, we have pairs of vectors . in related to each other as

Moreover, these pairs are ordered with if the corresponding singular values satisfy Finally, it is not difficult to show that the matrix . with rank at most that is closest in Frobenius norm to the matrix is given by where is the diagonal matrix containing only the greatest nonzero singular values on the diagonal. By eliminating the parts of , and , that do not, because of all the zero entries in participate in the product , one obtains a factorization , where , is an matrix, is an matrix, and is an diagonal matrix with the largest singular values of on the diagonal. Principal component analysis (PCA) employs this approximate factorization of a matrix into low-rank components for dimensionality reduction of high-dimensional data.

Moving from linear algebra to category theory, one finds a remarkably similar story. Given two categories and the analogy of a , - matrix is something called a *profunctor*, which is a set-valued functor As before, the .“ here and in what follows is used to indicate the variance of functors and is needed for accuracy, but can on first reading be ignored. Experts will surely know of situations in which this ” is involved in interesting mathematical dualities, but for the analogy with linear algebra described here, it can be thought of as indicating a sort of transpose between rows and columns. If both domain categories are finite sets viewed as discrete categories, then a profunctor is simply a collection of sets indexed by pairs of elements—that is, a matrix whose entries are sets instead of numbers. Again, by simple currying, a profunctor defines a pair of functors and defined on objects by and As in the linear algebra setting, the functor . can be pictured as the row of sets in the matrix -th which defines a functor , where the object of -th is mapped to the set in the row -th The functor . can be similarly be pictured as the column of -th Thinking of a category as embedded in its category of presheaves via the Yoneda (or co-Yoneda) embedding, the functors . and can be extended in a unique way to functors and that preserve colimits and limits, respectively.

Now, just as the composition of a linear map and its transpose define linear maps with special properties, the functors and are adjoint functors with special properties. This particular adjunction is known as the *Isbell adjunction*, which John Baez recently called “a jewel of mathematics” in a January 2023 column article in this publication Bae23. Objects that are fixed up to isomorphism under the composite functors and are called the *nuclei* of the profunctor and are analogous to the left and right singular vectors of a matrix. One can organize the nuclei into pairs of objects in where ,

The nuclei themselves have significant structure—they organize into a category that is complete and cocomplete. The pairs of singular vectors of a matrix have some structure—they are ordered by the magnitude of their singular values, and the magnitudes themselves are quite important. The nuclei of a profunctor have a different, in some ways more intricate, structure because one can take categorical limits and colimits of diagrams of pairs, allowing the pairs to be combined in various algebraic ways. In the context of linguistics, this is significant because the nuclei are like abstract units and the categorical limits and colimits provide ways to manipulate those abstract units. This is illustrated concretely in the next section. For now, interpret word embeddings obtained from the singular vectors of a matrix as a way to overlay the structure of a vector space on meanings. For certain semantic aspects of language, like semantic similarity, a vector space structure is a good fit, but overlaying a vector space structure could veil others. The Isbell adjunction provides a different structure that may help illuminate other structural features of language.

For flexibility, it is useful to look at the Isbell adjunction in the enriched setting. If the base category is instead of then a profunctor , between two finite sets and viewed as discrete categories enriched over , is just a function , which is the same as a relation on , The functors . and are known objects in the theory of formal concept analysis GW99. The function maps a subset to the set and maps a subset to the set The fixed objects of . and are known as *formal concepts*. They are organized into pairs with

and the set of all formal concepts is partially ordered with if and only if which is equivalent to Moreover, . forms a complete lattice, so, like the singular vectors of a matrix, there is a least and a greatest formal concept, and more. The product and coproduct, for example, of formal concepts are defined by and The point is that limits and colimits of formal concepts have simple, finite formulas that are similar to, but not exactly, the union and intersection of sets and give an idea of the kind of algebraic structures one would see on the nucleus of a profunctor. .

## Structures in the Real World

If there’s any place where neural techniques have indisputably surpassed more principled approaches to language, it is their capacity to exhibit surprisingly high performance on empirical linguistic data. Whatever the nature of formal language models to come, it will certainly be decisive to judge their quality and relevance in the real world. In this section, we illustrate how the tools of linear algebra and enriched category theory work in practice, and in the conclusion we will share how the empirical capabilities of the enriched category theory presented here can be used to do more.

To start, consider the English Wikipedia corpus comprising all Wikipedia articles in English as of March 2022 Wik. If we consider this corpus as a purely syntactic object without assuming any linguistic structure, then the corpus appears as a long sequence of a finite set of independent tokens or characters. To simplify things, let’s restrict ourselves to the 40 most frequent characters in that corpus (excluding punctuation), which account for more than 99.7% of occurrences. So, our initial set contains the following elements:

Now let and, in line with what we have seen in the previous sections, consider a matrix representing some linguistically relevant measure of the association between the elements of and A straightforward choice for . is the empirical probability that the characters are the left and right contexts of the character For instance, we have that . while reflecting that it is over a hundred times more probable to see the sequence than given , as the center character.

Considered as elements of each character is independent and as different as it can be from all the others. However, embedding them , via the matrix and leveraging the relationships they exhibit in concrete linguistic practices as reflected by a corpus brings out revealing structural features. Indeed, if we perform an SVD on the induced operator we can obtain a vector representation of each character.Footnote^{1} Figure 1 shows a plot of all characters in as points in a three-dimensional space, where the coordinates are given by the singular vectors corresponding to the three largest singular values, scaled by those singular values. We can see how what were originally unrelated elements now appear organized into clusters in the embedding space with identifiable linguistic significance. Namely, the elements are distinguished as vowels, consonants, digits, and special characters.

^{1}

For reasons beyond the scope of this paper, it’s convenient to take the square roots of the entries and center the matrix around before performing the SVD.

What’s more, the dimensions of the embedding space defined by the singular vectors of have a little bit of natural structure—there is a canonical order given by the singular values that is endowed with linguistic significance. Looking at the first three singular vectors, we see that the first one discriminates between digits and letters, the second one distinguishes vowels from the rest, and the third one identifies special characters (see Figure 2). The rapid decay of the successive singular values indicates that dimensions beyond three adds only marginal further distinctions (Figure 3).

While the decomposition into singular values and vectors reveals important structural features—each singular vector discriminates between elements in a reasonable way, and the corresponding singular values work to cluster elements into distinct types—the linear algebra in this narrow context seems to run aground. However, as discussed in the previous sections, we can gain further and different structural insights by considering our sets and as categories or enriched categories. As a primitive illustration, view and as discrete categories enriched over For the next step, a . matrix -valued is required. A simple and rather unsophisticated choice is to establish a cutoff value (such as to change the same matrix ) used in the SVD above into a matrix. All the entries of -valued less than the cutoff are replaced with and all the entries above the cutoff are replaced with , .

Then, extend to obtain functors and and look at the fixed objects of and which are organized as described earlier into formal concepts , that form a highly structured and complete lattice. Visualizing the lattice in its entirety is challenging in a static two-dimensional image. To give the idea in these pages, one can look at a sublattice defined by selecting a single character and looking at only those concepts for which contains the selected character. For Figure 4, the characters and are selected and to further reduce the complexity of the images, only nodes representing a large number of contexts, are drawn. ,

Right away the lattices make clear the distinction between digits and letters, but there is also more. Each set of characters is associated to an explicit dual set of contexts suggesting a principle of compositionality—namely, the elements of the corresponding classes can be freely composed to produce a sequence belonging to the corpus. Such composition reveals relevant features from a linguistic viewpoint. While digits tend to compose with other digits or special characters, vowels compose mostly with consonants. Strictly speaking, a similar duality was present in the SVD analysis, since left singular vectors are canonically paired (generically) in a one-to-one fashion with right singular vectors, which discriminate between contexts in based on the characters for which they are contexts. The formal concepts, on the other hand, as fixed objects of and display dualities between large but discrete classes of characters. Numbers, consonants, and letters are all distinguished, but finer distinctions are also made. Moreover, the collection of such dual classes is not just a set of independent elements but carries the aforementioned operations , and allowing one to perform algebraic operations on the concept level.