What does “Mathematics of Language” mean? Before we approach that question, what do languages and mathematics have in common?
Well, let’s start with a practical experiment to establish the relevance of these questions first. Try to read the following sentence:
If you are a native English speaker, you were most probably able to read the above (enciphered) sentence. This essay covers how this phenomenon connects our language usage to mathematics, and how we can benefit from this connection.
Let’s start with analysing what is going on with the above-enciphered sentence.
This essay is supported by Generatebg
The Mathematics of Language: Less = More
If you were perceptive enough, you would have noticed that I just replaced all the vowels in the English language with a triangle symbol in the enciphered sentence. So, in order to make sense of it, all you had to do was to “guess” the missing vowels correctly.
So, in essence, we just established that we can still make sense of a sentence even if we take the letters (in this case, the vowels) out of words. If you are still sceptical about this “claim”, let’s test it one more time using SMS-lingo:
If we can communicate more efficiently using lesser letters, why do we still use long-form sentences with complex grammatical structures and rules? To answer this question, we need to understand Shannon entropy first.
Claude Shannon: The Underrated Genius
Claude Elwood Shannon is, in my opinion, one of the most underestimated science figures of all time.
As a 21-year-old master’s degree student, he single-handedly defined the fate of an entire branch of science (now known as information theory) in his master’s thesis.
Not only did he ask ground-breaking questions in his thesis, but he also went on to provide most of the answers as well. Usually, such progress in science happens through a series of minor innovations over decades. But Shannon made a discrete jump possible, which led to the digital computer revolution that we still enjoy today.
From computers to smartphones, Shannon’s revolutionary master’s thesis is still being used in our current technological devices. On a personal note, some of my own technological innovations have been inspired by Shannon’s work. Needless to say, he is one of my all-time heroes!
Among all of his influential contributions, two notable ones are relevant for our current purposes:
1. Development of Shannon entropy as a measure of information content in a message.
2. Prediction and entropy of printed English (reference link at the end of the essay).
Shannon Entropy
Shannon defined information entropy (known as Shannon entropy) as a measure of the amount of information that could be transmitted by a message. To understand this further, let us take an intuitive approach rather than look at the mathematical formulation.
Let us consider a coin that features “heads” on both sides. Now consider this as an information system. We are interested in the least number of questions we need to ask to ascertain the state after a toss.
When we toss such a rigged coin, regardless of whichever side comes on top, we know that the outcome would be “heads”. Therefore, no question is necessary. According to Shannon’s approach, such a message system has zero entropy.
Now, consider a fair coin with “heads” on one side and “tails” on the other side. If we toss the coin twice, what is the least number of questions we need to ask to ascertain each substate? The answer is two — one for the outcome of each toss.
According to Shannon’s approach, such an information system would carry two bits of entropy. A “bit” here is a measure of information. That’s right — it was in Shanon’s master thesis that the term “bit” was expressed for the first time. Shannon attributed the usage of the word to fellow mathematician, John Tukey.
Now, consider a two-letter English word whose alphabets are not known. What is the least number of questions that we need to ask to ascertain the word? Well, what if we use the following approach?
Question: Does the first letter belong to the first half of the English alphabet or the latter half?
Based on the answer, one could repeat the process to ascertain the letter:
In this example, we were able to ascertain the state using 4 steps. This also means that this information system contains 4 bits of entropy. To save us some effort, let us say that we figured out that the second letter was “O” using 5 steps. Therefore, the word “GO” contains 9 bits of entropy.
Order and Randomness
You see, the word “GO” contains 9 bits of entropy only if you consider the occurrence of the English alphabet as a random variable.
However, the English language conveniently features order in the form of grammar and other conventions. A few examples of the order are the following:
1. The letter “E” appears much more often than “Z”.
2. The letter “Q” is highly likely to be followed by the letter “U”.
If you think this is fascinating enough, not only did Shannon take notice of all this, but he dove deeper! Through the application of statistics and experimentation, he recorded patterns and calculated the entropy of the English language to be 2.62 bits per letter on average.
Now, where do we go from here with this knowledge?
The Mathematics of Language: Redundancy
Shannon was interested in using these results to develop models of prediction and compression for information transmission (in the English language). In a sense, he wanted to maximise entropy available in a given bandwidth architecture.
Broadly speaking, most natural phenomena exhibit a trend where entropy only increases with time (empirical observation). Rather paradoxically, the history of the English spelling system suggests that we are moving from higher entropy to lower entropy over time.
This is because human beings (and nature) have slightly different intentions. It appears that whenever the importance of a message increases, we prefer to use long formulations with many redundancies packed into the text. With such an architecture, we make sure that even if a few letters or words are missed, the core of the message is still held intact.
For example, consider a case where you are texting your best friend to plan a party and another case where you are writing an application for your dream job. In the former, you are likely to use compressed short messages, whereas a job application demands a lot more attention to detail. This in turn leads to more redundancies in the language used for the dream job application.
Filling In the Blanks
The highly redundant nature of the English language combined with our cognitive pattern-seeking nature is what enables us to comprehend sentences even when vowels or letters are taken out.
As serial optimisers, we might think that there is room for more efficiency to be gained in our usage of the English language. But historically speaking, the evolution of the language has been in the exact opposite direction.
The gradual increase of structural redundancy has made English easier and easier to use over the past decades and centuries. This in turn has made the language significantly more resilient to errors in communication and robust in usage. It is then no wonder why English has thrived as a Lingua Franca across the globe over time.
Final Remarks
Have you ever wondered why a scratched Compact Disc (CD) works fine most of the time? It is because CDs have redundant bits and error correction protocols encoded onto them. In this way, they are more resilient to errors. So, the rabbit hole goes beyond just the English language.
The moral of this essay is that not everything is worth optimising for efficiency. Sometimes, resilience to error and robustness are of higher priority than just blind efficiency!
Reference: Claude Shannon (research paper).
I hope you found this article interesting and useful. If you’d like to get notified when interesting content gets published here, consider subscribing.
Further reading that might interest you: How Does The The Human Brain Miss The Second ‘The’? and Are We Living In A Simulation?
Comments