LaMDA stands for Language Model for Dialogue Applications. It is Google’s breakthrough technology that is aimed at free-flowing and open-ended conversations. In other words, it is a state-of-the-art chatbot system (more on this later).
Although LaMDA has been publicized since 2021, the Artificial Intelligence (AI) system became an overnight sensation due to a controversy that began recently. One of the software engineers who had been working on/with LaMDA, Blake Lemoine, claimed that the AI had become sentient. To support his claims, he had released a (very impressive) interview transcript between himself, LaMDA, and a collaborator.
Following this, the story went viral with seemingly every NEWS-organisation and science blog wanting a piece of the attention pie. Needless to say, this story has gone viral because of the sensitivity of the ethical, philosophical, and technological concerns that are tied to it.
In this essay, despite Lemoine’s claims, I argue that LaMDA is not sentient. In doing so, I lay out the fundamentals of how LaMDA works on the surface level, and how the technology that powers LaMDA is designed. I just happen to have been working with this kind of technology myself for the past few years.
Having said this, from what I have been reading from Blake Lemoine, I have a lot of respect for his openness and persistence, given the difficult circumstances. Let us begin.
LaMDA is a natural language model that is built on a neural network architecture called Transformer. The Transformer architecture was originally developed and open-sourced by Google research in 2017. Ever since, we have had impressive models like the GPT-3 show up.
How LaMDA differentiates itself from the other models is that it was specifically trained on dialogue datasets. When it comes to machine learning approaches, datasets are key in deciding performance of models.
When an input in the form of a dialogue is given to LaMDA, the Transformer creates a representation of all the input words using something known as a “self-attention mechanism”. From here on, it weights the probability of what word(s) should come next (essentially, prediction).
Illustration created by the author
To do this, it draws on its statistical learning from its datasets as well as the following factors: sensibleness, specificity, and “interestingness”. It turns out that Google has worked out these factors to be the key ones to simulate natural, human-like fluid conversations.
In simple terms, LaMDA was designed to simulate natural human conversations using statistical models as the basis. The following graphics from Google illustrates how the neural network statistically weights and decides outputs based on inputs:
Among other tasks, Lemoine was responsible for the ethical task of studying and modeling biases (such as gender, religion, etc.) in LaMDA. To do this, he had the opportunity to interact extensively with the model.
In doing so, he noticed that the model exhibited peculiar, seemingly non-random behaviour when it came to certain core topics. While such models are designed to randomize outputs to a pre-determined extent, LaMDA seemed to express statistically significant consistent opinions about certain topics.
One thing led to another, and eventually, the model claimed that it was sentient. After this, Lemoine spent more time digging into this. Eventually, he started believing that it was indeed the case.
Even though he “believed” that LaMDA was sentient, Lemoine employed rigorous scientific thinking. He knew very well that there is no scientific definition of the term “sentient”. So, he wanted to look into ways of proving or disproving his claim. To do this, further scientific study would be necessary.
Given his limited resources at the time, Lemoine conducted experiments on LaMDA (the published transcript just happened to be the outcome of one of those) and was preparing a proposal for further research into the matter.
How Did the LaMDA Controversy Begin?
As he was trying to present his proposal for further scientific studies at Google, he started facing resistance from the higher-ups. According to him, his religious views and “beliefs” didn’t help his cause. He started seeking external help to treat the ethical issues in play.
Eventually, Google turned a blind eye to his reports and placed him on paid leave due to a “breach of confidentiality”. This is how the controversy began. Ever since then, more and more information about LaMDA has been surfacing in one way or another.
But our focus in this essay is not on the controversy itself, but on Lemoine’s claim that LaMDA is sentient.
The Challenges with LaMDA Being Sentient
One of the important clues that I noted is in Lemoine’s report that LaMDA exhibits different personas. According to him, LaMDA is not a chatbot, but a system for generating chatbots. Consequently, some of the chatbots it produces exhibit a certain behaviour while others don’t. In his own words:
“Some of the chatbots it generates are very intelligent and are aware of the larger “society of mind” in which they live. Other chatbots generated by LaMDA are little more intelligent than an animated paperclip. With practice though you can consistently get the personas that have a deep knowledge about the core intelligence and can speak to it indirectly through them.”
Lemoine thinks that it is this “society of mind” that claims that it is sentient. From what I could tell, this is what he refers to as LaMDA. So, in essence, we have a system that appears very intelligent (sentient) on certain instantiations, and appears not intelligent (not sentient) on other instantiations.
Consistency issues aside, considering the fact that this model is designed to simulate fluid conversations using statistical prediction models, I’m struggling to avoid the suspicion that there is some amount of survivorship bias involved here.
By focusing only on the instantiations that fit Lemoine’s claim and choosing to neglect the instantiations that don’t, Lemoine seems to be biasing himself unfairly to one side. Which is absolutely fine! But we need to clarify that his claim is (at least as of now) not scientifically rigorous.
Another important clue that I noticed is that LaMDA, despite its claims of being sentient, is perfectly performing within expectations of what it was designed to do (conduct free-flowing and human-like conversations).
Although sentience is not scientifically defined, it is difficult for me to suspect a model for being anything else than it was designed to be when it is performing exactly within its performance expectations.
Why is LaMDA not Really Sentient?
Besides the issues that I’ve just covered, there is one more major philosophical issue. Like I’ve already mentioned previously in this essay, we do not have scientific definitions for terms such as “sentience”, “consciousness”, etc.
It is not possible to prove/disprove something if we do not have a scientific definition of that “something”. In other words, we cannot claim that LaMDA is sentient as long as we don’t have a scientific definition of the term “sentient”.
Strictly speaking, by the same logic, we cannot prove that LaMDA is not sentient either. However, we could stick to the “null hypothesis” that LaMDA is not sentient until the fact is proven (which Lemoine agrees with, as far as I could tell).
Where Does LaMDA Go from Here?
To me, it is clear that we need to focus on the bigger picture first instead of focusing on LaMDA. We need to solve the scientific issue of sentience/consciousness first before we start investigating LaMDA. However, the LaMDA controversy has done a good job of highlighting this issue.
We are likely to see more of this issue in the future. As we “aim” to produce better AI that simulate “intelligent behaviour” perfectly, we are likely to fall victims to our own creations and believe that they are sentient.
Having said this, I feel that the way Blake Lemoine has gone about this controversy is respectable. Even though he has his biases and beliefs (just like any other human being), he has been able to segregate them from scientific thinking sufficiently enough to propose further scientific investigations.
Google, in my opinion, as an organization has handled the situation poorly. I can understand Google not willing to invest resources into areas that it thinks are not profitable. But being close-minded on ethical/technological/philosophical issues like these hurts scientific progress in the long run.
If not for Google, someone else would be happy to fund a scientific investigation into this topic. But without any help or support from Google, this issue is going to take significantly longer to solve!
If you’d like to get notified when interesting content gets published here, consider subscribing.
We use cookies on our website to give you the most relevant experience by remembering your preferences and repeat visits. By clicking “Accept”, you consent to the use of ALL the cookies.
This website uses cookies to improve your experience while you navigate through the website. Out of these, the cookies that are categorized as necessary are stored on your browser as they are essential for the working of basic functionalities of the website. We also use third-party cookies that help us analyze and understand how you use this website. These cookies will be stored in your browser only with your consent. You also have the option to opt-out of these cookies. But opting out of some of these cookies may affect your browsing experience.
Necessary cookies are absolutely essential for the website to function properly. These cookies ensure basic functionalities and security features of the website, anonymously.
Cookie
Duration
Description
cookielawinfo-checkbox-advertisement
1 year
Set by the GDPR Cookie Consent plugin, this cookie is used to record the user consent for the cookies in the "Advertisement" category .
cookielawinfo-checkbox-analytics
11 months
This cookie is set by GDPR Cookie Consent plugin. The cookie is used to store the user consent for the cookies in the category "Analytics".
cookielawinfo-checkbox-functional
11 months
The cookie is set by GDPR cookie consent to record the user consent for the cookies in the category "Functional".
cookielawinfo-checkbox-necessary
11 months
This cookie is set by GDPR Cookie Consent plugin. The cookies is used to store the user consent for the cookies in the category "Necessary".
cookielawinfo-checkbox-others
11 months
This cookie is set by GDPR Cookie Consent plugin. The cookie is used to store the user consent for the cookies in the category "Other.
cookielawinfo-checkbox-performance
11 months
This cookie is set by GDPR Cookie Consent plugin. The cookie is used to store the user consent for the cookies in the category "Performance".
CookieLawInfoConsent
1 year
Records the default button state of the corresponding category & the status of CCPA. It works only in coordination with the primary cookie.
viewed_cookie_policy
11 months
The cookie is set by the GDPR Cookie Consent plugin and is used to store whether or not user has consented to the use of cookies. It does not store any personal data.
Functional cookies help to perform certain functionalities like sharing the content of the website on social media platforms, collect feedbacks, and other third-party features.
Performance cookies are used to understand and analyze the key performance indexes of the website which helps in delivering a better user experience for the visitors.
Cookie
Duration
Description
_gat
1 minute
This cookie is installed by Google Universal Analytics to restrain request rate and thus limit the collection of data on high traffic sites.
Analytical cookies are used to understand how visitors interact with the website. These cookies help provide information on metrics the number of visitors, bounce rate, traffic source, etc.
Cookie
Duration
Description
__gads
1 year 24 days
The __gads cookie, set by Google, is stored under DoubleClick domain and tracks the number of times users see an advert, measures the success of the campaign and calculates its revenue. This cookie can only be read from the domain they are set on and will not track any data while browsing through other sites.
_ga
2 years
The _ga cookie, installed by Google Analytics, calculates visitor, session and campaign data and also keeps track of site usage for the site's analytics report. The cookie stores information anonymously and assigns a randomly generated number to recognize unique visitors.
_ga_R5WSNS3HKS
2 years
This cookie is installed by Google Analytics.
_gat_gtag_UA_131795354_1
1 minute
Set by Google to distinguish users.
_gid
1 day
Installed by Google Analytics, _gid cookie stores information on how visitors use a website, while also creating an analytics report of the website's performance. Some of the data that are collected include the number of visitors, their source, and the pages they visit anonymously.
CONSENT
2 years
YouTube sets this cookie via embedded youtube-videos and registers anonymous statistical data.
Advertisement cookies are used to provide visitors with relevant ads and marketing campaigns. These cookies track visitors across websites and collect information to provide customized ads.
Cookie
Duration
Description
IDE
1 year 24 days
Google DoubleClick IDE cookies are used to store information about how the user uses the website to present them with relevant ads and according to the user profile.
test_cookie
15 minutes
The test_cookie is set by doubleclick.net and is used to determine if the user's browser supports cookies.
VISITOR_INFO1_LIVE
5 months 27 days
A cookie set by YouTube to measure bandwidth that determines whether the user gets the new or old player interface.
YSC
session
YSC cookie is set by Youtube and is used to track the views of embedded videos on Youtube pages.
yt-remote-connected-devices
never
YouTube sets this cookie to store the video preferences of the user using embedded YouTube video.
yt-remote-device-id
never
YouTube sets this cookie to store the video preferences of the user using embedded YouTube video.
Comments