In the realm of artificial intelligence (AI), large language models (LLMs) are contributing to a slew of innovative use cases. I have been tinkering with these models for a few years now. Not long ago, I recall trying to stitch together an AI conversational partner (a goofy one at that) powered by GPT-2.
Fast forward to the present, here I am power-coding complex applications with the help of models like GPT-4. We went from goofing around to power productivity in a matter of few years. What changed, and where are we headed?
As I am beginning to work ever more intensively with and on large language models, I feel that I have some insight into their future and how we (as a society) might be able to cope with / benefit from them.
Contrary to popular belief today, most of the fundamental science behind LLMs is not new; some of these concepts are even decades old. What is new is the innovation on two primary fronts:
1. Hardware
2. Engineering
Many Machine Learning and Artificial Intelligence “Researchers” would like to believe that the current state of the art LLMs are marvels of science. I don’t think so.
These models are beautiful creations, but they do not add much to the body of scientific truths as we know it.
Instead, AI researchers use our current scientific knowledge to “engineer” complex adaptive systems (LLMs) that are able to solve the challenging problems that we throw at them.
Don’t get me wrong. I am not trying to take any credit away from our excellent AI researchers who have created such models.
I am just differentiating between “science” and “engineering”; one pushes our understanding of fundamental ground truths, and the other uses the said fundamental ground truths to build useful tools for society.
Speaking of tools, none of these “engineering” feats would have been possible without the advancements in computational hardware that we have had over the decades. It has been a key contributor to how these complex adaptive systems are able to function so time-efficiently.
Who is Driving the Current State of the Art?
As I see the landscape, two groups of people are driving the current state of the art of LLMs:
1. AI and Machine Learning Researchers
2. Hackers and Enthusiasts
While the researchers deliver discrete jumps in technological advancement, the hackers and enthusiasts push the boundaries of these models and find interesting and unique applications for them.
This dance of alliance between these two unlikely groups makes for a fascinating and exciting time to be involved with this field.
While I have friends who are active researchers in the field, I belong to the latter category than the former.
My Latest Project
I have been intensely working on a project that involves an image-generation large language model. At this point, you might be asking the following question:
“What does image generation have to do with language?”
Well, it does a lot. Think of it this way: we use words all the time to describe images and we use our “imagination” all the time to construct mental images from words we hear.
In essence, our language is connected to images and vice versa. Throw in advanced image algorithms, tensor calculus, tensor algebra, and a healthy dose of engineering genius, you are left with large language models that enable us to generate and edit images. What a fascinating time to live in?!
I am making rapid progress with my project and all the signs indicate that this will be a product launch soon. The word “product” unfortunately also means that I have to be conservative about what I reveal about it here.
So, that is all I can share about my project at this point, I guess. But what about the future? Where is this journey headed?
How to Read the Future of Large Language Models?
First and foremost, I would like to openly admit that I possess no crystal ball and therefore, cannot say anything about the future with 100% certainty.
However, I do have some unique insights using which I could make some educated bets.
I won’t beat around the bush. Here are 5 educated bets that I am willing to make given my current knowledge of LLMs
1. The use cases for LLMs will continue to increase geometrically.
2. LLMs will not function without human guidance/assistance. This is due to the notion of tail risks that arise from edge cases. No matter how complex your tensor operations are, nature will feature edge cases that are at least an order above and out of reach.
3. Advances in scientific truths will slow down growth of LLMs and will give birth to the next generation of (engineered) artificial intelligence architecture.
4. These newer generations of AI architectures (that are unlikely to be language models) will be the ones that will slowly inch towards performing without human guidance/assistance.
5. LLMs will still strongly serve future AI architectures (that are not language models themselves) as these systems are likely to employ higher-order languages that are beyond human comprehension.
One of more of these points might be controversial takes, but I am not aiming to be scientifically correct here. These are my bets, given my current knowledge of LLMs.
The Future of Humanity
As the current generation of LLMs develops, I believe that we will continue to see more and more complex human problems get solved and automated. As a result, our society will get more efficient and environment-friendly.
This will lead to another problem, though. Many human beings who are in the workforce today will not have much to do in the near-future economy. What then?
The future of LLMs — Illustration created by the author
One solution is Universal Basic Income (UBI). But this still does not solve the problem of “life’s purpose”. Another alternative could be human beings working under LLMs that aim to maximise human fulfilment as well as operational efficiency.
It is indeed a strange thought to not only work with LLMs but work for them.
But given the rate of development of LLMs, I think that it is wise for us to mentally prepare ourselves to cope with both the potentially fruitful as well as the adverse changes that await us in the near future.
If you’d like to get notified when interesting content gets published here, consider subscribing.
We use cookies on our website to give you the most relevant experience by remembering your preferences and repeat visits. By clicking “Accept”, you consent to the use of ALL the cookies.
This website uses cookies to improve your experience while you navigate through the website. Out of these, the cookies that are categorized as necessary are stored on your browser as they are essential for the working of basic functionalities of the website. We also use third-party cookies that help us analyze and understand how you use this website. These cookies will be stored in your browser only with your consent. You also have the option to opt-out of these cookies. But opting out of some of these cookies may affect your browsing experience.
Necessary cookies are absolutely essential for the website to function properly. These cookies ensure basic functionalities and security features of the website, anonymously.
Cookie
Duration
Description
cookielawinfo-checkbox-advertisement
1 year
Set by the GDPR Cookie Consent plugin, this cookie is used to record the user consent for the cookies in the "Advertisement" category .
cookielawinfo-checkbox-analytics
11 months
This cookie is set by GDPR Cookie Consent plugin. The cookie is used to store the user consent for the cookies in the category "Analytics".
cookielawinfo-checkbox-functional
11 months
The cookie is set by GDPR cookie consent to record the user consent for the cookies in the category "Functional".
cookielawinfo-checkbox-necessary
11 months
This cookie is set by GDPR Cookie Consent plugin. The cookies is used to store the user consent for the cookies in the category "Necessary".
cookielawinfo-checkbox-others
11 months
This cookie is set by GDPR Cookie Consent plugin. The cookie is used to store the user consent for the cookies in the category "Other.
cookielawinfo-checkbox-performance
11 months
This cookie is set by GDPR Cookie Consent plugin. The cookie is used to store the user consent for the cookies in the category "Performance".
CookieLawInfoConsent
1 year
Records the default button state of the corresponding category & the status of CCPA. It works only in coordination with the primary cookie.
viewed_cookie_policy
11 months
The cookie is set by the GDPR Cookie Consent plugin and is used to store whether or not user has consented to the use of cookies. It does not store any personal data.
Functional cookies help to perform certain functionalities like sharing the content of the website on social media platforms, collect feedbacks, and other third-party features.
Performance cookies are used to understand and analyze the key performance indexes of the website which helps in delivering a better user experience for the visitors.
Cookie
Duration
Description
_gat
1 minute
This cookie is installed by Google Universal Analytics to restrain request rate and thus limit the collection of data on high traffic sites.
Analytical cookies are used to understand how visitors interact with the website. These cookies help provide information on metrics the number of visitors, bounce rate, traffic source, etc.
Cookie
Duration
Description
__gads
1 year 24 days
The __gads cookie, set by Google, is stored under DoubleClick domain and tracks the number of times users see an advert, measures the success of the campaign and calculates its revenue. This cookie can only be read from the domain they are set on and will not track any data while browsing through other sites.
_ga
2 years
The _ga cookie, installed by Google Analytics, calculates visitor, session and campaign data and also keeps track of site usage for the site's analytics report. The cookie stores information anonymously and assigns a randomly generated number to recognize unique visitors.
_ga_R5WSNS3HKS
2 years
This cookie is installed by Google Analytics.
_gat_gtag_UA_131795354_1
1 minute
Set by Google to distinguish users.
_gid
1 day
Installed by Google Analytics, _gid cookie stores information on how visitors use a website, while also creating an analytics report of the website's performance. Some of the data that are collected include the number of visitors, their source, and the pages they visit anonymously.
CONSENT
2 years
YouTube sets this cookie via embedded youtube-videos and registers anonymous statistical data.
Advertisement cookies are used to provide visitors with relevant ads and marketing campaigns. These cookies track visitors across websites and collect information to provide customized ads.
Cookie
Duration
Description
IDE
1 year 24 days
Google DoubleClick IDE cookies are used to store information about how the user uses the website to present them with relevant ads and according to the user profile.
test_cookie
15 minutes
The test_cookie is set by doubleclick.net and is used to determine if the user's browser supports cookies.
VISITOR_INFO1_LIVE
5 months 27 days
A cookie set by YouTube to measure bandwidth that determines whether the user gets the new or old player interface.
YSC
session
YSC cookie is set by Youtube and is used to track the views of embedded videos on Youtube pages.
yt-remote-connected-devices
never
YouTube sets this cookie to store the video preferences of the user using embedded YouTube video.
yt-remote-device-id
never
YouTube sets this cookie to store the video preferences of the user using embedded YouTube video.
Comments