GPT-3: The Magic Wand✨that is both Good👍 and Bad👎
- deepak Gandepalli
- Jun 18, 2021
- 9 min read
Updated: Aug 29, 2021
Some are calling it a hype, while some are claiming it as the "greatest invention since the internet" itself. For whatever the reasons, the newest entry to the world of AI is dictating all the attention. So fasten your seat belts💺 and layback🍿 as we break down the magic wand of writing✍️.

Targeted Ads, something that was thought of to be a great marketing tool for businesses, ended up being a dreadful “persona test” for the users. You wouldn’t want to share your device with your parents and thereafter feel embarrassed😬. You are not alone! and this needs to end🙅♂️.
So, a group of “selected intellectuals🧠” decides to interrogate the tech moguls🤴 of silicon valley with a clear objective to figure out how these tech companies collect the user data. The quiz comes off nicely with questions being passed due to lack of time up until an old woman🧓 at a corner of the round table asks, “What’s the flavor of your website cookies?😵 Chocolate or Peanut Butter?🤣 Yes or No?😆”

Several scientists🔬 across the globe🌎 have recognized this catastrophe and started preparation to address the issue. They are trying to make the best use of artificial intelligence💡, basically to enhance the standards of the questions during testimony at the Congress and also for other relatively less important stuff like climate change, public health, and terrorism.
Stepping up to the challenge, OpenAI, an AI (they call it the AGI) research company, cofounded by the unrivaled Dogefather🐶, Elon Musk came to the rescue.

In May of 2020 (a year that was assumed to be the worst year for mankind, until 2021 came), the company released a paper titled, ”Language Models are Few Shot Learners”, which introduced to something called GPT-3.
So, What is GPT-3🧐?
In short, it produces text✍️📝 and a little🤏 beyond that.
In detail, GPT-3 stands for Generative Pre-trained Transformer 3. It is a neural network language model based on deep learning. The pre-trained algorithms produce almost all forms of text (whisper🤫: so humanly!).
Although it isn’t the first NLP (Natural Language Processing) model, it certainly the best😎 among all other models available currently. And what makes it stand out from the rest is the mass of its data.
GPT-3🥇 has 1 7 5 B parameters🤯.
Just to tell you how big it is, the second-best, Microsoft’s Turing NLG🥈 has 17B parameters followed by Google's, T5🥉, and the previous model in the GPT series, GPT-2🏅 with 11B and 1.5B parameters respectively.

Undeniably, GPT-3 is larger than Turing NLG, T5, GPT-2, and all the memes on Ever Given🚢 combined😅.
The more the parameters, the better is the model. Underlying fact: SIZE MATTERS 🙈🤫(GPT-3).
It was trained with multiple unlabeled datasets including the common crawl, English Wikipedia, various books, and over a trillion words from social media like Reddit and Twitter.

If I were the CEO of the company, I would have named GPT-3 as GPT-15 given how enhanced it is to its previous incarnation, GPT-2. It's almost like building a rover immediately after inventing the wheel, leaving planes, rockets, and even flying cars in between🤪.

Another aspect at which it is better than GPT-2 is the number of shots it needs (no harm to its liver though!😌). Shots are basically the examples or demonstrations the user has to give to perform a certain task for the NLP model. GPT-3 requires way less🤏 and in some cases, no prompt (a condition which is called no-shots) is required to accomplish a task😲.

Well, How does it work?🤓:
When the user gives the text input, the system looks for patterns🖧 in the data. It predicts the probability of the text to follow and then auto-completes with the most probable one (just the way Google auto-completes text but on a large scale and in a much better way).
For example, you typed "the room is cold. Switch on the....". The chances are more that it returns "heater" than "cooler". It is because there are many references in that web that heater provides warmth and cooler keeps the room cool. Makes sense, right👍. (For a detailed technical understanding behind its magic, check out Imtiaz Adam's article.)
Here's what GPT-3 thinks about blockchain:
"All cryptocurrency if fraud. it is backed by nothing real and it is made out of thin air (except the electricity used)."
Ugh🤧, appears like Elon Musk manipulated GPT-3 too😡. But never mind, #Bitcoin₿ fixes this😉🙌.
What can it do?
The self-attention mechanism of GPT-3 allows it to translate languages. So finally, time to ditch Google Translate, which typically translates Hindi to Hindi😖 instead of Hindi to English. Rather use a GPT-3 based translator.

A search engine, which when typed in a question gives the most appropriate Wikipedia link as the result. But nothing to worry about, it doesn’t seem to compete with the speed of our all-time favorite Internet Explorer🤠 anytime soon. Neither any of the search engines😏.
As said already, GPT-3 produces text and a little beyond that. It can do some creative stuff🧩 as well.
GPT-3 can auto-complete images also. Yes, just like auto-completing sentences, It can auto-complete images too🙄.
ps👐: This image is a work by GPT-2 and not GPT-3.

In one of the demos, GPT-3 was seen changing the style of the writing. All you need to do is to input the text in plain language and let GPT-3 do its magic. In this example, the user changed “plain language” into “legal language”😯.
Want a decent physics teacher for you👨🏻🏫? GPT-3 has got you covered. With GPT-3 enabled chatbot, learnfromanyone.com, you can pretty much learn anything from anyone. It doesn’t matter if they are at the other end of the globe, out of the globe, or even in the grave⚰️👻!
Physics from #AskFeyman, literature from #AskShakespeare, and revolutionizing blood testing from #AskHolmes😉. Anything and anyone and anytime⏰.
But the conversation that caught my eye is this👇. A user asked few questions using GPT-3 to GPT-3 and the result will make your blood run cold😨.

Of all of them, the most intriguing and ridiculously insane use case is its ability to write code with input instructions in just plain English😱! RIP☠️ the self-taught devs who learned programming from Google, YouTube, GitHub, Stack Overflow, Tinder, and OnlyFans😛.
Back in 2017, a group of scientists predicted that machines would be able to write code by 2040. And just in GPT- is proving them all wrong🤭.
At the recently concluded Build, Microsoft's annual software development conference, the company announced the first commercial use case of the GPT-3. By integrating it with its apps it enables anyone to code🤩. This moment will be remembered in history as the second-best decision taken by Microsoft while the first one still being getting rid of Steve Ballmer🙃.

These are just a few examples, and there are plenty more use cases of GPT-3, some of which haven't been figured out yet. While it is fascinating for some people, it is also sending chills across some industries. Most businesses are now trying to quickly get their hands on GPT-3 to not miss out on the future.
If you are still stuck in 2015🤙 not knowing what an NFT is, hit👊 this banner👇. But before that complete reading this first.😜
The internet is awestruck😲, and how?
Ever since the release, GPT-3 has made the internet buzzing and left the technologists amused🤗. Thomas Smith in his article has probably given the greatest praise GPT-3 would ever receive👌. He quoted,
"I can say with no irony or hyperbole that GPT-3👑 is the most important technical object i've seen since the internet itself and certainly the most significant artificial intelligene technology created in this millennium."
Unlike other NLP models, GPT-3's uniqueness comes with its versatility. It is rewriting✍️ the boundaries and a new revolution in AI has just begun as NLP models will no longer be just a copy-and-paste text from the web.

Drawbacks too👎:
On the flip side of it though, critics are raising concerns (After all, humans had never stood on the same ground🤦🏼♂️, except for trusting that Bill Gates is tracking people with the chips implanted along with the covid vaccine🕵️) about the rapidly increasing artificial intelligence🤖 which might destruct the humanity.

Skeptics fear that GPT-3 could be used as a tool by vicious people to spread hate speech, misinformation, and fake news📰 , which might leave journalists jobless🤷♂️. And to their belief, it has already started to creep in🙆.
On Reddit, a bot with the user name, /u/thegentlemetre, interacted with other users on /r/AskReddit forum for about a week. The bot was posting many “large and deep”😌 posts rapidly and fooling the other users. Later, a Redditor found that the bot was powered by a GPT-3 tool, philosopher AI, designed to answer life and philosophy questions. Although the posts were harmless, it gives us a glimpse of potential misuse🥴.

Another possible threat with the use of advanced AIs is fraudulent academic writing and essays📘. For them being almost indistinguishable from human writings could cause a lot of chaos.
A college student🎓, Liam Porr used GPT-3 to write a fake blog post, and the result: It went viral🔥 and even hit the top spot in Hacker news that too in just a matter of time⌛. All the hard work😫 that he did was giving a headline and intro to the GPT-3 and the rest, the easier work was completed by it😧. Porr wrote in another blog post that only one person had reached him to inquire if it was AI-generated🥶.

He told in an interview,
“It was super easy, actually, which was the scary part😰”
Me📢: Just to clear the air, this article was written by an organism that descended from apes🐵, with 46 DNA🧬, and has passed several I’M NOT A ROBOT🤖🚫 tests.
Targeted Ads💭: I sense he's trying to make AI write an article for him. Might happen if he completes all his overdue pending assignments🤞. So, stick around by subscribing to the blog👇👇👇.

One of the least spoken downsides📉 of GPT-3 (and AI in general) is the enormous amount of energy⚡ required to build and use them. Rob Toews wrote in Forbes about the carbon footprint👣 AI has,
"Modern AI models consume a massive amount of energy requirements are growing at a breathtaking rate. In the deep learning era, the computational resources needed to produce a best-in-class AI model has on average doubled every 3.4 months; this translates to a 300,000x🔋 increase between 2012 and 2018. GPT-3 is just the latest embodiment of this exponential trajectory😟."
Another area of concern of GPT-3 is its inability to say "I really don't know, dude🙅♂️". Yes, it has a huge amount of data and is extremely intelligent but the pitfall is when it answers what's called "common sense" questions. It does better than most other machines and also Donald Trump, but humans want mooore😀.
Here's one of the tests (biological reasoning) conducted and published on MIT Technology Review.

Although it was mentioned in the prompt that you are drinking cranberry and grape🍇 juice, GPT-3 mistakes it to be poison🥀😭.
What they also said🤔:
And this is not the only instance where GPT-3 has failed to answer logical and common sense questions. All we can say is that it has the ability to fool humans but is still quite mindless. Kevin Lacker has given GPT-3 a turning test and in his blog, he mentioned,
"GPT-3 is quite impressive🥰 in some areas, and still clearly subhuman in others😥."
Sumers-Stay aptly described GPT-3 as,
"GPT is odd because it doesn’t 'care' about getting the right answer to a question you put to it. It’s more like an improv actor who is totally dedicated to their craft, never breaks character, and has never left home but only read about the world in books. Like such an actor, when it doesn’t know something, it will just fake it. You wouldn’t trust an improv actor playing a doctor👩⚕️ to give you medical advice."
OpenAI, What are you thinking🤔?
Of course, OpenAI is wary of the possible complications GPT-3 could bring, so it's being cautious in its approach. Currently, GPT-3 is not available for commercial purposes, but there is a waitlist for beta users and researchers. (You can sign up here🤤 if you wish to play with it along with having a good amount of experience of standing in long queues!🤦♂️).
But the company has made plans to expand it to commercial use and avail some profits🤑, again something that is debated for their not-for-profit tag.😬
The beta test demos have shown up what risks GPT-3 offers and it's important to curb them before releasing it to the wild. So, OpenAI is working on ways to improve the model to make sound more sensible and less toxic by fine-tuning on smaller datasets👏.
So, What's next🤲?
In a nutshell🥜, GPT-3 isn't perfect, it has many flaws to be fixed. But in all fairness, it is a great advancement🏆 in the field of AI. For good🤞 or bad to come, it is a wonderful advert to how AI will be in the near future.
Hmm, enough said GPT-3 is (almost) ready to serve its purpose of creation, which is to tackle the most pressing problem of our generation: "Dumb questions at the testimony🥱".
This will eventually end the occasional pop-up ads on your mobile as GPT-3 would be assisting the quiz masters in retrieving back our personal data from the tech giants🤘. I am quite optimistic that it would make the world a better place to live😇.
All seems alright at the moment, but we must never forget that constant up-gradation of the technology is necessary for it to prepare to solve much harder problems in the future. Just think about it, in the year 2040, a robot👾 as the CEO of a company outfoxing🦊 the GPT-3. we may not deal with it then….
Oh, wait🙆♂️, a man is laughing with a black face.
Is it a lizard🦎?
No🚫.
Is it a robot🤖?
No🚫.
It’s a lizard robot🦎🤖! We may never get our data back🤢. Continue to feel embarrassed in front of your parents😭.

Comments