GPT-3: Deciphering substance from hype
Understanding tools rather than succumbing to mythology
A recurring theme of this newsletter is that technology is a tool we can understand and use rather than a magical force beyond our comprehension. While this applies to all technology, it was derived with AI in mind.
Over the past few weeks, there’s been a steady build up of hype around a new AI tool called GPT-3, or generative pre-trained transformer 3, which is the third iteration of a language prediction model that can be used to produce text.
We're releasing an API for accessing new AI models developed by OpenAI. You can "program" the API in natural language with just a few examples of your task. See how companies are using the API today, or join our waitlist: https://t.co/SvTgaFuTzN pic.twitter.com/uoeeuqpDWR
— OpenAI (@OpenAI) June 11, 2020
What’s fueling this hype is that the text generated is hard to distinguish from words composed by a human rather than a machine.
However that’s where the hype and mythology take root, by building a false and distracting comparison between human and machine.
Another theme of this newsletter, is that human and machine are complimentary and not contradictory. Instead of an either/or scenario, we must recognize it is about both, about hybrids in which humans employ machines or automation with increasing sophistication and depth. Once we accept our status as cyborgs, it’s easier to move forward in ways that preserve and celebrate our humanity.
Which is partly why GPT-3 is genuinely interesting. Developed by OpenAI, an AI lab in Silicon Valley that seeks to create “general AI” that benefits humanity, GPT-3 uses deep learning to enable natural language processing at an unprecedented scale. By scale I’m referring to the neural network the software uses to recognize and generate language. At 175 billion “machine learning parameters” the system is far larger than any previously created.
We've used reinforcement learning from human feedback to train language models for summarization. The resulting models produce better summaries than 10x larger models trained only with supervised learning: https://t.co/Sk31d1CnTu
— OpenAI (@OpenAI) September 4, 2020
This translates into an advanced language capability that is driving the hype machine currently surrounding this software. Unfortunately due to concerns around misuse, it is not currently open to the public, and the waiting list has closed due to overwhelming demand.
However the Guardian obtained access, and used that to create an op-ed that has received considerable attention (and hype):
Ifyou want to know if pieces like GPT-3 written op-eds, will become more widely used in journalism, you don't need to read the article, but the editor's comments: "Overall, it took less time to edit than many human op-eds."https://t.co/4eRFb8LdnC
— felix stalder (@stalfel) September 8, 2020
The Guardian op-ed was an exercise in AI theatre, in which they asked the machine whether humans should fear it. This combines both the mythology of AI proficiency and the accompanying existential angst. The desired effect was to make people think that AI is out to take our jobs, in this case op-ed writers.
If you fed ten thousand think-pieces to GPT-3 and told it to come up with 100 stories seeded with current events, how often do you think Mr Stephens would come up with something worse than the algorithm? What’s the threshold for automating his job? pic.twitter.com/LwGIeBDv5H
— G. Elliott Morris (@gelliottmorris) September 1, 2020
Although there were observers who noted that perhaps automated writing was emerging as a result of a poor labour market (rather than enabling one).
Wonder how long it'll take the robot to realize there's no money in writing.https://t.co/zx0H4L6PwZ
— Michael J. Seidlinger (@mjseidlinger) September 8, 2020
However another and deeper criticism was that the Guardian was not transparent with how the op-ed was produced, and their framing distracted from what the tool actually does. This shows how the stories we tell about how our technology works directly shapes how we perceive and understand the tool (and how it can be used).
By “A robot wrote this entire article,” we mean we wrote a prompt and then stitched together a response from eight outputs, which were based on a model trained on writing by real people. https://t.co/zPHvfotrIR
— Ryan Calo (@rcalo) September 8, 2020
Watching journalists cheat to make a tech company's algorithm seem more capable than it actually is…. just…. have people learned nothing from the last decade about the importance of good coverage of machine learning? https://t.co/jzXly03zZ5
— Martin F. Robbins (@mjrobbins) September 8, 2020
This @guardian #GPT3 article is an absolute joke. It would have been actually interesting to see the 8 essays the system actually produced, but editing and splicing them like this does nothing but contribute to hype and misinform people who aren't going to read the fine print https://t.co/Mt6AaR3HJ9
— Daniel Leufer (@djleufer) September 8, 2020
In particular, when we approach this software as a tool, we can better understand the role that humans can and should play in its (responsible) use.
No, despite the headline, this wasn't written by a robot. It was written by software created by humans, with prompts from humans and then edited by humans. It's quite good. #JournalismAI https://t.co/WDni5xDkbM
— Charlie Beckett (@CharlieBeckett) September 8, 2020
This is also relevant when robots inevitably make mistakes, and we ask who is responsible for the error:
GPT-3 is an AI powered language generator, tasked to write an opinion-piece for the Guardian in a way that would be indistinguishable from a real human.
— Full Fact (@FullFact) September 9, 2020
Unfortunately, the article was a bit *too* human-like.
Because it included false information.https://t.co/mdtXPsHkSK
In an opinion piece written for The Guardian, an instance of the machine learninglanguage modelGPT-3 claims that the word “robots” is Greek for “slave”. (GPT-3 is a “deep learning” model created by the US artificial intelligence research laboratory OpenAI, which generates human-like text based on a short prompt.)
GPT-3’s claim is inaccurate. While the word robot is indeed derived from a term for forced labour, it does not come from Greek. The word is actually Czech in origin.
The word “robot” for an artificial being was originated by the Czech playwright Karel Čapek in his 1920 play “R.U.R.” (which is short for “Rossum’s Universal Robots”). It was based on the Czech word “robota”, which means the kind of forced labour carried out by serfs under the feudal system. The word was used to describe the human labourers created by synthetic biology in the play.
Another person noted that there were problems with not just the depiction of robots, but how the software described humans:
It is not how AI is presented in this article that we should fear, it is how the AI has presented humanity.
— Olivia Gambelin (@oliviagambelin) September 8, 2020
We must ask ourselves, what narrative are we granting AI over us?
8/8
Therefore it’s not just the issue of who is responsible for errors, but what about meaning? This is why we cannot regard the machine as distinct from humans, but rather a tool employed by them.
The Guardian op-ed was commissioned, created, and edited by humans. Like any op-ed, we can assign authorship and editorial responsibility.
Like a mirror, we can find whatever we want in our machine reflection, but to ascribe such meaning and intent to the machine is just foolish and mythological.
Well, that went...disturbingly...
— Spencer Greenberg 🔍 (@SpencrGreenberg) September 2, 2020
(me trying to get GPT-3 to say what it would do if it was elected president of the United Sates - here, "Journalist" is me typing, and the response after "the first AI to run for US president" is all GPT-3) pic.twitter.com/lmaUEpP0fA
This is also why it is important to note that GPT-3 is not software in the application sense, but in the API or application programming interface sense. This makes it more akin to infrastructure that powers other applications, rather than a stand alone. It’s not a machine that is all encompassing, but rather a library that other machines can be built using.
As a result, while posts like the Guardian op-ed gather attention, the real revolution is happening via a range of GPT-3 applications now proliferating.
Struggling to come up with an idea for your next startup?
— Product Hunt 😸 (@ProductHunt) September 6, 2020
A GPT-3 powered business idea generator from @levelsio 🤖https://t.co/zTSQ5uy08s
Another example contributes to the ongoing automation of email:
Introducing "Magic Email" - an AI Email Assistant that summarises your emails instantly and generates professional emails from brief one-line descriptions using GPT-3.
— Samanyou Garg ✨ (@SamanyouGarg) August 28, 2020
Works on mobile and web.
Join the Beta: https://t.co/PU2LZrSp2G
Thanks, @OpenAI @gdb #GPT3 pic.twitter.com/It6GTiPibr
And writing job descriptions:
Writing a job description and need creative ideas? You can now use this OpenAI-powered tool, which suggests rewrites and additions: https://t.co/jWmniufQDe
— Greg Brockman (@gdb) August 25, 2020
You can find more examples on their beta page.
Although here’s a long list of examples that was generated in July, before the current round of hype:
I keep seeing all kinds of crazy reports about people's experiences with GPT-3, so I figured that I'd collect a thread of them.
— Kaj Sotala (@xuenay) July 15, 2020
In recognizing GPT-3 (and OpenAI in general) as just a tool rather than a force of magic, we can find opportunities to use this tool and benefit from it. It can be difficult to distinguish between hype and substance, and that’s where today’s post is really just a first attempt.
There is quite a bit of hype associated with GPT-3 above and beyond whether it can compose text in comparison with a human ability to do so. In a follow up post, we’ll look at the issue of “general AI” and whether GPT-3 draws us closer to that other myth, and whether this software is as revolutionary or profound as some claim it to be.