the health strategist
research institute for continuous transformation
– of health & care
Joaquim Cardoso MSc
Chief Researcher & Editor
March 19, 2023
EXECUTIVE SUMMARY
OpenAI has launched its new artificial intelligence model, GPT-4, which has been described as more advanced than its predecessor, GPT-3.
- OpenAI has kept the technical details about GPT-4 private due to concerns over both competitive advantage and safety implications.
- GPT-4 is accessible through the $20 paid version of ChatGPT and is capable of accepting both text and image input, generating and ingesting far bigger volumes of text than its predecessor, and has more advanced reasoning and parsing abilities.
- However, OpenAI has warned of its potential to provide detailed information on illegal activities and generate hateful and discriminatory speech.
- GPT-4 has already been used to create sophisticated chatbots by companies such as Morgan Stanley Wealth Management and Duolingo, demonstrating its moneymaking potential.
DEEP DIVE
GPT-4 from OpenAI shows advances — and moneymaking potential
Financial Times
March 18, 2023
Klinton Bicknell was let into one of the technology world’s great secrets last September. The head of AI at the language learning app Duolingo, was given rare access to GPT-4, a new artificial intelligence model created by Microsoft-backed OpenAI.
He soon discovered the new AI system was even more advanced than OpenAI’s earlier version used to power the hit ChatGPT chatbot that provides realistic answers in response to text prompts.
Within six months, Bicknell’s team had used GPT-4 to create a sophisticated chatbot of their own that human users could talk with, in order to practice conversational French, Spanish and English as if they were in real-world settings like airports or cafés.
“It was amazing how the model had such detailed and specialised knowledge of how languages work and of the correspondences between different languages,” said Bicknell. “With GPT-3, which we had already been using, this just would not be a viable feature.”
Duolingo is one of a handful of companies, including Morgan Stanley Wealth Management and online education group Khan Academy, given prior access to GPT-4, before it was launched more widely this week.
The release reveals how OpenAI has transformed from a research-focused group into a company worth almost $30bn, racing giants such as Google in efforts to commercialise AI technologies.
OpenAI announced that GPT-4 showed “human-level” performance on a range of standardised tests such as the US Bar exam and the SAT school tests, and showed off how its partners were using the AI software to create new products and services.
They’re cherry-picking these tasks, because there is no scientifically agreed-upon set of benchmarks
Alex Hanna, the Distributed AI Research Institute
But for the first time, OpenAI did not reveal any details about the technical aspects of GPT-4, such as what data it was trained on or the hardware and computing capacity used to deploy it, because of both the “competitive landscape and the safety implications”.
This represents a shift since OpenAI was created as a non-profit in 2015, in part, the brainchild of some of the tech world’s most radical thinkers, including Elon Musk and Peter Thiel. It was built on the principles of making AI accessible to everybody through scientific publications, and developing the technology safely.
A pivot in 2019 turned it into a profitmaking enterprise with a $1bn investment from Microsoft. That was followed this year by a further multibillion-dollar funding from the tech giant, with OpenAI quickly becoming a crucial part of Microsoft’s bet that AI systems will transform its business model and products.
This transformation led Musk, who left OpenAI’s board in 2018, to tweet this week that he was “still confused as to how a non-profit to which I donated ~$100mn somehow became a $30bn market cap for-profit. If this is legal, why doesn’t everyone do it?”
OpenAI’s lack of transparency regarding the technical details of GPT-4 has drawn criticism from others within the AI community.
“It’s so opaque, they’re saying ‘trust us, we’ve done the right thing’,” said Alex Hanna, director of research at the Distributed AI Research Institute (DAIR) and a former member of Google’s Ethical AI team. “They’re cherry-picking these tasks, because there is no scientifically agreed-upon set of benchmarks.”
GPT-4, which can be accessed through the $20 paid version of ChatGPT, has shown rapid improvement to earlier AI models on certain tasks. For instance, GPT-4 scored in the 90th percentile on the Uniform Bar Exam taken by would-be lawyers in the US. ChatGPT only reached the 10th percentile.
While OpenAI did not provide details, AI experts believe the size of the model is larger than previous generations and that it has had a lot more human training to fine tune it.
The most obvious new feature is that GPT-4 can accept input in both text and image form — although it only responds using text. This means users can upload a photo to ask the model to describe the picture in painstaking detail, request ideas for a meal made with ingredients present in the image, or ask it to explain the joke behind a visual meme.
GPT-4 is also able to generate and ingest far bigger volumes of text, compared to other models of its type: users can feed in up to 25,000 words compared with 3,000 words into ChatGPT. This means it can handle detailed financial documentation, literary works or technical manuals.
Its more advanced reasoning and parsing abilities mean it is far more proficient at analysing complex legal contracts for risks, said Winston Weinberg, co-founder of Harvey, an AI chatbot that was built using GPT-4 and is used by PwC and magic circle law firm Allen & Overy.
Despite these advances, OpenAI has warned of several risks and limitations of GPT-4. This includes its ability to provide detailed information on how to conduct illegal activities — including developing biological weapons, and generating hateful and discriminatory speech.
OpenAI put GPT-4 through a safety testing process known as red-teaming, where more than 50 external experts in disciplines ranging from medicinal chemistry to nuclear physics and misinformation were asked to try to break the model.
Paul Röttger, an AI researcher at the Oxford Internet Institute who focuses on the identification of toxic content online, was contracted by OpenAI for six months to try to elicit harmful responses from GPT-4 and provide feedback, on topics ranging from suicide or self-harm content, to graphic descriptions of violence or examples of extremism and hate speech.
He said that overall the model improved its responses over the months of testing, where it would initially hedge its answers but later become more unequivocal in its responses to bad prompts.
“On one hand, safety research has progressed since GPT-3, and there’s a lot of good ideas that went into making this model safer,” he said. “But at the same time, this model is so much more powerful and can do a lot more things than GPT-3, so the risk surface has gotten a lot bigger too.”
Originally published at https://www.ft.com on March 18, 2023.
Names mentioned (selected list)
- Klinton Bicknell — Head of AI at the language learning app Duolingo
- Morgan Stanley Wealth Management — A financial services firm that was given prior access to GPT-4
- Khan Academy — An online education group that was given prior access to GPT-4
- Google — A technology company that Alex Hanna, director of research at the Distributed AI Research Institute, is a former member of the Ethical AI team
- Elon Musk — Entrepreneur and former board member of OpenAI
- Peter Thiel — Entrepreneur and co-founder of PayPal, who was involved in the creation of OpenAI
- Distributed AI Research Institute (DAIR) — A research organization where Alex Hanna is the director of research
- Oxford Internet Institute — An academic department of the University of Oxford where Paul Röttger, an AI researcher, works
- PwC — A professional services firm that uses the AI chatbot Harvey, built using GPT-4
- Allen & Overy — A law firm that uses the AI chatbot Harvey, built using GPT-4
- Winston Weinberg — Co-founder of Harvey, an AI chatbot that was built using GPT-4 and is used by PwC and Allen & Overy.
- Microsoft