One of the key findings of the 2022 AI Index Report was that large language models (LLMs) are setting records on technical benchmarks thanks to advances in deep neural networks and computational power that allows them to be trained using huge amounts of data. LLMs are now surpassing human baselines in a number of complex language tasks, including English language understanding, text summarization, natural language inference, and machine translation.
A.I. Is Mastering Language. Should We Trust What It Says?, a recent NY Times Magazine article by science writer Steven Johnson, took a close look at one such LLM, the Generative Pre-Trained Transformer 3, generally referred to as GPT-3. GPT-3 was created by the AI research company OpenAI. It’s been trained with over 700 gigabytes of data from across the web, along with a large collection of text from digitized books. “Since GPT-3’s release, the internet has been awash with examples of the software’s eerie facility with language - along with its blind spots and foibles and other more sinister tendencies,” said Johnson.
“So far, the experiments with large language models have been mostly that: experiments probing the model for signs of true intelligence, exploring its creative uses, exposing its biases. But the ultimate commercial potential is enormous. If the existing trajectory continues, software like GPT-3 could revolutionize how we search for information in the next few years.” Instead of typing a few keywords into Google and getting back a long list of links that might have the answer, you’d ask GPT-3 what you’re looking for in English and get back a well-written, accurate answer. “Customer service could be utterly transformed: Any company with a product that currently requires a human tech-support team might be able to train an L.L.M. to replace them.”
Continue reading "Can We Trust Our Impressive AI Language Models?" »