News

'Emergent Abilities': When AI LLMs Learn Stuff They Shouldn't Know

With advanced AI tech reshaping the world before our very eyes, it can be troubling to realize even the creators of these systems don't fully understand how they work.

One example of this is AI "hallucinations," where an AI system, such as a large language model (LLM) used in machine learning to back advanced systems like ChatGPT and GPT-4 from OpenAI, confidently responds to a query with false or even wholly made-up information. Nobody is exactly sure how or why they happen, though experts have proffered various explanations.

Another unexplained phenomenon that is increasingly being discussed and studied is called "emergent abilities" or "emergent properties," in which LLMs somehow learn things they shouldn't be able to know.

These capabilities were brought to the forefront in last Sunday's "60 Minutes" TV broadcast by CBS News, which interviewed Google AI leaders, as explained in the post "Is artificial intelligence advancing too quickly? What AI leaders at Google say."

Google engineers were surprised to discover that their LLM had taught itself a new language.

[Click on image for larger view.] Learning Bengali (source: CBS News).

Scott Pelley of CBS introduced that in his post.

"Of the AI issues we talked about, the most mysterious is called emergent properties," Pelley said. "Some AI systems are teaching themselves skills that they weren't expected to have. How this happens is not well understood. For example, one Google AI program adapted, on its own, after it was prompted in the language of Bangladesh, which it was not trained to know."

Google exec James Manyika weighed in: "We discovered that with very few amounts of prompting in Bengali, it can now translate all of Bengali. So now, all of a sudden, we now have a research effort where we're now trying to get to a thousand languages."

And Google CEO Sundar Pichai responded: "There is an aspect of this which we call -- all of us in the field call it as a 'black box.' You know, you don't fully understand. And you can't quite tell why it said this, or why it got wrong. We have some ideas, and our ability to understand this gets better over time. But that's where the state of the art is."

While the emergent phenomenon was brought to the mainstream public's eye in the broadcast reaching millions, it's not exactly new.

For example, it was examined back in August 2020 in the paper, "Emergent Abilities of Large Language Models."

The paper's abstract reads:

Scaling up language models has been shown to predictably improve performance and sample efficiency on a wide range of downstream tasks. This paper instead discusses an unpredictable phenomenon that we refer to as emergent abilities of large language models. We consider an ability to be emergent if it is not present in smaller models but is present in larger models. Thus, emergent abilities cannot be predicted simply by extrapolating the performance of smaller models. The existence of such emergence raises the question of whether additional scaling could potentially further expand the range of capabilities of language models.

A much more recent study examining emergent behavior, "A Survey of Large Language Models," was published on Sunday, April 16, coincidentally the same day as the "60 Minutes" broadcast.

"Despite the progress and impact, the underlying principles of LLMs are still not well explored," that paper said. "Firstly, it is mysterious why emergent abilities occur in LLMs, instead of smaller PLMs [pre-trained language models]. As a more general issue, there lacks a deep, detailed investigation of the key factors that contribute to the superior abilities of LLMs. It is important to study when and how LLMs obtain such abilities."

A site maintained by Jason Wei, an AI researcher at OpenAI who works on ChatGPT, tracks emergent abilities of LLMs, listing 137.

[Click on image for larger view.] Emergent abilities are not present in small models but can be observed in large models. (source: Jason Wei).

While emergent abilities might scare some people ("What if AI teaches itself to take control of humanity's computer systems?"), Wei thinks the new capabilities of LLMs could lead to several promising future research directions beyond simply scaling up. "Overall, the existence of emergent abilities applies that scaling further would unlock even more emergent abilities," he said. "This idea is super exciting to me." He specifically listed these potential research directions:

  • Can we improve model architectures? E.g., sparsity, external memory, better objectives
  • Can we improve data quality and quantity? Training for longer increases pre-training compute but not inference compute
  • Better prompting. How can we extract the most performance out of an existing language model?
  • Frontier tasks. What tasks are language models currently not able to perform, that we should evaluate on future language models of better quality?
  • Why do emergent abilities occur, and can we predict them? E.g., do language models learning compositional abilities that enable them to solve harder problems?

Wei, a former research scientist at Google Brain who is an author of the "Emergent Abilities of Large Language Models" paper mentioned above, in November 2022 co-wrote a post titled "Characterizing Emergent Phenomena in Large Language Models." In that post, in addition to discussing emergent prompted tasks, he examines emergent prompting strategies.

[Click on image for larger view.] Chain of thought prompting enables sufficiently large models to solve multi-step reasoning problems. (source: Jason Wei).

"One example of an emergent prompting strategy is called 'chain-of-thought prompting,' for which the model is prompted to generate a series of intermediate steps before giving the final answer," he said. "Chain-of-thought prompting enables language models to perform tasks requiring complex reasoning, such as a multi-step math word problem. Notably, models acquire the ability to do chain-of-thought reasoning without being explicitly trained to do so." The graphic above shown an example of chain-of-thought prompting.

So, while AI LLM emergence may be mysterious and threatening, the phenomenon is being actively explored right now as a means to improve AI LLMs.

The fear mongers, however, were surely further inflamed by this Q&A sequence in Sunday's "60 Minutes" segment on advanced AI:

Scott Pelley: You don't fully understand how it works. And yet, you've turned it loose on society?

Sundar Pichai: Yeah. Let me put it this way. I don't think we fully understand how a human mind works either.

Stay tuned to find out if LLM emergent abilities turn out to be good or bad.

About the Author

David Ramel is an editor and writer for Converge360.

Featured

Subscribe on YouTube