“Artificial intelligence (AI) is already having a significant impact on the economy, and its influence is expected to grow significantly in the coming years … Overall, the effects of AI on the economy will depend on a variety of factors, including the rate of technological advancement, government policies and the ability of workers to adapt to new technologies.”
OK, who said that? Nobody, unless we’re ready to start calling large language models people. What I did was ask ChatGPT to describe the economic effects of artificial intelligence; that’s just an excerpt.
Many of us who’ve played around with large language models — which are being widely discussed under the rubric of artificial intelligence — have been shocked by how much they now manage to sound like people. And it’s a good bet that they or their descendants will eventually take over many tasks that are now done by humans.
Like previous leaps in technology, this will make the economy more productive but will also probably hurt some workers whose skills have been devalued. How large will these effects be? And how quickly will they come about? On the first question, nobody really knows. Predictions about the economic impact of technology are notoriously unreliable. On the second, history suggests that large economic effects from AI will take longer to materialize than many people expect.
Consider the effects of previous advances in computing. Gordon Moore, a founder of Intel — which introduced the microprocessor in 1971 — predicted that the number of transistors on a computer chip would double every two years — a prediction that proved stunningly accurate for half a century. The consequences of Moore’s Law are all around us, most obviously in the smartphones that almost everyone carries around these days.
For a long time, however, the economic payoff from this awesome rise in computing power was surprisingly elusive. For at least two decades after Moore’s Law kicked in, America, far from experiencing a productivity boom, suffered from a protracted productivity slowdown. The boom kicked in only during the 1990s, and even then it was a bit disappointing.
Why did a huge, prolonged surge in computing power take so long to pay off for the economy? In 1990, economic historian Paul David published one of my favorite economics papers, “The Dynamo and the Computer.” It drew a parallel between the effects of information technology and those of an earlier tech revolution, electrification.
As David noted, electric motors became widely available in the 1890s. But having a technology isn’t enough. You also have to figure out what to do with it.
To take full advantage of electrification, manufacturers had to rethink the design of factories. Pre-electric factories were multistory buildings with cramped working spaces, because that was necessary to make efficient use of a steam engine in the basement driving the machines through a system of shafts, gears and pulleys.
It took time to realize that having each machine driven by its own motor made it possible to have sprawling one-story factories with wide aisles allowing easy movement of materials and assembly lines. As a result, the big productivity gains from electrification didn’t materialize until after World War I.
Sure enough, as David, in effect, predicted, the economic payoff from information technology finally kicked in during the 1990s, as filing cabinets and secretaries taking dictation finally gave way to cubicle farms. The lag in this economic payoff even ended up being similar in length to the lagged payoff from electrification.
But this history still presents a few puzzles. One is why the first productivity boom from information technology (there may be another one coming, if the enthusiasm about chatbots is justified) was so short-lived; basically, it lasted only around a decade.
And even while it lasted, productivity growth during the IT boom was no higher than it was during the generation-long boom after World War II, which was notable in the fact that it didn’t seem to be driven by any radically new technology.
The great boom from the 1940s to around 1970 seems to have been largely based on the use of technologies, like the internal combustion engine, that had been around for decades — which should make us even more skeptical about trying to use recent technological developments to predict economic growth.
That’s not to say that artificial intelligence won’t have huge economic impacts. But history suggests that they won’t come quickly. ChatGPT and whatever follows are probably an economic story for the 2030s, not for the next few years.