Against the singularity hypothesis (Part 3: Further doubts)

Machine computation speeds have increased by a factor of about 1011 since World War II. By contrast, power plants have seen modest efficiency gains and face limited prospects given constraints like Carnot’s theorem. This distinction is important, because … output and growth end up being determined not by what we are good at, but by what is essential but hard to improve.

Aghion et al., “Artificial intelligence and economic growth

Listen to this post

1. Recap

This is Part 3 of a series based on my paper “Against the singularity hypothesis“.

The singularity hypothesis begins with the assumption that artificial agents will gain the ability to improve their own intelligence. From there, the singularity hypothesis holds that the intelligence of artificial agents will grow at a rapidly accelerating rate, producing an intelligence explosion: an event in which artificial agents rapidly become orders of magnitude more intelligent than their human creators.

Part 1 introduced and clarified the singularity hypothesis. Part 2 gave two preliminary reasons to doubt the singularity hypothesis. Today’s post gives three further reasons to doubt the singularity hypothesis. Together, these five reasons for doubt will place a strong burden on defenders of the singularity hypothesis to provide significant evidence in favor of their view. The remainder of this series will argue that this burden has not been met.

2. Bottlenecks

One reason why growth processes slow down is that they hit bottlenecks: single obstacles to improvement. Often a single bottleneck is enough to bring growth to a standstill until the bottleneck can be resolved.

A recent paper by Philippe Aghion and colleagues (2017) uses economic modeling to illustrate how bottlenecks can arise in the growth of computing capacities, and what their effects might be. By way of illustration, Aghion and colleagues offer the following example:

Machine computation speeds have increased by a factor of about 1011 since World War II. By contrast, power plants have seen modest efficiency gains and face limited prospects given constraints like Carnot’s theorem. This distinction is important, because with [technical condition omitted], output and growth end up being determined not by what we are good at, but by what is essential but hard to improve.

In this example, the potential inability of power generation to keep pace with growth in computing capacities forces growth in computing outputs to crawl along until energy needs can be met.

Aghion and colleagues’ point is quite general. Even if computing technology were to advance to the point where rapid increases in the intelligence of artificial agents were technologically feasible, those increases might not be realized until bottlenecks were resolved.

One class of bottleneck involves the physical resource constraints discussed in the next section. To rapidly grow the intelligence of artificial agents, we might well have to do all of the following (and plausibly a good bit more):

  • Substantially expand our capacity to generate energy.
  • Substantially expand our capacity to distribute energy through the electrical grid.
  • Procure key materials, including rare metals which must be mined and computer components which must be manufactured.
  • Build manufacturing capacity in capital-intensive industries such as semiconductor manufacturing, by building new plants, hiring and training new laborers, and securing relevant machinery.

In the next section, I will ask whether this could reasonably be done on the scale required to ground the singularity hypothesis. But here, I want to ask a more modest question: how confident are we that all of this could be done quickly enough to ground an intelligence explosion? Are we meant, for example, to see an increase in five orders of magnitude in our manufacturing, mining and energy generation/distribution capabilities over a space of months or years? How might this increase come about?

Other classes of bottlenecks may also begin to pinch. For example, if software improvements are meant to play a large role in driving the intelligence explosion, then we need to be in a position to speed up all major components of our best algorithms. It’s not so clear we can do that. For example, many computer scientists think that a good number of search algorithms are within sight of optimal performance. There is certainly room to speed them up, but how confident are we that we can quickly find a way to make all relevant search processes orders of magnitude faster than they are now?

The trouble with bottlenecks is that it only takes one bottleneck to slow growth. Placing high confidence in the singularity hypothesis requires placing high confidence in our ability to simultaneously avoid or quickly overcome all possible bottlenecks to growth. That is a tall order.

3. Physical constraints

As the capabilities of artificial agents grow skywards, we confront not merely temporary bottlenecks, but also relatively more permanent physical constraints, including resource constraints and fundamental physical laws.

Begin with resource constraints. The problem is not merely that it could take time to mine more metals, hire more laborers, and find more fossil fuels to burn. The problem is that in trying to produce artificial agents many orders of magnitude more intelligent than current systems, we might actually run out of metals to mine, laborers to hire, and fossil fuels to burn. Of course, that is not to say that there might not be enough materials somewhere in the known universe to meet resource demands. But if, for example, deep-space mining were required to build enough chips to produce radical superintelligence, then it would be relatively less plausible to expect radical superintelligence to emerge on a fast timeline.

Here is a partial list of some of the most troubling resource constraints:

  • Energy generation: Computation requires enormous amounts of energy. These demands are great enough that Sam Altman has said that even AGI, let alone superintelligence, is likely to require a fundamental breakthrough in energy generation, because existing energy resources may be greatly insufficient.
  • Computer chips: Computation requires hardware, and in particular, it requires chips. As the recent chip shortage has shown, manufacturing enough chips to meet global demand is far from trivial. Sam Altman has recently sought as much as five to seven trillion dollars to manufacture enough chips to bring about AGI. That is not a cheap price tag, and it raises the question of how much might be needed on more pessimistic estimates of the remaining gap to AGI, or even on optimistic estimates of the gap to superintelligence.
  • Minerals: Existing chips use scarce minerals such as silicon (for transistors) and gold (as a conductor), as well as a variety of rare earth metals. Both the supply of these minerals and our ability to extract them are deeply limited, and it is not clear how. we might scale up production by many orders of magnitude in a short time scale.

Many more resource constraints could be listed, but perhaps this is enough to make the point.

Turn next to physical laws. As we push our oldest and most successful tricks to their limits, fundamental physical phenomena that were previously irrelevant emerge as sizable obstacles to progress. Two examples deserve particular mention.

First, our smallest transistors are now about ten times as wide as a typical atom. We don’t currently have the capacity to build transistors much smaller than this. Even if we learn to overcome the fiendish difficulty of building ever-smaller transistors, these shrinking transistors will eventually be affected by microphysical phenomena such as quantum uncertainty that previously washed out for all relevant purposes, and which we would really rather not deal with.

Second, feeding more and more energy into computer chips generates increasing need for effective heat dissipation. For many years, heat dissipation advanced in lockstep with changes to the density of chips, but increasingly this is no longer the case, and many worry that the difficulty of dissipating large amounts of concentrated heat will pose a fundamental obstacle to current paradigms for increasing computing capabilities.

It is certainly possible that all of these physical constraints will be in principle superable, and that they can be overcome quickly enough that they will not emerge as bottlenecks to growth. But that view needs an argument, because there are identifiable reasons to suspect that physical constraints are emerging, and will continue to emerge as barriers to growth.

4. Sublinearity of intelligence growth in accessible improvements

Improvements in machine intelligence are driven by hardware and software improvements. I focus on hardware improvements in this section, since they are more easily measured, though I suspect similar conclusions could be drawn for software improvements.

What is the relationship between hardware improvements and improvements in machine intelligence? Suppose that the relationship were roughly linear. Then exponential growth in machine intelligence would require exponential growth in hardware capacities, and hyperbolic growth in machine intelligence would require hyperbolic hardware growth. This still would not be good for the singularity hypothesis. On this model, even an indefinite continuation of Moore’s law (which we saw in Part 2 to be implausible) would yield only exponential intelligence growth, and hyperbolic intelligence growth would require previously unheard-of hyperbolic hardware growth.

However, I think that matters are a bit worse than that. In this section, I argue that machine intelligence plausibly grows sublinearly in hardware capacities. If that is true, then even exponential growth in machine intelligence may require superexponential growth in hardware capacities. And hyperbolic growth? Not easy!

It is relatively unproblematic to measure hardware growth: most traditional measures such as transistor count and FLOPS show consistent exponential hardware growth over the past half-century or so. Has this translated into exponential intelligence growth? That is harder to say, since we would need a measure of machine intelligence, and there is no agreed-upon measure of machine intelligence.

I don’t think we need to choose a measure. I think that all, or nearly all measures of machine intelligence have grown subexponentially during a period of exponential hardware growth. If that is right, then machine intelligence has for some time grown subexponentially in hardware capacities, which gives us some reason to suspect future growth to be subexponential as well.

By way of illustration, Neil Thompson and colleagues look at performance gains in two areas of computer game-play (Chess and Go) as well as three areas that might be thought to depend a great deal on intelligence (protein folding, weather prediction, modeling underground oil reservoirs). I think these may be good illustrative measures of intelligence, but readers are welcome to replace them with their preferred alternative measure and repeat the analysis.

Thompson and colleagues find that recent exponential growth in hardware capacities has led only to linear growth in all five intelligence proxies. Here, for example, are their data for weather prediction: note that predictive improvement (reduced loss) is log-linear in compute growth, i.e. predictive improvement proceeds linearly against a background of exponential compute growth.

Similarly, Thompson and colleagues find a clear log-linear relationship between chess ELO and the computing power of chess engines:

If these and other similar measures are good proxies for intelligence, then intelligence looks to have grown log-linearly rather than linearly in hardware capacities for some time. That would not be good news for the singularity hypothesis.

5. Taking stock

Part 2 of this series gave two reasons for skepticism about the singularity hypothesis: (1) extraordinary claims require extraordinary evidence, and (2) diminishing research productivity threatens gains from self-improvement.

Today’s post gave three further reasons for skepticism: (3) any of a number of bottlenecks may halt growth, (4) resource constraints and fundamental physical laws emerge as obstacles to growth, and (5) intelligence may grow sublinearly in accessible hardware or software improvements, so that rapid growth in intelligence may require extremely rapid growth in these underlying quantities.

Together, these five reasons for skepticism place a strong burden on defenders of the singularity hypothesis to provide significant evidence in favor of their view. If that evidence cannot be produced, then we should not assign significant credence to the singularity hypothesis.

The remainder of this series will argue that this burden has not been met. As a result, I will conclude that we should not assign significant credence to the singularity hypothesis.

Comments

7 responses to “Against the singularity hypothesis (Part 3: Further doubts)”

  1. Vasco Grilo Avatar
    Vasco Grilo

    Thanks for the post, David. Relatedly, readers may want to check out “Explosive growth from AI automation: A review of the arguments” (https://arxiv.org/abs/2309.11690).

    “We examine whether substantial AI automation could accelerate global economic growth by about an order of magnitude, akin to the economic growth effects of the Industrial Revolution. We identify three primary drivers for such growth: 1) the scalability of an AI “labor force” restoring a regime of increasing returns to scale, 2) the rapid expansion of an AI labor force, and 3) a massive increase in output from rapid automation occurring over a brief period of time. Against this backdrop, we evaluate nine counterarguments, including regulatory hurdles, production bottlenecks, alignment issues, and the pace of automation. We tentatively assess these arguments, finding most are unlikely deciders. We conclude that explosive growth seems plausible with AI capable of broadly substituting for human labor, but high confidence in this claim seems currently unwarranted. Key questions remain about the intensity of regulatory responses to AI, physical bottlenecks in production, the economic value of superhuman abilities, and the rate at which AI automation could occur.”

    1. Vasco Grilo Avatar
      Vasco Grilo

      Here is a summary of the article on Epoch’s website: https://epochai.org/blog/explosive-growth-from-ai-a-review-of-the-arguments

    2. David Thorstad Avatar

      Thanks Vasco!

      This looks interesting. I’ll take a look. Nathan Barnard raised a similar issue in the comments on Part 2 of this series.

      The reason why Part 2 specifies the quantity that I am interested in (general intelligence) is that this paper deals primarily with the technological singularity (growth in the intelligence of artificial agents) not the economic singularity (growth in GDP or other economic measures). Some of the arguments in this paper might work against the economic singularity, though not always as well, but others are distinct from the technological singularity.

      While the economic singularity is decidedly a minority view among professional economists, many leading economists do take the idea seriously, particularly when it is given a less radical statement (see i.e. Nordhaus, “Are We Approaching an Economic Singularity? Information Technology and the Future of Economic Growth” for a good opinionated take on the relative plausibility of various way of stating the economic singularity hypothesis). I’m open to taking these views seriously, and the main reason why I don’t talk about them more is that I think they are already in the capable hands of leading economists and I don’t pretend for a moment that I can compete with them.

      I’m thinking about what to write at the conclusion of this series, and it might be interesting to write a comparison of the technological and economic singularity hypotheses. I’m not entirely sure I’m up to the task, because the math gets quite dense, but if I think I can pull it off I might well try.

  2. Kai Williams Avatar
    Kai Williams

    Hi David,
    Thank you again for the post, and the blog. I’ve found it really helpful. While I broadly buy the thesis against the singularity hypothesis, section four has been nagging me a little bit. It is definitely true that in the domains that Thompson et al looked at, intelligence has grown log-linear compared to compute. However, I’m not sure how meaningful that is across all of the domains we care about. If capabilities had grown linearly in these domains compared to computational power, computers would be so good at these problems that it would not be interesting to investigate them in a paper. For instance, a lot of the geometry problems in video game graphics are solvable in polynomial time, so over thirty years, there has been a dramatic improvement in the quality of video games graphics and modeling of the world.

    The question then becomes not whether “intelligence scales linearly with compute” but “in which domains does intelligence scale linearly (or faster) with compute?” There are probably domains which current computer (or AI systems) are not capable, but with more computing power, these capabilities emerge. (I’m thinking of the emergence of arithmetic skills in large LLMs, although that specific example might not carry over). Then, it may still seem like an intelligence explosion to us, if these domains are important enough.

    Taking the line of defense significantly weakens the singularity hypothesis, though. There are definitely important domains with the pattern you suggested. My intuition suggests that building accurate world models (as in weather forecasting) necessarily grows logarithmically compared to compute. Because the world is a chaotic system, prediction is an exponential-time problem. Even just less accurate world models would be a blow to the singularity hypothesis.

    Overall, I don’t think my objection is fatal, or even particularly powerful against your argument. But I do think it does add a wrinkle to the discussion of how responsive intelligence is to computational power.

    1. David Thorstad Avatar

      Thanks Kai, for your engagement and your kind words!

      You’re definitely right that it’s important to investigate, in a domain-specific or metric-specific manner, how intelligence scales with computational power. There are a great many things that people mean by intelligence, and some of the things that they mean might scale relatively well with computational power.

      Maybe the best way to read this section is as an open challenge for advocates of the singularity hypothesis to say what they would like to mean by intelligence and argue on this interpretation both that (a) the relevant quantity grows as quickly as the singularity hypothesis predicts, and (b) the relevant quantity is important enough to have the world-transforming implications that the singularity hypothesis is supposed to have. So for example, if we mean by (a) something like “how many vector operations of a fixed type can you perform in a second”, then on (a) it’s quite plausible that the relevant quantity scales linearly with compute, but now (b) requires a good deal of argument.

      I like your point about the difficulty of predicting chaotic systems. That’s a nice way to think about why predicting the future is hard.

      I hope this is helpful. Sorry if this is a bit shorter than my usual responses. I’m traveling this week and trying to balance speed/accuracy. There is, as we bounded rationality folks are wont to say, a speed/accuracy tradeoff, and one shouldn’t always side against speed in favor of accuracy.

  3. Chris Bao Avatar
    Chris Bao

    A linear increase in elo actually represents an exponential increase in ability (every 400 elo means ~10x greater score ratio), so this specific logarithmic relationship isn’t really convincing. Perhaps you could make a similar argument about other metrics, but I agree that it is difficult to make a convincing argument for either type of relationship.

    1. David Thorstad Avatar

      Thanks Chris! It’s a good point that there are multiple mathematical representations of what is going on with the ELO score and that the choice between them matters. In particular, as you mention, the ratio of expected scores (expected score for player A)/(expected score for player B) grows exponentially in the rating difference between the players.

      To be honest, when I wrote this paper I was trying to push the debate away from a point that I think is relevant here. When I wrote this paper, one of the most common arguments I heard against the singularity hypothesis was some version of the claim that it’s not possible to define intelligence, or that there’s no uncontroversial definition of intelligence. While I have some sympathy for the idea that defining intelligence is hard, I didn’t think it was an especially good way to knock out the singularity hypothesis entirely, so I said outright in the paper that I wouldn’t push on issues with defining intelligence.

      As I say in the paper, this is the one subsection where I can’t really do that. There are lots of different metrics we could claim to track “intelligence,” and these metrics grow at different rates, as in Chess ELO (linear) versus expected score ratio (exponential). So what we really need to do is to ask the advocate of the singularity hypothesis to tell us what they mean by intelligence and give us some plausible proxies which correlate with intelligence and grow at the claimed rate.

      In this specific case — whether intelligence tracks ELO or expected score ratio — I find myself a bit at a loss and more inclined to side with those who think it’s hard to say what intelligence amounts to in this case. I don’t want to go back to the bad old days of treating the difficulty of saying what intelligence amounts to as itself an objection to the singularity hypothesis. But I do think it is unclear in cases like this which quantity should be taken as the correct proxy for intelligence.

      There is, of course, another issue with using either Chess ELO or expected score ratios as a proxy for intelligence. That is that both quantities are almost certainly bounded above, and in some cases (e.g. checkers) we can hit the bound sooner than we expected. We don’t want to end up saying that a system which has solved checkers, chess, or any other game has reached an upper bound on intelligence. I don’t think many people think we’re close to solving chess, but as scores keep ticking upwards we might have more reason to worry about using any metric of chess performance to measure intelligence.

Leave a Reply

Discover more from Reflective altruism

Subscribe now to keep reading and get access to the full archive.

Continue reading