27 September 2023

Intelligence quota: The threat of Super-AI and how to beat it

Start the conversation

Wim Naudé* says governments must consider how to prevent an arms race for super-AI, or at least incentivise its creators not to cut corners with safety.


Image: Pete Linforth

The invention of an artificial super-intelligence has been a central theme in science fiction since at least the nineteenth century.

From E.M. Forster’s short story The Machine Stops (1909) to the recent HBO television series Westworld, writers have tended to portray this possibility as an unmitigated disaster.

But this issue is no longer one of fiction.

Prominent contemporary scientists and engineers are now also worried that super-AI could one day surpass human intelligence (an event known as the “singularity”) and become humanity’s “worst mistake”.

Current trends suggest we are set to enter an international arms race for such a technology.

Whichever high-tech firm or government lab succeeds in inventing the first super-AI will obtain a potentially world-dominating technology.

It is a winner-takes-all prize.

So, for those who want to stop such an event, the question is how to discourage this kind of arms race, or at least incentivise competing teams not to cut corners with AI safety.

A super-AI raises two fundamental challenges for its inventors, as philosopher Nick Bostrom and others have pointed out.

One is a control problem, which is how to make sure the super-AI has the same objectives as humanity.

Without this, the intelligence could deliberately, accidently or by neglect destroy humanity — an “AI disaster”.

The second is a political problem, which is how to ensure that the benefits of a super-intelligence do not go only to a small elite, causing massive social and wealth inequalities.

If a super-AI arms race occurs, it could lead competing groups to ignore these problems to develop their technology more quickly.

This could lead to a poor-quality or unfriendly super-AI.

One suggested solution is to use public policy to make it harder to enter the race to reduce the number of competing groups and improve the capabilities of those who do enter.

The fewer who compete, the less pressure there will be to cut corners to win.

But how can governments lessen the competition in this way?

My colleague Nicola Dimitri and I recently published a paper that tried to answer this question.

We first showed that in a typical winner-takes-all race, such as the one to build the first super-AI, only the most competitive teams will participate.

This is because the probability of actually inventing the super-AI is very small, and entering the race is very expensive because of the large investment in research and development needed.

Indeed, this seems to be the current situation with the development of simpler “narrow” AI.

Patent applications for this kind of AI are dominated by a few firms, and the vast bulk of AI research is done in just three regions (the US, China and Europe).

There also seem to be very few, if any, groups presently investing in building a super-AI.

This suggests reducing the number of competing groups isn’t the most important priority at the moment.

But even with smaller numbers of competitors in the race, the intensity of competition could still lead to the problems mentioned above.

So, to reduce the intensity of competition between groups striving to build a super-AI and raise their capabilities, governments could turn to public procurement and taxes.

Public procurement refers to all the things governments pay private companies to provide, from software for use in government agencies to contracts to run services.

Governments could impose constraints on any super-AI supplier that required them to address the potential problems and support complementary technologies to enhance human intelligence and integrate it with AI.

But governments could also offer to buy a less-than-best version of super-AI, effectively creating a “second prize” in the arms race and stopping it from being a winner-takes-all competition.

With an intermediate prize, which could be for inventing something close to (but not exactly) a super-AI, competing groups will have an incentive to invest and cooperate more, reducing the intensity of competition.

A second prize would also reduce the risk of failure and justify more investment, helping to increase the capabilities of the competing teams.

As for taxes, governments could set the tax rate on the group that invents super-AI according to how friendly or unfriendly the AI is.

A high enough tax rate would essentially mean the nationalisation of the super-AI.

This would strongly discourage private firms from cutting corners for fear of losing their product to the state.

Public good not private monopoly

This idea may require better global coordination of taxation and regulation of super-AI.

But it wouldn’t need all governments to be involved.

In theory, a single country or region (such as the EU) could carry the costs and effort involved in tackling the problems and ethics of super-AI.

But all countries would benefit and super-AI would become a public good rather than an unstoppable private monopoly.

Of course, all this depends on super-AI actually being a threat to humanity.

And some scientists don’t think it will be.

We might naturally engineer away the risks of super-AI over time.

Some think humans might even merge with AI.

Whatever the case, our planet and its inhabitants will benefit enormously from making sure we get the best from AI, a technology that is still in its infancy.

For this, we need a better understanding of what role government can play.

* Wim Naudé is Professorial Fellow in the Maastricht Economic and Social Research Institute on Innovation and Technology at the United Nations University.

This article first appeared at theconversation.com.

Start the conversation

Be among the first to get all the Public Sector and Defence news and views that matter.

Subscribe now and receive the latest news, delivered free to your inbox.

By submitting your email address you are agreeing to Region Group's terms and conditions and privacy policy.