READ MORE:
• An artificial-intelligence first: Voice-mimicking software reportedly used in a major theft
• Could company directors be replaced by artificial intelligence?
• UK scientists have trained artificial intelligence to evaluate when people will die
• China is catching up to the US on research into artificial intelligence
Getting that ultimate control problem right could usher in a golden age of abundance. Getting it wrong could result in humanity's extinction. Prof Russell fears it may take a Chernobyl-scale tragedy in AI to alert us to the vital importance of ensuring control.
For the moment, the professor is something of an outlier in the AI community in sounding such alarms. Although he co-wrote a previous textbook on AI that is used by most universities around the world, Prof Russell is critical of what he calls the standard model of AI and the "denialism" of many in the industry.
He is now working with some other researchers and tech companies to design what he calls "provably beneficial" AI systems. The idea is to create more deferential programs that would periodically check in with humans to ensure that they are doing the right thing, rather than blindly following one fixed objective. The goal should be constantly to optimise human preferences.
Stating the challenge so simply is to open a wonder-box of complexity, though. Since the dawn of thought, philosophers have been disputing what makes us happy. Who decides those preferences? How do we weigh the views of the old against the young — and the unborn? What happens if, like King Midas, we make a dumb choice and then change our minds?
Experiments with "wireheading" — artificial brain stimulation — also suggest our preferences can be manipulated. In the wrong conditions, wired-up humans would probably behave just like lab rats do, maniacally pressing a dopamine-releasing pleasure button until we die.
Of course, the challenges of optimising human preferences are not unique to AI. That lofty aim is, in theory, what drives democratic governments every day. It is also supposed to explain the workings of the market. But Prof Russell argues that shareholder capitalism is a good example of a system with a bad objective function, or target. We have instructed corporations to maximise shareholder returns, ignoring other societal preferences, such as preservation of the environment.
"Think of the fossil fuel industry as an AI system. We set it up with this fixed objective of maximising its profit and it has won," Prof Russell says. "It is a machine that has human components but functions as a machine. It has outwitted the rest of the human race because it is a sort of a superintelligent entity."
Just as environmental economists model how to price in negative externalities — by imposing carbon taxes, for example — so computer scientists will have to design more flexible AI programs that reflect a range of societal preferences.
To that end, Prof Russell is launching a computer science course at Berkeley with input from the philosophy and economics departments. The new Centre for Artificial Intelligence at University College London is also adopting a cross-disciplinary approach.
In writing the instruction manual for our future, software engineers will have to create a constant and unbreakable dialogue between machines and humans. AI is too important to be left solely to the computer scientists.
Written by: John Thornhill
© Financial Times