NewsBite

Self-taught, ‘superhuman’ AI now even smarter, says creators

A GOOGLE-owned artificially intelligent computer is “no longer constrained by limits of human knowledge,” its creators say.

Jet Packs, Artificial Intelligence, Data and the Future of Tech

THE computer that stunned humanity by beating the best mortal players at a strategy board game requiring “intuition” has become even smarter, its creators claim.

Even more startling, the updated version of AlphaGo is entirely self-taught — a major step towards the rise of machines that achieve superhuman abilities “with no human input”, they reported in the science journal Nature.

Dubbed AlphaGo Zero, the Artificial Intelligence (AI) system learnt by itself, within days, to master the ancient Chinese board game known as “Go” — said to be the most complex two-person challenge ever invented.

It came up with its own, novel moves to eclipse all the Go acumen humans have acquired over thousands of years.

After just three days of self-training it was put to the ultimate test against AlphaGo, its forerunner which previously dethroned the top human champs.

AlphaGo Zero won by 100 games to zero.

“AlphaGo Zero not only rediscovered the common patterns and openings that humans tend to play ... it ultimately discarded them in preference for its own variants which humans don’t even know about or play at the moment,” said AlphaGo lead researcher David Silver.

The 3000-year-old Chinese game played with black and white stones on a board has more move configurations possible than there are atoms in the Universe.

AlphaGo made world headlines with its shock 4-1 victory in March 2016 over 18-time Go champion Lee Se-Dol, one of the game’s all-time masters.

Go champion Lee Se-Dol plays the Google-owned AlphaGo computer. Picture: Google
Go champion Lee Se-Dol plays the Google-owned AlphaGo computer. Picture: Google
Spectators watch a big screen showing live footage of the third game of the Google DeepMind Challenge Match between Lee Se-Dol, one of the greatest modern players of the ancient board game Go, and the Google-developed supercomputer. Picture: Jung Yeon-Je
Spectators watch a big screen showing live footage of the third game of the Google DeepMind Challenge Match between Lee Se-Dol, one of the greatest modern players of the ancient board game Go, and the Google-developed supercomputer. Picture: Jung Yeon-Je

Lee’s defeat showed that AI was progressing faster than widely thought, said experts at the time who called for rules to make sure powerful AI always remains completely under human control.

In May this year, an updated AlphaGo Master program beat world Number One Ke Jie in three matches out of three.

NOT CONSTRAINED BY HUMANS

Unlike its predecessors which trained on data from thousands of human games before practising by playing against itself, AlphaGo Zero did not learn from humans, or by playing against them, according to researchers at DeepMind, the Google-owned British artificial intelligence (AI) company developing the system.

“All previous versions of AlphaGo ... were told: ‘Well, in this position the human expert played this particular move, and in this other position the human expert played here’,” Silver said in a video explaining the advance.

AlphaGo Zero skipped this step.

Instead, it was programmed to respond to reward — a positive point for a win versus a negative point for a loss.

Starting with just the rules of Go and no instructions, the system learnt the game, devised strategy and improved as it competed against itself — starting with “completely random play” to figure out how the reward is earned. This is a trial-and-error process known as “reinforcement learning”.

Unlike its predecessors, AlphaGo Zero “is no longer constrained by the limits of human knowledge,” Silver and DeepMind CEO Demis Hassabis wrote in a blog.

Amazingly, AlphaGo Zero used a single machine — a human brain-mimicking “neural network” -- compared to the multiple-machine “brain” that beat Lee.

It had four data processing units compared to AlphaGo’s 48, and played 4.9 million training games over three days compared to 30 million over several months.

BEGINNING OF THE END?

“People tend to assume that machine learning is all about big data and massive amounts of computation but actually what we saw with AlphaGo Zero is that algorithms matter much more,” said Silver.

The findings suggested that AI based on reinforcement learning performed better than those that rely on human expertise, Satinder Singh of the University of Michigan wrote in a commentary also carried by Nature.

“However, this is not the beginning of any end because AlphaGo Zero, like all other successful AI so far, is extremely limited in what it knows and in what it can do compared with humans and even other animals,” he said.

AlphaGo Zero’s ability to learn on its own “might appear creepily autonomous”, added Anders Sandberg of the Future of Humanity Institute at Oxford University.

But there was an important difference, he told AFP, “between the general-purpose smarts humans have and the specialised smarts” of computer software.

“What DeepMind has demonstrated over the past years is that one can make software that can be turned into experts in different domains ... but it does not become generally intelligent,” he said.

It was also worth noting that AlphaGo was not programming itself, said Sandberg.

“The clever insights making Zero better was due to humans, not any piece of software suggesting that this approach would be good. I would start to get worried when that happens.”

Is it the dawn of the robot takeover? Picture: Justine Emard
Is it the dawn of the robot takeover? Picture: Justine Emard

Original URL: https://www.news.com.au/technology/innovation/design/selftaught-superhuman-ai-now-even-smarter-says-creators/news-story/ff7799e215e3e98f88828ab31c84fb9a