What's new

New Google AlphaZero AI beats #1 champion chess program after teaching itself in only four hours

Hamartia Antidote

ELITE MEMBER
Joined
Nov 17, 2013
Messages
35,183
Reaction score
30
Country
United States
Location
United States
https://www.theguardian.com/technol...on-program-teaching-itself-to-play-four-hours

3497.jpg


Google’s artificial intelligence sibling DeepMind repurposes Go-playing AI to conquer chess and shogi without aid of human knowledge

AlphaZero, the game-playing AI created by Google sibling DeepMind, has beaten the world’s best chess-playing computer program, having taught itself how to play in under four hours.

The repurposed AI, which has repeatedly beaten the world’s best Go players as AlphaGo, has been generalised so that it can now learn other games. It took just four hours to learn the rules to chess before beating the world champion chess program, Stockfish 8, in a 100-game match up.

AlphaZero won or drew all 100 games, according to a non-peer-reviewed research paper published with Cornell University Library’s arXiv.

“Starting from random play, and given no domain knowledge except the game rules, AlphaZero achieved within 24 hours a superhuman level of play in the games of chess and shogi [a similar Japanese board game] as well as Go, and convincingly defeated a world-champion program in each case,” said the paper’s authors that include DeepMind founder Demis Hassabis, who was a child chess prodigy reaching master standard at the age of 13.

“It’s a remarkable achievement, even if we should have expected it after AlphaGo,” former world chess champion Garry Kasparov told Chess.com. “We have always assumed that chess required too much empirical knowledge for a machine to play so well from scratch, with no human knowledge added at all.”

Computer programs have been able to beat the best human chess players ever since IBM’s Deep Blue supercomputer defeated Kasparov on 12 May 1997.


FacebookTwitterPinterest
Chess enthusiasts watch World Chess champion Garry Kasparov on a television monitor in 1997. Photograph: Stan Honda/AFP/Getty Images
DeepMind said the difference between AlphaZero and its competitors is that its machine-learning approach is given no human input apart from the basic rules of chess. The rest it works out by playing itself over and over with self-reinforced knowledge. The result, according to DeepMind, is that AlphaZero took an “arguably more human-like approach” to the search for moves, processing around 80,000 positions per second in chess compared to Stockfish 8’s 70m.

After winning 25 games of chess versus Stockfish 8 starting as white, with first-mover advantage, a further three starting with black and drawing a further 72 games, AlphaZero also learned shogi in two hours before beating the leading program Elmo in a 100-game matchup. AlphaZero won 90 games, lost eight and drew 2.

The new generalised AlphaZero was also able to beat the “super human” former version of itself AlphaGo at the Chinese game of Go after only eight-hours of self-training, winning 60 games and losing 40 games.

While experts said the results are impressive, and have potential across a wide-range of applications to complement human knowledge, professor Joanna Bryson, a computer scientist and AI researcher at the University of Bath, warned that it was “still a discrete task”.
 
Last edited:
https://www.technologyreview.com/s/...ss-shows-the-power-and-the-peculiarity-of-ai/

Alpha Zero’s “Alien” Chess Shows the Power, and the Peculiarity, of AI

gettyimages-515689368.jpg


The latest AI program developed by DeepMind is not only brilliant and remarkably flexible—it’s also quite weird.

DeepMind published a paper this week describing a game-playing program it developed that proved capable of mastering chess and the Japanese game Shoju, having already mastered the game of Go.

Demis Hassabis, the founder and CEO of DeepMind and an expert chess player himself, presented further details of the system, called Alpha Zero, at an AI conference in California on Thursday. The program often made moves that would seem unthinkable to a human chess player.

“It doesn’t play like a human, and it doesn’t play like a program,” Hassabis said at the Neural Information Processing Systems (NIPS) conference in Long Beach. “It plays in a third, almost alien, way.”

Besides showing how brilliant machine-learning programs can be at a specific task, this shows that artificial intelligence can be quite different from the human kind. As AI becomes more commonplace, we might need to be conscious of such “alien” behavior.

Alpha Zero is a more general version of AlphaGo, the program developed by DeepMind to play the board game Go. In 24 hours, Alpha Zero taught itself to play chess well enough to beat one of the best existing chess programs around.

What’s also remarkable, though, Hassabis explained, is that it sometimes makes seemingly crazy sacrifices, like offering up a bishop and queen to exploit a positional advantage that led to victory. Such sacrifices of high-value pieces are normally rare. In another case the program moved its queen to the corner of the board, a very bizarre trick with a surprising positional value. “It’s like chess from another dimension,” Hassabis said.

Hassabis speculates that because Alpha Zero teaches itself, it benefits from not following the usual approach of assigning value to pieces and trying to minimize losses. “Maybe our conception of chess has been too limited,” he said. “It could be an important moment for chess. We can graft it into our own play.”

The game of chess has a long history in artificial intelligence. The best programs, developed and refined over decades, incorporate huge amounts of human intelligence. Although in 1996 IBM’s Deep Blue beat the world champion at the time, that program, like other conventional chess programs, required careful hand-programming.

The original AlphaGo, designed specifically for Go, was a big deal because it was capable of learning to play a game that is enormously complex and is difficult to teach, requiring an instinctive sense of board positions. AlphaGo mastered Go by ingesting thousands of example games and then practicing against another version of itself. It did this partially by training a large neural network using an approach known as reinforcement learning, which is modeled on the way animals seem to learn (see “Google’s AI Masters Go a Decade Earlier Than Expected”).

DeepMind has since demonstrated a version of the program, called AlphaGo Zero, that learns without any example games, instead relying purely on self-play (see “AlphaGo Zero Shows Machines Can Become Superhuman Without Any Help”). Alpha Zero improves further still by showing that the same program can master three different types of board games.

Alpha Zero’s achievements are impressive, but it still needs to play many more practice games than a human chess master. Hassabis says this may be because humans benefit from other forms of learning, such as reading about how to play the game and watching other people play.

Still, some experts caution that the program’s capabilities, while remarkable, should be taken in context. Speaking after Hassabis, Gary Marcus, a professor at NYU, said that a great deal of human knowledge went into building Alpha Zero. And he suggests that human intelligence seems to involve some innate capabilities, such as an intuitive ability to develop language.

Josh Tenenbaum, a professor at MIT who studies human intelligence, said that if we want to develop real, human-level artificial intelligence, we should study the flexibility and creativity that humans exhibit. He pointed, among other examples, to the intelligence of Hassabis and his colleagues in devising, designing, and building the program in the first place. “That’s almost as impressive as a queen in the corner,” he quipped.
 
Damn man:o:
 

I think it a foregone conclusion we are all doomed. When they load this on a Quantum computer and it pulls a split second Matrix style "I know how to fly a helicopter" thing we are all in big trouble.
 

Pakistan Affairs Latest Posts

Back
Top Bottom