Difference between revisions of "Gaming"

From
Jump to: navigation, search
(John Conway: The Game of Life ...1970)
(Google DeepMind AlphaStar: StarCraft II ... 2019)
Line 72: Line 72:
 
<youtube>x5Q79XCxMVc</youtube>
 
<youtube>x5Q79XCxMVc</youtube>
 
<youtube>eZCI7zu_DlM</youtube>
 
<youtube>eZCI7zu_DlM</youtube>
 +
<youtube>UuhECwm31dM</youtube>
  
 
== [http://openai.com/ OpenAI]: Dota 2 ...2018 ==
 
== [http://openai.com/ OpenAI]: Dota 2 ...2018 ==

Revision as of 08:23, 8 August 2020

Youtube search... ...Google search


google-2019-expanding-the-gamescape.png


NVIDIA: 40 Years on, PAC-MAN ...2020

  • GameGAN, a generative adversarial network trained on 50,000 PAC-MAN episodes, produces a fully functional version of the dot-munching classic without an underlying game engine.

OpenAI: Hide and Seek ... 2019

Brown & Sandholm: 6-player Poker ...2019

Google DeepMind AlphaStar: StarCraft II ... 2019

OpenAI: Dota 2 ...2018

Google DeepMind AlphaGo Zero: Go ...2016

Google DeepMind: Atari video games ...2015

IBM: Watson: Jeopardy ...2011

IBM: Deep Blue: Chess ...1997

John Conway: The Game of Life (GoL) ...1970

Gospers_glider_gun.gif

The Rules

  • For a space that is 'populated':
    • Each cell with one or no neighbors dies, as if by solitude.
    • Each cell with four or more neighbors dies, as if by overpopulation.
    • Each cell with two or three neighbors survives.
  • For a space that is 'empty' or 'unpopulated'
    • Each cell with three neighbors becomes populated.


Donald Waterman: Draw Poker ...1968

Donald Michie: Noughts and Crosses ...1960

MENACE (the Machine Educable Noughts And Crosses Engine) “learns” to play Noughts and Crosses by playing the game repeatedly against another player, each time refining its strategy until after having played a certain number of games it becomes almost perfect and its opponent is only able to draw or lose against it. The learning process involves being “punished” for losing and “rewarded” for drawing or winning, in much the same way that a child learns. This type of machine learning is called Reinforcement Learning (RL). Menace: the Machine Educable Noughts And Crosses Engine | Oliver Child - Chalkdust

MENACE makes a move when the human player randomly picks a bead out of the box that represents the game’s current state. The colour of the bead determines where MENACE will move. In some versions of MENACE, there were beads that only represented more blatant moves such as the side, centre, or corner. The human player chooses the beads at random, just like a neural network’s weights are random at the start. Also like weights, the beads are adjusted when there is failure or success. At the end of each game, if MENACE loses, each bead MENACE used is removed from each box. If MENACE wins, three beads the same as the colour used during each individual turn are added to their respective box. If if the game resulted in a draw, one bead is added. How 300 Matchboxes Learned to Play Tic-Tac-Toe Using MENACE | Caspar Wylie - Open Data Science (ODSC)

img3.jpg



Hexapawn

Arthur Samuel: Checkers ...1950s

Fun!

Books

Invent Your Own Computer Games with Python | Al Sweigart

51mpkckeu4L._SX376_BO1,204,203,200_.jpg

Deep Learning and the Game of Go | Max Pumperla, Kevin Ferguson

51LpAeEYhzL._SX397_BO1,204,203,200_.jpg

Hands-On Deep Learning for Games: Leverage the power of neural networks and reinforcement learning to build intelligent games | Micheal Lanham

517S9nvodoL._SX404_BO1,204,203,200_.jpg

Machine learning and Artificial Intelligence 2.0 with Big Data: Building Video Games using Python 3.7 and Pygame | Narendra Mohan Mittal

41eHxTsXXgL.jpg