Difference between revisions of "Gaming"

From
Jump to: navigation, search
m (More...)
m (Other Videos)
 
Line 242: Line 242:
 
= More... =
 
= More... =
  
= Other Videos =
+
 
 
<youtube>HnkVoOdTiSo</youtube>
 
<youtube>HnkVoOdTiSo</youtube>
 
<youtube>Q8BX0nXfPjY</youtube>
 
<youtube>Q8BX0nXfPjY</youtube>

Latest revision as of 09:07, 17 November 2024

Youtube search... ... Quora ...Google search ...Google News ...Bing News



Gaming Evolution

Meta: Diplomacy 2022

Cicero, has demonstrated the ability to play the strategy game Diplomacy at a level that rivals human performance. Cicero can engage in game conversations and negotiations without most human players realizing they are interacting with a machine. During gameplay on an online league, Cicero sent over 5,000 messages to human players, and its identity as an AI remained undetected. Cicero's performance was impressive, ranking in the top 10% of players. The integration of AI into the game of Diplomacy has shown that machines can effectively mimic human negotiation tactics and strategic thinking. Cicero's achievements in Diplomacy are a testament to the potential of AI in complex human interactions. As AI continues to evolve, it will undoubtedly transform the landscape of diplomacy, offering new tools and methods to support diplomatic efforts.

NVIDIA: 40 Years on, PAC-MAN ...2020

  • GameGAN, a generative adversarial network trained on 50,000 PAC-MAN episodes, produces a fully functional version of the dot-munching classic without an underlying game engine.

OpenAI: Hide and Seek ... 2019

Meta: Brown & Sandholm: 6-player Poker ...2019

  • Occlusions
  • Facebook and Carnegie Mellon (CMU) ‘superhuman’ poker AI beats human pros, ‘It can bluff better than any human.’ | James Vincent - The Verge

Google DeepMind AlphaStar: StarCraft II ... 2019

OpenAI: Dota 2 ...2018

Google DeepMind AlphaGo Zero: Go ...2016

AlphaGo is a computer program developed by Google DeepMind that uses artificial intelligence (AI) to play the board game Go. In 2016, AlphaGo made history by becoming the first computer program to defeat a professional Go player, Lee Sedol, in a five-game match.

During the second game of the match, AlphaGo made a surprising move, known as Move 37, which stunned the Go community and left Lee Sedol speechless. The move involved placing a stone in an unexpected location on the board, which initially appeared to be a mistake. However, as the game progressed, it became clear that the move was part of a complex strategy that allowed AlphaGo to gain an advantage over Lee Sedol. Move 37 is significant because it demonstrated the power of AlphaGo's AI algorithms and its ability to think creatively and strategically. The move was not based on any known human strategy or prior knowledge of the game, but rather on AlphaGo's own analysis and evaluation of the board position.



What would have happened with human-in-the-loop on Move 37?



The move highlighted the limitations of human intuition and the potential for AI to uncover new insights and strategies in complex domains. If a human expert had been involved in the decision-making process for Move 37, they might have questioned AlphaGo's choice and suggested a more conventional move. This could have prevented AlphaGo from making the unexpected and seemingly risky move that ultimately led to its victory.


Andrew Jackson & Josh Hoak: Minigo ...2018

an open source, unofficial implementation of AlphaGo Zero using Reinforcement Learning (RL) approaches can be massively parallelized, so Kubernetes seems like a natural fit, as Kubernetes is all about reducing the overhead for managing applications. However, it can be daunting to wade into Kubernetes and Machine Learning, especially when you add in hardware accelerators like GPUs or TPUs! This talk will break down how you can use Kubernetes and TensorFlow to create, in relatively few lines of code, a tabula rasa AI that can play the game of Go, inspired by the AlphaZero algorithm published by Deepmind. This talk will rely on GPUs, TPUs, TensorFlow, KubeFlow, and large-scale Kubernetes Engine clusters. Uses self-play with Monte Carlo Tree Search and refining the Policy/Value along the way.

Google DeepMind: Atari video games ...2015

IBM: Watson: Jeopardy ...2011

IBM: Deep Blue: Chess ...1997

John Conway: The Game of Life (GoL) ...1970

Gospers_glider_gun.gif

The Rules

  • For a space that is 'populated':
    • Each cell with one or no neighbors dies, as if by solitude.
    • Each cell with four or more neighbors dies, as if by overpopulation.
    • Each cell with two or three neighbors survives.
  • For a space that is 'empty' or 'unpopulated'
    • Each cell with three neighbors becomes populated.

Donald Waterman: Draw Poker ...1968

Martin Gardner: Hexapawn ...1962

A simple game on a 3x3 grid, where each side has 3 chess pawns. The objective is to get a pawn to the other side of the board, or leave the opponent unable to move. Normal chess rules apply except that the pawns are not allowed a double move from their starting position. Not really intended as a two-player game, it was designed to demonstrate an artificial intelligence learning technique by using beads in matchboxes. (Old enough to remember matchboxes?) Twenty-four matchboxes were used to represent the possible moves. Essentially, there were two phases. The first phase was to "teach" the matchbox computer to play the game, then a second phase allowed the matchbox computer to play other opponents. The learning speed depended on the skill of the opponent in the teaching phase. Martin Gardner first published this in his Mathematical Games column in March 1962, and subsequently in his book, "The Unexpected Hanging". Board Game Geek

Donald Michie: Noughts and Crosses ...1960

MENACE (the Machine Educable Noughts And Crosses Engine) “learns” to play Noughts and Crosses by playing the game repeatedly against another player, each time refining its strategy until after having played a certain number of games it becomes almost perfect and its opponent is only able to draw or lose against it. The learning process involves being “punished” for losing and “rewarded” for drawing or winning, in much the same way that a child learns. This type of machine learning is called Reinforcement Learning (RL). Menace: the Machine Educable Noughts And Crosses Engine | Oliver Child - Chalkdust

MENACE makes a move when the human player randomly picks a bead out of the box that represents the game’s current state. The colour of the bead determines where MENACE will move. In some versions of MENACE, there were beads that only represented more blatant moves such as the side, centre, or corner. The human player chooses the beads at random, just like a neural network’s weights are random at the start. Also like weights, the beads are adjusted when there is failure or success. At the end of each game, if MENACE loses, each bead MENACE used is removed from each box. If MENACE wins, three beads the same as the colour used during each individual turn are added to their respective box. If if the game resulted in a draw, one bead is added. How 300 Matchboxes Learned to Play Tic-Tac-Toe Using MENACE | Caspar Wylie - Open Data Science (ODSC)

img3.jpg




Arthur Samuel: Checkers ...1950s

More...

Books

Invent Your Own Computer Games with Python | Al Sweigart

51mpkckeu4L._SX376_BO1,204,203,200_.jpg

Deep Learning and the Game of Go | Max Pumperla, Kevin Ferguson

51LpAeEYhzL._SX397_BO1,204,203,200_.jpg

Hands-On Deep Learning for Games: Leverage the power of neural networks and reinforcement learning to build intelligent games | Micheal Lanham

517S9nvodoL._SX404_BO1,204,203,200_.jpg

Machine learning and Artificial Intelligence 2.0 with Big Data: Building Video Games using Python 3.7 and Pygame | Narendra Mohan Mittal

41eHxTsXXgL.jpg