About 49,100 results
Open links in new tab
  1. What is the significance of move 37? (to a non go player)

    Feb 26, 2023 · 1 I have seen (and googled) information for Game 2, Move 37 in the AlphaGo vs. Lee Sedol match However it is difficult to find information concerning this move that doesn't …

  2. Did Alphago zero actually beat Alphago 100 games to 0?

    Oct 21, 2020 · 2 tl;dr Did AlphaGo and AlphaGo play 100 repetitions of the same sequence of boards, or were there 100 different games? Background: Alphago was the first superhuman go …

  3. terminology - AlphaGo Zero MCTS a misnomer? - Artificial …

    Dec 19, 2024 · AlphaGo Zero utilizes an algorithm it calls "Monte Carlo Tree Search" in the papers, but it seems like the algorithm is fully deterministic since it doesn't utilize random …

  4. Newest 'alphago' Questions - Artificial Intelligence Stack Exchange

    For questions related to DeepMind's AlphaGo, which is the first computer Go program to beat a human professional Go player without handicaps on a full-sized 19x19 board. AlphaGo was …

  5. deep learning - What is the input to AlphaGo's neural network ...

    Jun 8, 2020 · AlphaGo Zero only uses the black and white stones from the Go board as its input, whereas previous versions of AlphaGo included a small number of hand-engineered features. …

  6. How does Alpha Go Zero MCTS work in parallel?

    Sep 25, 2023 · To understand how AlphaGo Zero performs parallel simulations think of each simulation as a separate agent that interacts with the search tree. Each agent starts from the …

  7. Difference between AlphaGo's policy network and value network

    Mar 29, 2016 · If anyone else stumbles upon this old question, like me, you'll be pleased to know that AlphaGo's successor, "AlphaGo Zero", as well as its successor "AlphaZero" do indeed get …

  8. Why AlphaGo didn't use Deep Q-Learning?

    Apr 29, 2020 · In the previous research, in 2015, Deep Q-Learning shows its great performance on single player Atari Games. But why do AlphaGo's researchers use CNN + MCTS instead of …

  9. Initialising DQN with weights from imitation learning rather than ...

    Nov 14, 2020 · In AlphaGo, the authors initialised a policy gradient network with weights trained from imitation learning. I believe this gives it a very good starting policy for the policy gradient …

  10. Would AlphaZero perform better if made with transformers?

    Feb 24, 2024 · AlphaZero utilized a residual convolutional neural network to estimate move policy and position value. If it was rebuilt today, would it be more efficient and powerful if they used a …