Question

Answer the following about the methods used by Google’s DeepMind to train AlphaStar, an agent developed to play StarCraft II that reached the highest rank of Grandmaster in 2019. For 10 points each:
[10h] Two answers required. The reinforcement learning procedure used by AlphaStar is based on a policy gradient algorithm in a framework named for these two entities. A popular RL algorithm is named for “Asynchronous Advantage” and these two entities, where policy and value functions are simultaneously learned and updated.
ANSWER: actors and critics [accept (Asynchronous) Advantage Actor-Critic; prompt on A2C or A3C]
[10m] The supervised and reinforcement stages of AlphaStar combined losses using this optimizer. Momentum and RMSProp are precursors to this often-default ML optimization method that has a four-letter acronym.
ANSWER: Adam algorithm [or Adaptive Moment Estimation]
[10e] The multi-agent stage of AlphaStar avoids solely using naive self-play because of its tendency to chase these constructs, leading to an infinite loop. In graphs, these constructs are paths that have the same first and last vertex.
ANSWER: cycles [or circuits]
<Science - Other Science - Computer Science>

Back to bonuses

Summary

2024 ARGOS @ McMaster11/17/2024Y514.00100%40%0%

Data

I'd prefer to have the team name be Christensen et al. than anything that Erik cooks up Simpson Agonistes: The Crisis of Donut001010
You cannot go to Aarhus to see his peat-brown head / With eyes like ripening fruitModerator Can't Neg me While in Alpha0101020
The Only Existing Manuscript from A Clockwork OrangeRyan Wesley Routh's 10 000 NATO-trained Afghan Quizbowlers001010
Communism is Soviet power plus the yassification of the whole countryShe Dicer On My Argonaute Till I RNA Interfere0101020
as rational as the square root of two power bottomsTensei Shitara Flashcard Data Ken001010